Hi,
Can we specify which subset of machines to use for different jobs? E.g. We
set machine A as namenode, and B, C, D as datanodes. Then for job 1, we have
a mapreduce that runs on B C and for job 2, the map-reduce runs on C D.
Regards,
Raakhi
Hi Raakhi,
Unfortunately there is no built-in way of doing this. You'd have to
instantiate two entirely separate Hadoop clusters to accomplish what you're
trying to do, which isn't an uncommon thing to do.
I'm not sure why you're hoping to have this behavior, but the fair share
scheduler might