Does Hadoop support the environment where nodes join and leave without a 
preconfigured file like "hadoop-site.xml" ?  The characteristic is that none of 
the IP addresses and node names of any machines are stable.  They will change 
after the machine is reboot after crash.

Before that, I use a simple way of just configuring my hadoop-site.xml and use 
the startup scripts that takes care of everything.  But for the dynamic IP 
address scenario, that doesn't seem to work.  Can someone suggest a solution 
how to deal with this scenario ?

Here are the considerations ...

Startup Discovery Scenario
===========================
How does a NameNode knows a newly joined DataNode ?
How does a new DataNode knows the existing NameNode ?
How does a JobTracker knows a newly joined TaskTracker ?
How does a new TaskTracker knows the existing JobTracker ?

Fail Recovery Scenario
=======================
Lets say a NameNode crash, and then another NameNode (at a different address) 
starts up.  How does the new NameNode learnt about other DataNodes ?
How does other DataNodes learn about this new NameNode ?

Lets say a JobTracker crash, and then another JobTracker (at a different 
address) starts up.  How does the new JobTracker learnt about other 
TaskTrackers ?
How does other TaskTrackers learn about this new JobTracker ?

Lets say a DataNode crash, and then another DataNode (at a different address) 
starts up.  How does the new DataNode learnt about the existing NameNode ?
How does the existing NameNode learn about this new DataNode ?

Lets say a TaskTracker crash, and then another TaskTracker (at a different 
address) starts up.  How does the new TaskTracker learnt about the existing 
JobTracker ?
How does the existing JobTracker learn about this new TaskTracker ?


Rgds,
Ricky

Reply via email to