Hi All, Ive been working michael nolls multi-node cluster setup example (Running_Hadoop_On_Ubuntu_Linux) for hadoop and I have a working setup. I then on my slave machine -- which is currently running a datanode killed the process in an effort to try to simulate some sort of failure on the slave machine datanode. I had assumed that the namenode would have been polling its datanodes and thus attempted to bring up any node that goes down. On looking at my slave machine it seems that the datanode process is still down (I've checked jps).
Obviously im missing something ! Does hadoop look after its datanodes ? Is there a config setting that i may have missed ? Do I need to create some sort of external tool to pool and attempt to bring up nodes that have gone down ? Thanks Will -- View this message in context: http://www.nabble.com/How-does-an-offline-Datanode-come-back-up---tp20192214p20192214.html Sent from the Hadoop lucene-users mailing list archive at Nabble.com.