>From the timestamp point of view, the only directory seems to be modified and removed is the Current directory under dfs.home.dir. However, the storage file under dfs.home.dir is untouched since the datanode started.
On Tue, Apr 12, 2011 at 8:02 AM, felix gao <gre1...@gmail.com> wrote: > The xml files have not been changed for more than two months, so that > should not be the reason. Even the in_use.lock is more than a month old. > However, we did shut it down few days ago and restarted it afterward. Then > the second shutdown might not be clean. > > > On Tue, Apr 12, 2011 at 7:52 AM, Ayon Sinha <ayonsi...@yahoo.com> wrote: > >> The datanode used the dfs config xml file to tell the datanode process, >> what disks are available for storage. Can you check that the config xml has >> all the partitions mentioned and has not been overwritten during the restore >> process? >> >> -Ayon >> See My Photos on Flickr <http://www.flickr.com/photos/ayonsinha/> >> Also check out my Blog for answers to commonly asked >> questions.<http://dailyadvisor.blogspot.com> >> >> >> ------------------------------ >> *From:* felix gao <gre1...@gmail.com> >> *To:* hdfs-user@hadoop.apache.org >> *Sent:* Tue, April 12, 2011 7:46:31 AM >> *Subject:* Question regarding datanode been wiped by hadoop >> >> What reason/condition would cause a datanode’s blocks to be removed? Our >> cluster had a one of its datanodes crash because of bad RAM. After the >> system was upgraded and the datanode/tasktracker brought online the next day >> we noticed the amount of space utilized was minimal and the cluster was >> rebalancing blocks to the datanode. It would seem the prior blocks were >> removed. Was this because the datanode was declared dead? What is the >> criteria for a namenode to decide (Assuming its the namenode) when a >> datanode should remove prior blocks? >> > >