Hi,

I have a case of a corrupt HDFS (according to bin/hadoop fsck) and I'm trying 
not to lose the precious data in it.  I accidentally run bin/hadoop namenode 
-format on a *new DN* that I just added to the cluster.  Is it possible for 
that to corrupt HDFS?  I also had to explicitly kill DN daemons before that, 
because bin/stop-all.sh didn't stop them for some reason (it always did so 
before).

Is there any way to salvage the data?  I have a 4-node cluster with replication 
factor of 3, though fsck reports lots of under-replicated blocks:

  ********************************
  CORRUPT FILES:        3355
  MISSING BLOCKS:       3462
  MISSING SIZE:         17708821225 B
  ********************************
 Minimally replicated blocks:   28802 (89.269775 %)
 Over-replicated blocks:        0 (0.0 %)
 Under-replicated blocks:       17025 (52.76779 %)
 Mis-replicated blocks:         0 (0.0 %)
 Default replication factor:    3
 Average block replication:     1.7750744
 Missing replicas:              17025 (29.727087 %)
 Number of data-nodes:          4
 Number of racks:               1


The filesystem under path '/' is CORRUPT


What can one do at this point to save the data?  If I run bin/hadoop fsck -move 
or -delete will I lose some of the data?  Or will I simply end up with fewer 
block replicas and will thus have to force re-balancing in order to get back to 
a "safe" number of replicas?

Thanks,
Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch

Reply via email to