Files still rotting in DFS of latest Hadoop
-------------------------------------------

         Key: HADOOP-124
         URL: http://issues.apache.org/jira/browse/HADOOP-124
     Project: Hadoop
        Type: Bug

  Components: dfs  
 Environment: ~30 node cluster
    Reporter: Bryan Pendleton


DFS files are still rotting.

I suspect that there's a problem with block accounting/detecting identical 
hosts in the namenode. I have 30 physical nodes, with various numbers of local 
disks, meaning that my current 'bin/hadoop dfs -report" shows 80 nodes after a 
full restart. However, when I discovered the  problem (which resulted in losing 
about 500gb worth of temporary data because of missing blocks in some of the 
larger chunks) -report showed 96 nodes. I suspect somehow there were extra 
datanodes running against the same paths, and that the namenode was counting 
those as replicated instances, which then showed up over-replicated, and one of 
them was told to delete its local block, leading to the block actually getting 
lost.

I will debug it more the next time the situation arises. This is at least the 
5th time I've had a large amount of file data "rot" in DFS since January.

-- 
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators:
   http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see:
   http://www.atlassian.com/software/jira

Reply via email to