I think might have come up before (but couldn't find something in the
archives) and was meant to be not serious but we just had a node spit
out 5000 errors per hour of these
2007-08-19T20:50:37+00:00 [192.168.165.136] [user.err]
[hadoop.org.apache.hadoop.dfs.DataNode] DataXCeiver
2007-08-19T20:50:37+00:00 [192.168.165.136] [user.err]
java.io.IOException: Block blk_-6080554387210237245 is valid, and
cannot be written to.
2007-08-19T20:50:37+00:00 [192.168.165.136] [user.err] at
org.apache.hadoop.dfs.FSDataset.writeToBlock(FSDataset.java:490)
2007-08-19T20:50:37+00:00 [192.168.165.136] [user.err] at
org.apache.hadoop.dfs.DataNode$DataXceiver.writeBlock(DataNode.java:734)
2007-08-19T20:50:37+00:00 [192.168.165.136] [user.err] at
org.apache.hadoop.dfs.DataNode$DataXceiver.run(DataNode.java:563)
2007-08-19T20:50:37+00:00 [192.168.165.136] [user.err] at
java.lang.Thread.run(Thread.java:595)
which cannot be good. We are still running version 0.10.1 (I know -
we need to upgrade!) ...but could someone please explain what this
really means?
The node was still listed as in service and the fsck also confirmed
the hdfs to be OK.
cheers
--
Torsten
- is valid, and cannot be written to Torsten Curdt
-