hi, i use CDH4.4

On Fri, Feb 21, 2014 at 12:04 PM, Ted Yu <yuzhih...@gmail.com> wrote:

> Which hadoop release are you using ?
>
> Cheers
>
>
> On Thu, Feb 20, 2014 at 8:57 PM, ch huang <justlo...@gmail.com> wrote:
>
>>  hi,maillist:
>>           i see the following info in my hdfs log ,and the block belong
>> to the file which write by scribe ,i do not know why
>> is there any limit in hdfs system ?
>>
>> 2014-02-21 10:33:30,235 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: opReadBlock
>> BP-1043055049-192.168.11.11-1382442676609:blk_-8536558734938003208_3823240
>> received exc
>> eption java.io.IOException: Replica gen stamp < block genstamp,
>> block=BP-1043055049-192.168.11.11-1382442676609:blk_-8536558734938003208_3823240,
>> replica=ReplicaWaitingToBeRecov
>> ered, blk_-8536558734938003208_3820986, RWR
>>   getNumBytes()     = 35840
>>   getBytesOnDisk()  = 35840
>>   getVisibleLength()= -1
>>   getVolume()       = /data/4/dn/current
>>   getBlockFile()    =
>> /data/4/dn/current/BP-1043055049-192.168.11.11-1382442676609/current/rbw/blk_-8536558734938003208
>>   unlinked=false
>> 2014-02-21 10:33:30,235 WARN
>> org.apache.hadoop.hdfs.server.datanode.DataNode:
>> DatanodeRegistration(192.168.11.12,
>> storageID=DS-754202132-192.168.11.12-50010-1382443087835, infoP
>> ort=50075, ipcPort=50020,
>> storageInfo=lv=-40;cid=CID-0e777b8c-19f3-44a1-8af1-916877f2506c;nsid=2086828354;c=0):Got
>> exception while serving BP-1043055049-192.168.11.11-1382442676
>> 609:blk_-8536558734938003208_3823240 to /192.168.11.15:56564
>> java.io.IOException: Replica gen stamp < block genstamp,
>> block=BP-1043055049-192.168.11.11-1382442676609:blk_-8536558734938003208_3823240,
>> replica=ReplicaWaitingToBeRecovered, b
>> lk_-8536558734938003208_3820986, RWR
>>   getNumBytes()     = 35840
>>   getBytesOnDisk()  = 35840
>>   getVisibleLength()= -1
>>   getVolume()       = /data/4/dn/current
>>   getBlockFile()    =
>> /data/4/dn/current/BP-1043055049-192.168.11.11-1382442676609/current/rbw/blk_-8536558734938003208
>>   unlinked=false
>>         at
>> org.apache.hadoop.hdfs.server.datanode.BlockSender.<init>(BlockSender.java:205)
>>         at
>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:326)
>>         at
>> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opReadBlock(Receiver.java:92)
>>         at
>> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:64)
>>         at
>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:221)
>>         at java.lang.Thread.run(Thread.java:744)
>> 2014-02-21 10:33:30,236 ERROR
>> org.apache.hadoop.hdfs.server.datanode.DataNode: ch12:50010:DataXceiver
>> error processing READ_BLOCK operation  src: /192.168.11.15:56564 dest: /
>> 192.168.11.12:50010
>> java.io.IOException: Replica gen stamp < block genstamp,
>> block=BP-1043055049-192.168.11.11-1382442676609:blk_-8536558734938003208_3823240,
>> replica=ReplicaWaitingToBeRecovered, blk_-8536558734938003208_3820986, RWR
>>   getNumBytes()     = 35840
>>   getBytesOnDisk()  = 35840
>>   getVisibleLength()= -1
>>   getVolume()       = /data/4/dn/current
>>   getBlockFile()    =
>> /data/4/dn/current/BP-1043055049-192.168.11.11-1382442676609/current/rbw/blk_-8536558734938003208
>>   unlinked=false
>>         at
>> org.apache.hadoop.hdfs.server.datanode.BlockSender.<init>(BlockSender.java:205)
>>         at
>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:326)
>>         at
>> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opReadBlock(Receiver.java:92)
>>         at
>> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:64)
>>         at
>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:221)
>>         at java.lang.Thread.run(Thread.java:744)
>>
>
>

Reply via email to