datanode
fail or
dfs
full

Sent from Meizu MX


-------- 原始邮件 --------
发件人:Subroto <ssan...@datameer.com>
时间:2012-6-20 15:07
收件人:hdfs-user@hadoop.apache.org
主题:Re: DFSClient not able to write

Thanks Todd,

Yups… the problem was Disk Space….. I checked the NameNode UI which mentioned 
DFS Remaining as 0%

Is it possible to get some more detailed exception message in logs because, 
none of the logs (DataNode or Client) mention about this problem. ?
May be the NameNode UI also can indicate the same…..

Just like to know where the point:
> 5*BLOCK_SIZE
is documented….  :-(

Thanks again...

Cheers,
Subroto Sanyal

On Jun 19, 2012, at 6:22 PM, Todd Lipcon wrote:

> Hi Subroto,
> 
> Anything in the DN logs? Maybe the one DN is out of disk space? You
> need to have 5*BLOCK_SIZE free.
> 
> -Todd
> 
> On Tue, Jun 19, 2012 at 1:29 AM, Subroto <ssan...@datameer.com> wrote:
>> Hi,
>> 
>> When DFS cluster;
>> The datanodes keep on dumping the logs:
>> 2012-06-19 04:05:04,842 INFO 
>> org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl: Adding 
>> block pool BP-1233167859-10.10.35.8-1340093005960
>> 2012-06-19 04:05:04,861 INFO 
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool 
>> BP-1233167859-10.10.35.8-1340093005960 (storage id 
>> DS-1255092928-10.244.15.192-50010-1340093104449) service to 
>> ec2-107-22-39-220.compute-1.amazonaws.com/10.10.35.8:9000 beginning 
>> handshake with NN
>> 2012-06-19 04:05:04,979 INFO 
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Block pool Block pool 
>> BP-1233167859-10.10.35.8-1340093005960 (storage id 
>> DS-1255092928-10.244.15.192-50010-1340093104449) service to 
>> ec2-107-22-39-220.compute-1.amazonaws.com/10.10.35.8:9000 successfully 
>> registered with NN
>> 2012-06-19 04:05:04,980 INFO 
>> org.apache.hadoop.hdfs.server.datanode.DataNode: For namenode 
>> ec2-107-22-39-220.compute-1.amazonaws.com/10.10.35.8:9000 using 
>> DELETEREPORT_INTERVAL of 300000 msec  BLOCKREPORT_INTERVAL of 21600000msec 
>> Initial delay: 0msec; heartBeatInterval=3000
>> 2012-06-19 04:05:05,258 INFO 
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Namenode Block pool 
>> BP-1233167859-10.10.35.8-1340093005960 (storage id 
>> DS-1255092928-10.244.15.192-50010-1340093104449) service to 
>> ec2-107-22-39-220.compute-1.amazonaws.com/10.10.35.8:9000 trying to claim 
>> ACTIVE state with txid=24
>> 2012-06-19 04:05:05,258 INFO 
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Acknowledging ACTIVE 
>> Namenode Block pool BP-1233167859-10.10.35.8-1340093005960 (storage id 
>> DS-1255092928-10.244.15.192-50010-1340093104449) service to 
>> ec2-107-22-39-220.compute-1.amazonaws.com/10.10.35.8:9000
>> 2012-06-19 04:05:05,394 INFO 
>> org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 blocks 
>> took 1 msec to generate and 135 msecs for RPC and NN processing
>> 2012-06-19 04:05:05,394 INFO 
>> org.apache.hadoop.hdfs.server.datanode.DataNode: sent block report, 
>> processed 
>> command:org.apache.hadoop.hdfs.server.protocol.FinalizeCommand@49d31859
>> 2012-06-19 04:05:05,396 INFO 
>> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Periodic Block 
>> Verification Scanner initialized with interval 504 hours for block pool 
>> BP-1233167859-10.10.35.8-1340093005960.
>> 2012-06-19 04:05:05,400 INFO 
>> org.apache.hadoop.hdfs.server.datanode.DataBlockScanner: Added 
>> bpid=BP-1233167859-10.10.35.8-1340093005960 to blockPoolScannerMap, new 
>> size=1
>> 2012-06-19 04:05:09,845 INFO 
>> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new 
>> period : work left in prev period : 0.00%
>> 2012-06-19 04:05:14,848 INFO 
>> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new 
>> period : work left in prev period : 0.00%
>> 2012-06-19 04:05:19,849 INFO 
>> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new 
>> period : work left in prev period : 0.00%
>> 2012-06-19 04:05:24,850 INFO 
>> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new 
>> period : work left in prev period : 0.00%
>> 2012-06-19 04:05:29,851 INFO 
>> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new 
>> period : work left in prev period : 0.00%
>> 2012-06-19 04:05:34,852 INFO 
>> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new 
>> period : work left in prev period : 0.00%
>> 2012-06-19 04:05:39,853 INFO 
>> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new 
>> period : work left in prev period : 0.00%
>> 2012-06-19 04:05:44,854 INFO 
>> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new 
>> period : work left in prev period : 0.00%
>> 2012-06-19 04:05:49,855 INFO 
>> org.apache.hadoop.hdfs.server.datanode.BlockPoolSliceScanner: Starting a new 
>> period : work left in prev period : 0.00%
>> 
>> The DFSClient operations are also not taking place properly:
>> 2012-06-19 04:04:56,653 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer 
>> Exception
>> java.io.IOException: File /hbase/hbase.version could only be replicated to 0 
>> nodes instead of minReplication (=1).  There are 1 datanode(s) running and 
>> no node(s) are excluded in this operation.
>>        at 
>> org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1256)
>>        at 
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1977)
>>        at 
>> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:470)
>>        at 
>> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:292)
>>        at 
>> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:42602)
>>        at 
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:427)
>>        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:916)
>>        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1692)
>>        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1688)
>>        at java.security.AccessController.doPrivileged(Native Method)
>>        at javax.security.auth.Subject.doAs(Subject.java:396)
>>        at 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1232)
>> 
>> Please let me know if there is any configuration which needs to be set…..
>> The hadoop version being used is: 2.0.0
>> 
>> 
>> 
>> Cheers,
>> Subroto Sanyal
> 
> 
> 
> -- 
> Todd Lipcon
> Software Engineer, Cloudera

Reply via email to