Hi,
I am running Hbase in pseudo distributed mode.( Hbase 0.94.7 and Hadoop
1.1.2).
I am getting certain exceptions in Hadoop's namenode and data node files
which are :-

Namenode :-

2013-10-18 10:33:37,218 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
NameSystem.heartbeatCheck: lost heartbeat from 192.168.20.30:50010
2013-10-18 10:33:37,242 INFO org.apache.hadoop.net.NetworkTopology:
Removing a node: /default-rack/192.168.20.30:50010
2013-10-18 10:35:27,606 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
transactions: 64 Total time for transactions(ms): 1Number
of transactions batched in Syncs: 0 Number of syncs: 43 SyncTimes(ms): 86
2013-10-18 10:35:27,614 ERROR
org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
as:hadoop cause:java.io.IOException: File /h
base/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e
could only be replicated to 0 nodes, instead of 1
2013-10-18 10:35:27,895 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 9 on 9000, call
addBlock(/hbase/event_data/433b61f2a4ebff8f2e4b89890508a
3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e,
DFSClient_hb_rs_hbase.rummycircle.com,60020,1382012725057, null) from
192.168.20.30:44990: error: java.io.I
OException: File
/hbase/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e
could only be replicated to 0 nodes, instead
 of 1
java.io.IOException: File
/hbase/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e
could only be replicated to 0 nodes
, instead of 1
    at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1639)
    at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:736)
    at sun.reflect.GeneratedMethodAccessor10.invoke(Unknown Source)
    at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:578)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1393)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1389)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:396)
    at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1149)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1387)


Data node :-

2013-10-18 06:13:14,499 WARN
org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
192.168.20.30:50010, storageID=DS-1816106352-192.16
8.20.30-50010-1369314076237, infoPort=50075, ipcPort=50020):Got exception
while serving blk_-3215981820534544354_52215 to /192.168.20.30:
java.net.SocketTimeoutException: 480000 millis timeout while waiting for
channel to be ready for write. ch :
java.nio.channels.SocketChannel[connected
 local=/192.168.20.30:50010 remote=/192.168.20.30:36188]
    at
org.apache.hadoop.net.SocketIOWithTimeout.waitForIO(SocketIOWithTimeout.java:246)
    at
org.apache.hadoop.net.SocketOutputStream.waitForWritable(SocketOutputStream.java:159)
    at
org.apache.hadoop.net.SocketOutputStream.transferToFully(SocketOutputStream.java:198)
    at
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendChunks(BlockSender.java:392)
    at
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendBlock(BlockSender.java:490)
    at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:202)
    at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:104)
    at java.lang.Thread.run(Thread.java:662)







-- 
Thanks and Regards,
Vimal Jain

Reply via email to