I will try that if i get them next time. Could anyone please give the cause of this exceptions ?
On Fri, Oct 18, 2013 at 4:03 PM, divye sheth <divs.sh...@gmail.com> wrote: > I would recommend you to stop the cluster and then start the daemons one by > one. > 1. stop-dfs.sh > 2. hadoop-daemon.sh start namenode > 3. hadoop-daemon.sh start datanode > > This will show start up errors if any, also verify if the datanode is able > to communicate with the namenode. > > Thanks > Divye Sheth > > > On Fri, Oct 18, 2013 at 3:51 PM, Vimal Jain <vkj...@gmail.com> wrote: > > > Hi, > > I am running Hbase in pseudo distributed mode.( Hbase 0.94.7 and Hadoop > > 1.1.2). > > I am getting certain exceptions in Hadoop's namenode and data node files > > which are :- > > > > Namenode :- > > > > 2013-10-18 10:33:37,218 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* > > NameSystem.heartbeatCheck: lost heartbeat from 192.168.20.30:50010 > > 2013-10-18 10:33:37,242 INFO org.apache.hadoop.net.NetworkTopology: > > Removing a node: /default-rack/192.168.20.30:50010 > > 2013-10-18 10:35:27,606 INFO > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of > > transactions: 64 Total time for transactions(ms): 1Number > > of transactions batched in Syncs: 0 Number of syncs: 43 SyncTimes(ms): 86 > > 2013-10-18 10:35:27,614 ERROR > > org.apache.hadoop.security.UserGroupInformation: > PriviledgedActionException > > as:hadoop cause:java.io.IOException: File /h > > > > > base/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e > > could only be replicated to 0 nodes, instead of 1 > > 2013-10-18 10:35:27,895 INFO org.apache.hadoop.ipc.Server: IPC Server > > handler 9 on 9000, call > > addBlock(/hbase/event_data/433b61f2a4ebff8f2e4b89890508a > > 3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e, > > DFSClient_hb_rs_hbase.rummycircle.com,60020,1382012725057, null) from > > 192.168.20.30:44990: error: java.io.I > > OException: File > > > > > /hbase/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e > > could only be replicated to 0 nodes, instead > > of 1 > > java.io.IOException: File > > > > > /hbase/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e > > could only be replicated to 0 nodes > > , instead of 1 > > at > > > > > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1639) > > at > > > org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:736) > > at sun.reflect.GeneratedMethodAccessor10.invoke(Unknown Source) > > at > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > > at java.lang.reflect.Method.invoke(Method.java:597) > > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:578) > > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1393) > > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1389) > > at java.security.AccessController.doPrivileged(Native Method) > > at javax.security.auth.Subject.doAs(Subject.java:396) > > at > > > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1149) > > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1387) > > > > > > Data node :- > > > > 2013-10-18 06:13:14,499 WARN > > org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration( > > 192.168.20.30:50010, storageID=DS-1816106352-192.16 > > 8.20.30-50010-1369314076237, infoPort=50075, ipcPort=50020):Got exception > > while serving blk_-3215981820534544354_52215 to /192.168.20.30: > > java.net.SocketTimeoutException: 480000 millis timeout while waiting for > > channel to be ready for write. ch : > > java.nio.channels.SocketChannel[connected > > local=/192.168.20.30:50010 remote=/192.168.20.30:36188] > > at > > > > > org.apache.hadoop.net.SocketIOWithTimeout.waitForIO(SocketIOWithTimeout.java:246) > > at > > > > > org.apache.hadoop.net.SocketOutputStream.waitForWritable(SocketOutputStream.java:159) > > at > > > > > org.apache.hadoop.net.SocketOutputStream.transferToFully(SocketOutputStream.java:198) > > at > > > > > org.apache.hadoop.hdfs.server.datanode.BlockSender.sendChunks(BlockSender.java:392) > > at > > > > > org.apache.hadoop.hdfs.server.datanode.BlockSender.sendBlock(BlockSender.java:490) > > at > > > > > org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:202) > > at > > > > > org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:104) > > at java.lang.Thread.run(Thread.java:662) > > > > > > > > > > > > > > > > -- > > Thanks and Regards, > > Vimal Jain > > > -- Thanks and Regards, Vimal Jain