I will try that if i get them next time.
Could anyone please give the cause of this exceptions ?


On Fri, Oct 18, 2013 at 4:03 PM, divye sheth <divs.sh...@gmail.com> wrote:

> I would recommend you to stop the cluster and then start the daemons one by
> one.
> 1. stop-dfs.sh
> 2. hadoop-daemon.sh start namenode
> 3. hadoop-daemon.sh start datanode
>
> This will show start up errors if any, also verify if the datanode is able
> to communicate with the namenode.
>
> Thanks
> Divye Sheth
>
>
> On Fri, Oct 18, 2013 at 3:51 PM, Vimal Jain <vkj...@gmail.com> wrote:
>
> > Hi,
> > I am running Hbase in pseudo distributed mode.( Hbase 0.94.7 and Hadoop
> > 1.1.2).
> > I am getting certain exceptions in Hadoop's namenode and data node files
> > which are :-
> >
> > Namenode :-
> >
> > 2013-10-18 10:33:37,218 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
> > NameSystem.heartbeatCheck: lost heartbeat from 192.168.20.30:50010
> > 2013-10-18 10:33:37,242 INFO org.apache.hadoop.net.NetworkTopology:
> > Removing a node: /default-rack/192.168.20.30:50010
> > 2013-10-18 10:35:27,606 INFO
> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
> > transactions: 64 Total time for transactions(ms): 1Number
> > of transactions batched in Syncs: 0 Number of syncs: 43 SyncTimes(ms): 86
> > 2013-10-18 10:35:27,614 ERROR
> > org.apache.hadoop.security.UserGroupInformation:
> PriviledgedActionException
> > as:hadoop cause:java.io.IOException: File /h
> >
> >
> base/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e
> > could only be replicated to 0 nodes, instead of 1
> > 2013-10-18 10:35:27,895 INFO org.apache.hadoop.ipc.Server: IPC Server
> > handler 9 on 9000, call
> > addBlock(/hbase/event_data/433b61f2a4ebff8f2e4b89890508a
> > 3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e,
> > DFSClient_hb_rs_hbase.rummycircle.com,60020,1382012725057, null) from
> > 192.168.20.30:44990: error: java.io.I
> > OException: File
> >
> >
> /hbase/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e
> > could only be replicated to 0 nodes, instead
> >  of 1
> > java.io.IOException: File
> >
> >
> /hbase/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e
> > could only be replicated to 0 nodes
> > , instead of 1
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1639)
> >     at
> >
> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:736)
> >     at sun.reflect.GeneratedMethodAccessor10.invoke(Unknown Source)
> >     at
> >
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> >     at java.lang.reflect.Method.invoke(Method.java:597)
> >     at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:578)
> >     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1393)
> >     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1389)
> >     at java.security.AccessController.doPrivileged(Native Method)
> >     at javax.security.auth.Subject.doAs(Subject.java:396)
> >     at
> >
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1149)
> >     at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1387)
> >
> >
> > Data node :-
> >
> > 2013-10-18 06:13:14,499 WARN
> > org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
> > 192.168.20.30:50010, storageID=DS-1816106352-192.16
> > 8.20.30-50010-1369314076237, infoPort=50075, ipcPort=50020):Got exception
> > while serving blk_-3215981820534544354_52215 to /192.168.20.30:
> > java.net.SocketTimeoutException: 480000 millis timeout while waiting for
> > channel to be ready for write. ch :
> > java.nio.channels.SocketChannel[connected
> >  local=/192.168.20.30:50010 remote=/192.168.20.30:36188]
> >     at
> >
> >
> org.apache.hadoop.net.SocketIOWithTimeout.waitForIO(SocketIOWithTimeout.java:246)
> >     at
> >
> >
> org.apache.hadoop.net.SocketOutputStream.waitForWritable(SocketOutputStream.java:159)
> >     at
> >
> >
> org.apache.hadoop.net.SocketOutputStream.transferToFully(SocketOutputStream.java:198)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockSender.sendChunks(BlockSender.java:392)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockSender.sendBlock(BlockSender.java:490)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:202)
> >     at
> >
> >
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:104)
> >     at java.lang.Thread.run(Thread.java:662)
> >
> >
> >
> >
> >
> >
> >
> > --
> > Thanks and Regards,
> > Vimal Jain
> >
>



-- 
Thanks and Regards,
Vimal Jain

Reply via email to