This is because the HDFS cluster is not started successfully. You can check the 
namenode/datanode logs for further information.

Regards,
Jingcheng

-----Original Message-----
From: ankit beohar [mailto:ankitbeoha...@gmail.com] 
Sent: Friday, July 29, 2016 1:06 PM
To: d...@hbase.apache.org
Cc: user@hbase.apache.org
Subject: Re: Hbase USERT

Hi All,

We will monitor memory usages meanwhile below are the error:-

org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
/hbase/WALs/quickstart.cloudera,60020,1469713295051/quickstart.cloudera%2C60020%2C1469713295051..meta.1469766981975.meta
could only be replicated to 0 nodes instead of minReplication (=1).  There are 
0 datanode(s) running and no node(s) are excluded in this operation.

        at
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1595)

        at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3287)

        at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:677)

        at
org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.addBlock(AuthorizationProviderProxyClientProtocol.java:213)

        at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:485)

        at
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)

        at
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617)

        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073)

        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086)

        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082)

        at java.security.AccessController.doPrivileged(Native Method)

        at javax.security.auth.Subject.doAs(Subject.java:415)

        at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1693)

        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080)



        at org.apache.hadoop.ipc.Client.call(Client.java:1471)

        at org.apache.hadoop.ipc.Client.call(Client.java:1408)

        at
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230)

        at com.sun.proxy.$Proxy22.addBlock(Unknown Source)

        at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:404)

        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

        at java.lang.reflect.Method.invoke(Method.java:606)

        at
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:256)

        at
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104)

        at com.sun.proxy.$Proxy23.addBlock(Unknown Source)

        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

        at java.lang.reflect.Method.invoke(Method.java:606)

        at
org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:279)

        at com.sun.proxy.$Proxy24.addBlock(Unknown Source)

        at
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1704)

        at
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1500)

        at
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:668)

2016-07-28 21:37:49,143 WARN
org.apache.hadoop.hbase.regionserver.wal.ProtobufLogWriter: Failed to write 
trailer, non-fatal, continuing...

org.apache.hadoop.ipc.RemoteException(java.io.IOException): File
/hbase/WALs/quickstart.cloudera,60020,1469713295051/quickstart.cloudera%2C60020%2C1469713295051.null0.1469766981975
could only be replicated to 0 nodes instead of minReplication (=1).  There are 
0 datanode(s) running and no node(s) are excluded in this operation.

        at
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1595)

        at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3287)

        at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:677)

        at
org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.addBlock(AuthorizationProviderProxyClientProtocol.java:213)

        at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:485)




Best Regards,
ANKIT BEOHAR


On Fri, Jul 29, 2016 at 9:10 AM, Du, Jingcheng <jingcheng...@intel.com>
wrote:

> Hi Ankit,
>
> It seems like a memory issue in region servers. Did you monitor the 
> memory usage in region servers during the run? How about to increase the heap 
> size?
> Do you get exceptions after the region server goes down? Mind sharing 
> it here?
>
> Regards,
> Jingcheng
>
> -----Original Message-----
> From: Dima Spivak [mailto:dspi...@cloudera.com]
> Sent: Friday, July 29, 2016 11:32 AM
> To: user@hbase.apache.org
> Subject: Re: Hbase USERT
>
> Hey Ankit,
>
> Moving the dev list to bcc and adding the user mailing list as the 
> recipient. Maybe a fellow user can offer some suggestions.
>
> All the best,
>   Dima
>
> On Thursday, July 28, 2016, ankit beohar <ankitbeoha...@gmail.com> wrote:
>
> > Hi Hbase,
> >
> > My use case is :- I am getting files and I want to insert the 
> > records in hbase with rowkey if rowkey available I have to update 
> > the values with
> > old+new values.
> >
> > For this I wrote MR job and get the values of each rowkey and in If 
> > else I manage my update and insert but with only 0.1 millions 
> > records hbase region server goes down.
> >
> > Any idea on this?
> >
> > I tried to incorporate Phoenix upsert also but with this same error
> occurs.
> >
> > Please help me out this.
> >
> > Best Regards,
> > ANKIT BEOHAR
> >
>

Reply via email to