Hbase client and server are in same version 0.98.7. We are having complete
downtime of about 30min and high cpu usage in the node and network in the
cluster.

On Sat, Aug 6, 2016 at 11:38 AM, kiran <kiran.sarvabho...@gmail.com> wrote:

> We are also facing the same issue. Please tell us what is the solution. I
> have increased the rpc timeout and caching is reduced but with no effect.
> We are using hbase 0.98.7. Please suggest a work around as we are facing
> the issue very frequently now and we are having downtime in production.
>
> On Fri, Oct 30, 2015 at 9:21 AM, 聪聪 <175998...@qq.com> wrote:
>
>> There is a view child nodes loop code  http://paste2.org/zm8GE7xH
>>
>>
>>
>>
>> ------------------ 原始邮件 ------------------
>> 发件人: "蒲聪-北京";<175998...@qq.com>;
>> 发送时间: 2015年10月30日(星期五) 上午10:24
>> 收件人: "user"<user@hbase.apache.org>;
>>
>> 主题: 回复: 回复: Hbase cluster is suddenly unable to respond
>>
>>
>>
>>
>>
>> The client code is http://paste2.org/p3BXkKtV
>>
>>
>> Is the client version compatible with it?
>> I see  that the client version is hbase0.96.1.1-hadoop2
>>
>>
>>
>>
>> ------------------ 原始邮件 ------------------
>> 发件人: "Ted Yu";<yuzhih...@gmail.com>;
>> 发送时间: 2015年10月30日(星期五) 凌晨0:08
>> 收件人: "user@hbase.apache.org"<user@hbase.apache.org>;
>>
>> 主题: Re: 回复: Hbase cluster is suddenly unable to respond
>>
>>
>>
>> Client side, have they tried increasing direct memory size ?
>> -XX:MaxDirectMemorySize=
>>
>> Do you know how wide the rows returned may get ?
>>
>> Cheers
>>
>> On Thu, Oct 29, 2015 at 9:03 AM, 聪聪 <175998...@qq.com> wrote:
>>
>> > Developers feedback their client has the following error:
>> >
>> > [2015/10/29 19:20:42.260][WARN][RpcClient:724] IPC Client (1904394969)
>> > connection to l-hbase28.data.cn8.qunar.com/192.168.44.32:60020 from
>> > tomcat: unexpected exception receiving call responses
>> >
>> > java.lang.OutOfMemoryError: Direct buffer memory
>> >
>> >     at java.nio.Bits.reserveMemory(Bits.java:633) ~[na:1.6.0_20]
>> >
>> >     at java.nio.DirectByteBuffer.<init>(DirectByteBuffer.java:95)
>> > ~[na:1.6.0_20]
>> >
>> >     at java.nio.ByteBuffer.allocateDirect(ByteBuffer.java:288)
>> > ~[na:1.6.0_20]
>> >
>> >     at sun.nio.ch.Util.getTemporaryDirectBuffer(Util.java:57)
>> > ~[na:1.6.0_20]
>> >
>> >     at sun.nio.ch.IOUtil.read(IOUtil.java:205) ~[na:1.6.0_20]
>> >
>> >     at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:236)
>> > ~[na:1.6.0_20]
>> >
>> >     at
>> > org.apache.hadoop.net.SocketInputStream$Reader.performIO(
>> SocketInputStream.java:57)
>> > ~[hadoop-common-2.2.0.jar:na]
>> >
>> >     at
>> > org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithT
>> imeout.java:142)
>> > ~[hadoop-common-2.2.0.jar:na]
>> >
>> >     at
>> > org.apache.hadoop.net.SocketInputStream.read(SocketInputStre
>> am.java:161)
>> > ~[hadoop-common-2.2.0.jar:na]
>> >
>> >     at
>> > org.apache.hadoop.net.SocketInputStream.read(SocketInputStre
>> am.java:131)
>> > ~[hadoop-common-2.2.0.jar:na]
>> >
>> >     at java.io.FilterInputStream.read(FilterInputStream.java:116)
>> > ~[na:1.6.0_20]
>> >
>> >     at java.io.FilterInputStream.read(FilterInputStream.java:116)
>> > ~[na:1.6.0_20]
>> >
>> >     at
>> > org.apache.hadoop.hbase.ipc.RpcClient$Connection$PingInputSt
>> ream.read(RpcClient.java:555)
>> > ~[hbase-client-0.96.1.1-hadoop2.jar:0.96.1.1-hadoop2]
>> >
>> >     at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
>> > ~[na:1.6.0_20]
>> >
>> >     at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>> > ~[na:1.6.0_20]
>> >
>> >     at java.io.DataInputStream.read(DataInputStream.java:132)
>> > ~[na:1.6.0_20]
>> >
>> >     at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:192)
>> > ~[hadoop-common-2.2.0.jar:na]
>> >
>> >     at
>> > org.apache.hadoop.hbase.ipc.RpcClient$Connection.readRespons
>> e(RpcClient.java:1101)
>> > ~[hbase-client-0.96.1.1-hadoop2.jar:0.96.1.1-hadoop2]
>> >
>> >     at
>> > org.apache.hadoop.hbase.ipc.RpcClient$Connection.run(RpcClie
>> nt.java:721)
>> > ~[hbase-client-0.96.1.1-hadoop2.jar:0.96.1.1-hadoop2]
>> >
>> >
>> >
>> >
>> >
>> >
>> >
>> > ------------------ 原始邮件 ------------------
>> > 发件人: "Ted Yu";<yuzhih...@gmail.com>;
>> > 发送时间: 2015年10月29日(星期四) 晚上10:48
>> > 收件人: "user@hbase.apache.org"<user@hbase.apache.org>;
>> >
>> > 主题: Re: 回复: Hbase cluster is suddenly unable to respond
>> >
>> >
>> >
>> > I took a look at the jstack.
>> > The threads involving RpcServer$Connection.readAndProcess() were in
>> > RUNNABLE state, not BLOCKED or IN_NATIVE state - as described in
>> > HBASE-11277
>> > .
>> >
>> > The protobuf exception shown in your earlier email corresponded to the
>> > following in hbase-protocol/src/main/protobuf/Client.proto :
>> >
>> > message GetRequest {
>> >   required RegionSpecifier region = 1;
>> >   required Get get = 2;
>> > }
>> >
>> > Are all your hbase clients running in the same version ?
>> >
>> > Cheers
>> >
>> > On Thu, Oct 29, 2015 at 7:28 AM, 聪聪 <175998...@qq.com> wrote:
>> >
>> > > the regionserver jstack log is    http://paste2.org/yLDJeXgL
>> > >
>> > >
>> > >
>> > >
>> > > ------------------ 原始邮件 ------------------
>> > > 发件人: "蒲聪-北京";<175998...@qq.com>;
>> > > 发送时间: 2015年10月29日(星期四) 晚上9:10
>> > > 收件人: "user"<user@hbase.apache.org>;
>> > >
>> > > 主题: 回复: 回复: Hbase cluster is suddenly unable to respond
>> > >
>> > >
>> > >
>> > > hi Ted:
>> > >
>> > >
>> > > Yesterday around 14:40,one of regionservers hang once against.At that
>> > time
>> > > I saw web ui can not open.Hbase cluster is  unable to respond.Very
>> > anxious,
>> > > hoping to get help!
>> > >
>> > >
>> > > jstack log is as follows:
>> > > "RpcServer.reader=4,port=60020" daemon prio=10 tid=0x00007f4466146800
>> > > nid=0x12d3 runnable [0x00007f3bebe58000]
>> > >    java.lang.Thread.State: RUNNABLE
>> > >     at sun.nio.ch.NativeThread.current(Native Method)
>> > >     at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:325)
>> > >     - locked <0x00007f3d27360fb0> (a java.lang.Object)
>> > >     - locked <0x00007f3d27360f90> (a java.lang.Object)
>> > >     at
>> > org.apache.hadoop.hbase.ipc.RpcServer.channelIO(RpcServer.java:2402)
>> > >     at
>> > > org.apache.hadoop.hbase.ipc.RpcServer.channelRead(RpcServer.
>> java:2368)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc
>> ess(RpcServer.java:1476)
>> > >     at
>> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer
>> ver.java:770)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL
>> oop(RpcServer.java:563)
>> > >     - locked <0x00007f3c584ce990> (a
>> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp
>> cServer.java:538)
>> > >     at
>> > >
>> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>> Executor.java:1145)
>> > >     at
>> > >
>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>> lExecutor.java:615)
>> > >     at java.lang.Thread.run(Thread.java:744)
>> > >
>> > >
>> > > "RpcServer.reader=3,port=60020" daemon prio=10 tid=0x00007f4466145000
>> > > nid=0x12d2 runnable [0x00007f3bebf59000]
>> > >    java.lang.Thread.State: RUNNABLE
>> > >     at sun.nio.ch.NativeThread.current(Native Method)
>> > >     at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:325)
>> > >     - locked <0x00007f3d27360530> (a java.lang.Object)
>> > >     - locked <0x00007f3d27360510> (a java.lang.Object)
>> > >     at
>> > org.apache.hadoop.hbase.ipc.RpcServer.channelIO(RpcServer.java:2402)
>> > >     at
>> > > org.apache.hadoop.hbase.ipc.RpcServer.channelRead(RpcServer.
>> java:2368)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc
>> ess(RpcServer.java:1476)
>> > >     at
>> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer
>> ver.java:770)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL
>> oop(RpcServer.java:563)
>> > >     - locked <0x00007f3c584cf7d8> (a
>> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp
>> cServer.java:538)
>> > >     at
>> > >
>> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>> Executor.java:1145)
>> > >     at
>> > >
>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>> lExecutor.java:615)
>> > >     at java.lang.Thread.run(Thread.java:744)
>> > >
>> > >
>> > >
>> > >
>> > >
>> > > region server log :
>> > > 2015-10-28 14:38:19,801 INFO  [Thread-15] regionserver.HRegion:
>> Finished
>> > > memstore flush of ~3.6 M/3820648, currentsize=536/536 for region
>> > >
>> > order_history,2801xyz140618175732642$3,1418829598639.afc8534
>> 71a8cd4184bc9e7be00b8eea0.
>> > > in 45ms, sequenceid=9599960557, compaction requested=true
>> > > 2015-10-28 14:38:32,693 ERROR [RpcServer.handler=3,port=60020]
>> > > regionserver.HRegionServer:
>> > > org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException:
>> > > Expected nextCallSeq: 1 But the nextCallSeq got from client: 0;
>> > > request=scanner_id: 16740356019163164014 number_of_rows: 10
>> > close_scanner:
>> > > false next_call_seq: 0
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.regionserver.HRegionServer.scan(
>> HRegionServer.java:3007)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$
>> ClientService$2.callBlockingMethod(ClientProtos.java:26929)
>> > >     at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:21
>> 46)
>> > >     at
>> > > org.apache.hadoop.hbase.ipc.RpcServer$Handler.run(RpcServer.
>> java:1851)
>> > > 2015-10-28 14:38:32,693 ERROR [RpcServer.handler=9,port=60020]
>> > > regionserver.HRegionServer:
>> > > org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException:
>> > > Expected nextCallSeq: 1 But the nextCallSeq got from client: 0;
>> > > request=scanner_id: 16740356019163164014 number_of_rows: 10
>> > close_scanner:
>> > > false next_call_seq: 0
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.regionserver.HRegionServer.scan(
>> HRegionServer.java:3007)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$
>> ClientService$2.callBlockingMethod(ClientProtos.java:26929)
>> > >     at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:21
>> 46)
>> > >     at
>> > > org.apache.hadoop.hbase.ipc.RpcServer$Handler.run(RpcServer.
>> java:1851)
>> > >
>> > > 2015-10-28 14:38:32,696 WARN  [RpcServer.reader=2,port=60020]
>> > > ipc.RpcServer: Unable to read call parameter from client
>> 192.168.37.135
>> > > com.google.protobuf.UninitializedMessageException: Message missing
>> > > required fields: region, get
>> > >     at
>> > >
>> > com.google.protobuf.AbstractMessage$Builder.newUninitialized
>> MessageException(AbstractMessage.java:770)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$
>> GetRequest$Builder.build(ClientProtos.java:4474)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$
>> GetRequest$Builder.build(ClientProtos.java:4406)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.processRequ
>> est(RpcServer.java:1689)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.processOneR
>> pc(RpcServer.java:1631)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc
>> ess(RpcServer.java:1491)
>> > >     at
>> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer
>> ver.java:770)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL
>> oop(RpcServer.java:563)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp
>> cServer.java:538)
>> > >     at
>> > >
>> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>> Executor.java:1145)
>> > >     at
>> > >
>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>> lExecutor.java:615)
>> > >     at java.lang.Thread.run(Thread.java:744)
>> > > 2015-10-28 14:38:32,696 WARN  [RpcServer.reader=1,port=60020]
>> > > ipc.RpcServer: Unable to read call parameter from client
>> 192.168.37.135
>> > > com.google.protobuf.UninitializedMessageException: Message missing
>> > > required fields: region, get
>> > >     at
>> > >
>> > com.google.protobuf.AbstractMessage$Builder.newUninitialized
>> MessageException(AbstractMessage.java:770)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$
>> GetRequest$Builder.build(ClientProtos.java:4474)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$
>> GetRequest$Builder.build(ClientProtos.java:4406)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.processRequ
>> est(RpcServer.java:1689)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.processOneR
>> pc(RpcServer.java:1631)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc
>> ess(RpcServer.java:1491)
>> > >     at
>> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer
>> ver.java:770)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL
>> oop(RpcServer.java:563)
>> > >     at
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp
>> cServer.java:538)
>> > >     at
>> > >
>> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>> Executor.java:1145)
>> > >     at
>> > >
>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>> lExecutor.java:615)
>> > >     at java.lang.Thread.run(Thread.java:744)
>> > >
>> > >
>> > >
>> > > ------------------ 原始邮件 ------------------
>> > > 发件人: "蒲聪-北京";<175998...@qq.com>;
>> > > 发送时间: 2015年10月26日(星期一) 晚上9:28
>> > > 收件人: "user"<user@hbase.apache.org>;
>> > >
>> > > 主题: 回复: 回复: Hbase cluster is suddenly unable to respond
>> > >
>> > >
>> > >
>> > >
>> > >
>> > > Thank you very much!
>> > >
>> > >
>> > > ------------------ 原始邮件 ------------------
>> > > 发件人: "Ted Yu";<yuzhih...@gmail.com>;
>> > > 发送时间: 2015年10月26日(星期一) 晚上8:28
>> > > 收件人: "user"<user@hbase.apache.org>;
>> > >
>> > > 主题: Re: 回复: Hbase cluster is suddenly unable to respond
>> > >
>> > >
>> > >
>> > > The fix from HBASE-11277 may solve your problem - if you collect stack
>> > > trace during the hang, we would have more clue.
>> > >
>> > > I suggest upgrading to newer release such as 1.1.2 or 0.98.15
>> > >
>> > > Cheers
>> > >
>> > > > On Oct 26, 2015, at 12:42 AM, 聪聪 <175998...@qq.com> wrote:
>> > > >
>> > > > hi,Ted:
>> > > >
>> > > >
>> > > > I use the HBase version is hbase-0.96.0.
>> > > > Around 17:33,other region servers also appeared in this warn log.I
>> > don't
>> > > know if it's normal or not.At that time I saw web ui can not open.I
>> > restart
>> > > the regionserver  then hbase back to normal. Is it possible  bug
>> > > HBASE-11277?
>> > > >
>> > > >
>> > > > Regionserver on the log basically almost  this warn log
>> > > > mater on the log  is as follows:
>> > > > 2015-10-21 22:15:43,575 INFO  [CatalogJanitor-l-namenode2:60000]
>> > > master.CatalogJanitor: Scanned 672 catalog row(s), gc'd 0 unreferenced
>> > > merged region(s) and 1 unreferenced parent region(s)
>> > > > 2015-10-23 17:47:25,617 INFO  [RpcServer.handler=28,port=60000]
>> > > master.HMaster: Client=hbase//192.168.39.19 set balanceSwitch=false
>> > > > 2015-10-23 17:49:45,513 WARN  [RpcServer.handler=24,port=60000]
>> > > ipc.RpcServer: (responseTooSlow):
>> > >
>> > {"processingtimems":70266,"call":"ListTableDescriptorsByName
>> space(org.apache.hadoop.hbase.protobuf.generated.MasterProto
>> s$ListTableDescriptorsByNamespaceRequest)","client":"
>> > > 192.168.39.22:60292
>> > >
>> > ","starttimems":1445593715207,"queuetimems":0,"class":"HMast
>> er","responsesize":704,"method":"ListTableDescriptorsByNamespace"}
>> > > > 2015-10-23 17:49:45,513 WARN  [RpcServer.handler=6,port=60000]
>> > > ipc.RpcServer: (responseTooSlow):
>> > >
>> > {"processingtimems":130525,"call":"ListTableDescriptorsByNam
>> espace(org.apache.hadoop.hbase.protobuf.generated.Maste
>> rProtos$ListTableDescriptorsByNamespaceRequest)","client":"
>> > > 192.168.39.22:60286
>> > >
>> > ","starttimems":1445593654945,"queuetimems":0,"class":"HMast
>> er","responsesize":704,"method":"ListTableDescriptorsByNamespace"}
>> > > > 2015-10-23 17:49:45,513 WARN  [RpcServer.handler=24,port=60000]
>> > > ipc.RpcServer: RpcServer.respondercallId: 130953 service:
>> MasterService
>> > > methodName: ListTableDescriptorsByNamespace size: 48 connection:
>> > > 192.168.39.22:60292: output error
>> > > > 2015-10-23 17:49:45,513 WARN  [RpcServer.handler=6,port=60000]
>> > > ipc.RpcServer: RpcServer.respondercallId: 130945 service:
>> MasterService
>> > > methodName: ListTableDescriptorsByNamespace size: 48 connection:
>> > > 192.168.39.22:60286: output error
>> > > > 2015-10-23 17:49:45,513 WARN  [RpcServer.handler=6,port=60000]
>> > > ipc.RpcServer: RpcServer.handler=6,port=60000: caught a
>> > > ClosedChannelException, this means that the server was processing a
>> > request
>> > > but the client went away. The error message was: null
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > > ------------------ 原始邮件 ------------------
>> > > > 发件人: "Ted Yu";<yuzhih...@gmail.com>;
>> > > > 发送时间: 2015年10月23日(星期五) 晚上11:39
>> > > > 收件人: "user@hbase.apache.org"<user@hbase.apache.org>;
>> > > >
>> > > > 主题: Re: Hbase cluster is suddenly unable to respond
>> > > >
>> > > >
>> > > >
>> > > > Were other region servers functioning normally around 17:33 ?
>> > > >
>> > > > Which hbase release are you using ?
>> > > >
>> > > > Can you pastebin more of the region server log ?
>> > > >
>> > > > Thanks
>> > > >
>> > > >> On Fri, Oct 23, 2015 at 8:28 AM, 聪聪 <175998...@qq.com> wrote:
>> > > >>
>> > > >> hi,all:
>> > > >>
>> > > >>
>> > > >> This afternoon,The whole Hbase cluster is suddenly unable to
>> > > respond.after
>> > > >> I restart a regionserver after,the cluster has recovered.I don't
>> know
>> > > the
>> > > >> cause of the trouble.I hope I can get help from you.
>> > > >>
>> > > >>
>> > > >> Regionserver on the log is as follows:
>> > > >> 2015-10-23 17:28:49,335 INFO  [regionserver60020.logRoller]
>> > wal.FSHLog:
>> > > >> moving old hlog file /hbase/WALs/l-hbase30.data.cn8.qunar.com
>> > > >> ,60020,1442810406218/l-hbase30.data.cn8.qunar.com
>> > > %2C60020%2C1442810406218.1445580462689
>> > > >> whose highest sequenceid is 9071525521 to /hbase/oldWALs/
>> > > >> l-hbase30.data.cn8.qunar.com%2C60020%2C1442810406218.1445580462689
>> > > >> 2015-10-23 17:33:31,375 WARN  [RpcServer.reader=8,port=60020]
>> > > >> ipc.RpcServer: RpcServer.listener,port=60020: count of bytes read:
>> 0
>> > > >> java.io.IOException: Connection reset by peer
>> > > >>        at sun.nio.ch.FileDispatcherImpl.read0(Native Method)
>> > > >>        at sun.nio.ch.SocketDispatcher.re
>> ad(SocketDispatcher.java:39)
>> > > >>        at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)
>> > > >>        at sun.nio.ch.IOUtil.read(IOUtil.java:197)
>> > > >>        at
>> > sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:379)
>> > > >>        at
>> > > >> org.apache.hadoop.hbase.ipc.RpcServer.channelRead(RpcServer.
>> java:2368)
>> > > >>        at
>> > > >>
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc
>> ess(RpcServer.java:1403)
>> > > >>        at
>> > > >>
>> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer
>> ver.java:770)
>> > > >>        at
>> > > >>
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL
>> oop(RpcServer.java:563)
>> > > >>        at
>> > > >>
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp
>> cServer.java:538)
>> > > >>        at
>> > > >>
>> > >
>> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>> Executor.java:1145)
>> > > >>        at
>> > > >>
>> > >
>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>> lExecutor.java:615)
>> > > >>        at java.lang.Thread.run(Thread.java:744)
>> > > >> 2015-10-23 17:33:31,779 WARN  [RpcServer.reader=2,port=60020]
>> > > >> ipc.RpcServer: RpcServer.listener,port=60020: count of bytes read:
>> 0
>> > > >> java.io.IOException: Connection reset by peer
>> > > >>        at sun.nio.ch.FileDispatcherImpl.read0(Native Method)
>> > > >>        at sun.nio.ch.SocketDispatcher.re
>> ad(SocketDispatcher.java:39)
>> > > >>        at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)
>> > > >>        at sun.nio.ch.IOUtil.read(IOUtil.java:197)
>> > > >>        at
>> > sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:379)
>> > > >>        at
>> > > >> org.apache.hadoop.hbase.ipc.RpcServer.channelRead(RpcServer.
>> java:2368)
>> > > >>        at
>> > > >>
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc
>> ess(RpcServer.java:1403)
>> > > >>        at
>> > > >>
>> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer
>> ver.java:770)
>> > > >>        at
>> > > >>
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL
>> oop(RpcServer.java:563)
>> > > >>        at
>> > > >>
>> > >
>> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp
>> cServer.java:538)
>> > > >>        at
>> > > >>
>> > >
>> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>> Executor.java:1145)
>> > > >>        at
>> > > >>
>> > >
>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>> lExecutor.java:615)
>> > > >>        at java.lang.Thread.run(Thread.java:744)
>> > >
>> >
>>
>
>
>
> --
> Thank you
> Kiran Sarvabhotla
>
> -----Even a correct decision is wrong when it is taken late
>
>


-- 
Thank you
Kiran Sarvabhotla

-----Even a correct decision is wrong when it is taken late

Reply via email to