Can you provide jstack of region server(s) ? Was there anything interesting in the logs ?
Thanks BTW 0.98.7 is quite old. Please consider upgrading. On Fri, Aug 5, 2016 at 11:10 PM, kiran <kiran.sarvabho...@gmail.com> wrote: > Hbase client and server are in same version 0.98.7. We are having complete > downtime of about 30min and high cpu usage in the node and network in the > cluster. > > On Sat, Aug 6, 2016 at 11:38 AM, kiran <kiran.sarvabho...@gmail.com> > wrote: > > > We are also facing the same issue. Please tell us what is the solution. I > > have increased the rpc timeout and caching is reduced but with no effect. > > We are using hbase 0.98.7. Please suggest a work around as we are facing > > the issue very frequently now and we are having downtime in production. > > > > On Fri, Oct 30, 2015 at 9:21 AM, 聪聪 <175998...@qq.com> wrote: > > > >> There is a view child nodes loop code http://paste2.org/zm8GE7xH > >> > >> > >> > >> > >> ------------------ 原始邮件 ------------------ > >> 发件人: "蒲聪-北京";<175998...@qq.com>; > >> 发送时间: 2015年10月30日(星期五) 上午10:24 > >> 收件人: "user"<user@hbase.apache.org>; > >> > >> 主题: 回复: 回复: Hbase cluster is suddenly unable to respond > >> > >> > >> > >> > >> > >> The client code is http://paste2.org/p3BXkKtV > >> > >> > >> Is the client version compatible with it? > >> I see that the client version is hbase0.96.1.1-hadoop2 > >> > >> > >> > >> > >> ------------------ 原始邮件 ------------------ > >> 发件人: "Ted Yu";<yuzhih...@gmail.com>; > >> 发送时间: 2015年10月30日(星期五) 凌晨0:08 > >> 收件人: "user@hbase.apache.org"<user@hbase.apache.org>; > >> > >> 主题: Re: 回复: Hbase cluster is suddenly unable to respond > >> > >> > >> > >> Client side, have they tried increasing direct memory size ? > >> -XX:MaxDirectMemorySize= > >> > >> Do you know how wide the rows returned may get ? > >> > >> Cheers > >> > >> On Thu, Oct 29, 2015 at 9:03 AM, 聪聪 <175998...@qq.com> wrote: > >> > >> > Developers feedback their client has the following error: > >> > > >> > [2015/10/29 19:20:42.260][WARN][RpcClient:724] IPC Client > (1904394969) > >> > connection to l-hbase28.data.cn8.qunar.com/192.168.44.32:60020 from > >> > tomcat: unexpected exception receiving call responses > >> > > >> > java.lang.OutOfMemoryError: Direct buffer memory > >> > > >> > at java.nio.Bits.reserveMemory(Bits.java:633) ~[na:1.6.0_20] > >> > > >> > at java.nio.DirectByteBuffer.<init>(DirectByteBuffer.java:95) > >> > ~[na:1.6.0_20] > >> > > >> > at java.nio.ByteBuffer.allocateDirect(ByteBuffer.java:288) > >> > ~[na:1.6.0_20] > >> > > >> > at sun.nio.ch.Util.getTemporaryDirectBuffer(Util.java:57) > >> > ~[na:1.6.0_20] > >> > > >> > at sun.nio.ch.IOUtil.read(IOUtil.java:205) ~[na:1.6.0_20] > >> > > >> > at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:236) > >> > ~[na:1.6.0_20] > >> > > >> > at > >> > org.apache.hadoop.net.SocketInputStream$Reader.performIO( > >> SocketInputStream.java:57) > >> > ~[hadoop-common-2.2.0.jar:na] > >> > > >> > at > >> > org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithT > >> imeout.java:142) > >> > ~[hadoop-common-2.2.0.jar:na] > >> > > >> > at > >> > org.apache.hadoop.net.SocketInputStream.read(SocketInputStre > >> am.java:161) > >> > ~[hadoop-common-2.2.0.jar:na] > >> > > >> > at > >> > org.apache.hadoop.net.SocketInputStream.read(SocketInputStre > >> am.java:131) > >> > ~[hadoop-common-2.2.0.jar:na] > >> > > >> > at java.io.FilterInputStream.read(FilterInputStream.java:116) > >> > ~[na:1.6.0_20] > >> > > >> > at java.io.FilterInputStream.read(FilterInputStream.java:116) > >> > ~[na:1.6.0_20] > >> > > >> > at > >> > org.apache.hadoop.hbase.ipc.RpcClient$Connection$PingInputSt > >> ream.read(RpcClient.java:555) > >> > ~[hbase-client-0.96.1.1-hadoop2.jar:0.96.1.1-hadoop2] > >> > > >> > at java.io.BufferedInputStream.read1(BufferedInputStream. > java:256) > >> > ~[na:1.6.0_20] > >> > > >> > at java.io.BufferedInputStream.read(BufferedInputStream.java:317) > >> > ~[na:1.6.0_20] > >> > > >> > at java.io.DataInputStream.read(DataInputStream.java:132) > >> > ~[na:1.6.0_20] > >> > > >> > at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:192) > >> > ~[hadoop-common-2.2.0.jar:na] > >> > > >> > at > >> > org.apache.hadoop.hbase.ipc.RpcClient$Connection.readRespons > >> e(RpcClient.java:1101) > >> > ~[hbase-client-0.96.1.1-hadoop2.jar:0.96.1.1-hadoop2] > >> > > >> > at > >> > org.apache.hadoop.hbase.ipc.RpcClient$Connection.run(RpcClie > >> nt.java:721) > >> > ~[hbase-client-0.96.1.1-hadoop2.jar:0.96.1.1-hadoop2] > >> > > >> > > >> > > >> > > >> > > >> > > >> > > >> > ------------------ 原始邮件 ------------------ > >> > 发件人: "Ted Yu";<yuzhih...@gmail.com>; > >> > 发送时间: 2015年10月29日(星期四) 晚上10:48 > >> > 收件人: "user@hbase.apache.org"<user@hbase.apache.org>; > >> > > >> > 主题: Re: 回复: Hbase cluster is suddenly unable to respond > >> > > >> > > >> > > >> > I took a look at the jstack. > >> > The threads involving RpcServer$Connection.readAndProcess() were in > >> > RUNNABLE state, not BLOCKED or IN_NATIVE state - as described in > >> > HBASE-11277 > >> > . > >> > > >> > The protobuf exception shown in your earlier email corresponded to the > >> > following in hbase-protocol/src/main/protobuf/Client.proto : > >> > > >> > message GetRequest { > >> > required RegionSpecifier region = 1; > >> > required Get get = 2; > >> > } > >> > > >> > Are all your hbase clients running in the same version ? > >> > > >> > Cheers > >> > > >> > On Thu, Oct 29, 2015 at 7:28 AM, 聪聪 <175998...@qq.com> wrote: > >> > > >> > > the regionserver jstack log is http://paste2.org/yLDJeXgL > >> > > > >> > > > >> > > > >> > > > >> > > ------------------ 原始邮件 ------------------ > >> > > 发件人: "蒲聪-北京";<175998...@qq.com>; > >> > > 发送时间: 2015年10月29日(星期四) 晚上9:10 > >> > > 收件人: "user"<user@hbase.apache.org>; > >> > > > >> > > 主题: 回复: 回复: Hbase cluster is suddenly unable to respond > >> > > > >> > > > >> > > > >> > > hi Ted: > >> > > > >> > > > >> > > Yesterday around 14:40,one of regionservers hang once against.At > that > >> > time > >> > > I saw web ui can not open.Hbase cluster is unable to respond.Very > >> > anxious, > >> > > hoping to get help! > >> > > > >> > > > >> > > jstack log is as follows: > >> > > "RpcServer.reader=4,port=60020" daemon prio=10 > tid=0x00007f4466146800 > >> > > nid=0x12d3 runnable [0x00007f3bebe58000] > >> > > java.lang.Thread.State: RUNNABLE > >> > > at sun.nio.ch.NativeThread.current(Native Method) > >> > > at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java: > 325) > >> > > - locked <0x00007f3d27360fb0> (a java.lang.Object) > >> > > - locked <0x00007f3d27360f90> (a java.lang.Object) > >> > > at > >> > org.apache.hadoop.hbase.ipc.RpcServer.channelIO(RpcServer.java:2402) > >> > > at > >> > > org.apache.hadoop.hbase.ipc.RpcServer.channelRead(RpcServer. > >> java:2368) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc > >> ess(RpcServer.java:1476) > >> > > at > >> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer > >> ver.java:770) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL > >> oop(RpcServer.java:563) > >> > > - locked <0x00007f3c584ce990> (a > >> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp > >> cServer.java:538) > >> > > at > >> > > > >> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool > >> Executor.java:1145) > >> > > at > >> > > > >> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo > >> lExecutor.java:615) > >> > > at java.lang.Thread.run(Thread.java:744) > >> > > > >> > > > >> > > "RpcServer.reader=3,port=60020" daemon prio=10 > tid=0x00007f4466145000 > >> > > nid=0x12d2 runnable [0x00007f3bebf59000] > >> > > java.lang.Thread.State: RUNNABLE > >> > > at sun.nio.ch.NativeThread.current(Native Method) > >> > > at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java: > 325) > >> > > - locked <0x00007f3d27360530> (a java.lang.Object) > >> > > - locked <0x00007f3d27360510> (a java.lang.Object) > >> > > at > >> > org.apache.hadoop.hbase.ipc.RpcServer.channelIO(RpcServer.java:2402) > >> > > at > >> > > org.apache.hadoop.hbase.ipc.RpcServer.channelRead(RpcServer. > >> java:2368) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc > >> ess(RpcServer.java:1476) > >> > > at > >> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer > >> ver.java:770) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL > >> oop(RpcServer.java:563) > >> > > - locked <0x00007f3c584cf7d8> (a > >> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp > >> cServer.java:538) > >> > > at > >> > > > >> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool > >> Executor.java:1145) > >> > > at > >> > > > >> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo > >> lExecutor.java:615) > >> > > at java.lang.Thread.run(Thread.java:744) > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > region server log : > >> > > 2015-10-28 14:38:19,801 INFO [Thread-15] regionserver.HRegion: > >> Finished > >> > > memstore flush of ~3.6 M/3820648, currentsize=536/536 for region > >> > > > >> > order_history,2801xyz140618175732642$3,1418829598639.afc8534 > >> 71a8cd4184bc9e7be00b8eea0. > >> > > in 45ms, sequenceid=9599960557, compaction requested=true > >> > > 2015-10-28 14:38:32,693 ERROR [RpcServer.handler=3,port=60020] > >> > > regionserver.HRegionServer: > >> > > org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException: > >> > > Expected nextCallSeq: 1 But the nextCallSeq got from client: 0; > >> > > request=scanner_id: 16740356019163164014 number_of_rows: 10 > >> > close_scanner: > >> > > false next_call_seq: 0 > >> > > at > >> > > > >> > org.apache.hadoop.hbase.regionserver.HRegionServer.scan( > >> HRegionServer.java:3007) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ > >> ClientService$2.callBlockingMethod(ClientProtos.java:26929) > >> > > at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:21 > >> 46) > >> > > at > >> > > org.apache.hadoop.hbase.ipc.RpcServer$Handler.run(RpcServer. > >> java:1851) > >> > > 2015-10-28 14:38:32,693 ERROR [RpcServer.handler=9,port=60020] > >> > > regionserver.HRegionServer: > >> > > org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException: > >> > > Expected nextCallSeq: 1 But the nextCallSeq got from client: 0; > >> > > request=scanner_id: 16740356019163164014 number_of_rows: 10 > >> > close_scanner: > >> > > false next_call_seq: 0 > >> > > at > >> > > > >> > org.apache.hadoop.hbase.regionserver.HRegionServer.scan( > >> HRegionServer.java:3007) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ > >> ClientService$2.callBlockingMethod(ClientProtos.java:26929) > >> > > at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:21 > >> 46) > >> > > at > >> > > org.apache.hadoop.hbase.ipc.RpcServer$Handler.run(RpcServer. > >> java:1851) > >> > > > >> > > 2015-10-28 14:38:32,696 WARN [RpcServer.reader=2,port=60020] > >> > > ipc.RpcServer: Unable to read call parameter from client > >> 192.168.37.135 > >> > > com.google.protobuf.UninitializedMessageException: Message missing > >> > > required fields: region, get > >> > > at > >> > > > >> > com.google.protobuf.AbstractMessage$Builder.newUninitialized > >> MessageException(AbstractMessage.java:770) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ > >> GetRequest$Builder.build(ClientProtos.java:4474) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ > >> GetRequest$Builder.build(ClientProtos.java:4406) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.processRequ > >> est(RpcServer.java:1689) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.processOneR > >> pc(RpcServer.java:1631) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc > >> ess(RpcServer.java:1491) > >> > > at > >> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer > >> ver.java:770) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL > >> oop(RpcServer.java:563) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp > >> cServer.java:538) > >> > > at > >> > > > >> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool > >> Executor.java:1145) > >> > > at > >> > > > >> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo > >> lExecutor.java:615) > >> > > at java.lang.Thread.run(Thread.java:744) > >> > > 2015-10-28 14:38:32,696 WARN [RpcServer.reader=1,port=60020] > >> > > ipc.RpcServer: Unable to read call parameter from client > >> 192.168.37.135 > >> > > com.google.protobuf.UninitializedMessageException: Message missing > >> > > required fields: region, get > >> > > at > >> > > > >> > com.google.protobuf.AbstractMessage$Builder.newUninitialized > >> MessageException(AbstractMessage.java:770) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ > >> GetRequest$Builder.build(ClientProtos.java:4474) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ > >> GetRequest$Builder.build(ClientProtos.java:4406) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.processRequ > >> est(RpcServer.java:1689) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.processOneR > >> pc(RpcServer.java:1631) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc > >> ess(RpcServer.java:1491) > >> > > at > >> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer > >> ver.java:770) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL > >> oop(RpcServer.java:563) > >> > > at > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp > >> cServer.java:538) > >> > > at > >> > > > >> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool > >> Executor.java:1145) > >> > > at > >> > > > >> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo > >> lExecutor.java:615) > >> > > at java.lang.Thread.run(Thread.java:744) > >> > > > >> > > > >> > > > >> > > ------------------ 原始邮件 ------------------ > >> > > 发件人: "蒲聪-北京";<175998...@qq.com>; > >> > > 发送时间: 2015年10月26日(星期一) 晚上9:28 > >> > > 收件人: "user"<user@hbase.apache.org>; > >> > > > >> > > 主题: 回复: 回复: Hbase cluster is suddenly unable to respond > >> > > > >> > > > >> > > > >> > > > >> > > > >> > > Thank you very much! > >> > > > >> > > > >> > > ------------------ 原始邮件 ------------------ > >> > > 发件人: "Ted Yu";<yuzhih...@gmail.com>; > >> > > 发送时间: 2015年10月26日(星期一) 晚上8:28 > >> > > 收件人: "user"<user@hbase.apache.org>; > >> > > > >> > > 主题: Re: 回复: Hbase cluster is suddenly unable to respond > >> > > > >> > > > >> > > > >> > > The fix from HBASE-11277 may solve your problem - if you collect > stack > >> > > trace during the hang, we would have more clue. > >> > > > >> > > I suggest upgrading to newer release such as 1.1.2 or 0.98.15 > >> > > > >> > > Cheers > >> > > > >> > > > On Oct 26, 2015, at 12:42 AM, 聪聪 <175998...@qq.com> wrote: > >> > > > > >> > > > hi,Ted: > >> > > > > >> > > > > >> > > > I use the HBase version is hbase-0.96.0. > >> > > > Around 17:33,other region servers also appeared in this warn log.I > >> > don't > >> > > know if it's normal or not.At that time I saw web ui can not open.I > >> > restart > >> > > the regionserver then hbase back to normal. Is it possible bug > >> > > HBASE-11277? > >> > > > > >> > > > > >> > > > Regionserver on the log basically almost this warn log > >> > > > mater on the log is as follows: > >> > > > 2015-10-21 22:15:43,575 INFO [CatalogJanitor-l-namenode2:60000] > >> > > master.CatalogJanitor: Scanned 672 catalog row(s), gc'd 0 > unreferenced > >> > > merged region(s) and 1 unreferenced parent region(s) > >> > > > 2015-10-23 17:47:25,617 INFO [RpcServer.handler=28,port=60000] > >> > > master.HMaster: Client=hbase//192.168.39.19 set balanceSwitch=false > >> > > > 2015-10-23 17:49:45,513 WARN [RpcServer.handler=24,port=60000] > >> > > ipc.RpcServer: (responseTooSlow): > >> > > > >> > {"processingtimems":70266,"call":"ListTableDescriptorsByName > >> space(org.apache.hadoop.hbase.protobuf.generated.MasterProto > >> s$ListTableDescriptorsByNamespaceRequest)","client":" > >> > > 192.168.39.22:60292 > >> > > > >> > ","starttimems":1445593715207,"queuetimems":0,"class":"HMast > >> er","responsesize":704,"method":"ListTableDescriptorsByNamespace"} > >> > > > 2015-10-23 17:49:45,513 WARN [RpcServer.handler=6,port=60000] > >> > > ipc.RpcServer: (responseTooSlow): > >> > > > >> > {"processingtimems":130525,"call":"ListTableDescriptorsByNam > >> espace(org.apache.hadoop.hbase.protobuf.generated.Maste > >> rProtos$ListTableDescriptorsByNamespaceRequest)","client":" > >> > > 192.168.39.22:60286 > >> > > > >> > ","starttimems":1445593654945,"queuetimems":0,"class":"HMast > >> er","responsesize":704,"method":"ListTableDescriptorsByNamespace"} > >> > > > 2015-10-23 17:49:45,513 WARN [RpcServer.handler=24,port=60000] > >> > > ipc.RpcServer: RpcServer.respondercallId: 130953 service: > >> MasterService > >> > > methodName: ListTableDescriptorsByNamespace size: 48 connection: > >> > > 192.168.39.22:60292: output error > >> > > > 2015-10-23 17:49:45,513 WARN [RpcServer.handler=6,port=60000] > >> > > ipc.RpcServer: RpcServer.respondercallId: 130945 service: > >> MasterService > >> > > methodName: ListTableDescriptorsByNamespace size: 48 connection: > >> > > 192.168.39.22:60286: output error > >> > > > 2015-10-23 17:49:45,513 WARN [RpcServer.handler=6,port=60000] > >> > > ipc.RpcServer: RpcServer.handler=6,port=60000: caught a > >> > > ClosedChannelException, this means that the server was processing a > >> > request > >> > > but the client went away. The error message was: null > >> > > > > >> > > > > >> > > > > >> > > > > >> > > > > >> > > > > >> > > > > >> > > > > >> > > > > >> > > > > >> > > > > >> > > > ------------------ 原始邮件 ------------------ > >> > > > 发件人: "Ted Yu";<yuzhih...@gmail.com>; > >> > > > 发送时间: 2015年10月23日(星期五) 晚上11:39 > >> > > > 收件人: "user@hbase.apache.org"<user@hbase.apache.org>; > >> > > > > >> > > > 主题: Re: Hbase cluster is suddenly unable to respond > >> > > > > >> > > > > >> > > > > >> > > > Were other region servers functioning normally around 17:33 ? > >> > > > > >> > > > Which hbase release are you using ? > >> > > > > >> > > > Can you pastebin more of the region server log ? > >> > > > > >> > > > Thanks > >> > > > > >> > > >> On Fri, Oct 23, 2015 at 8:28 AM, 聪聪 <175998...@qq.com> wrote: > >> > > >> > >> > > >> hi,all: > >> > > >> > >> > > >> > >> > > >> This afternoon,The whole Hbase cluster is suddenly unable to > >> > > respond.after > >> > > >> I restart a regionserver after,the cluster has recovered.I don't > >> know > >> > > the > >> > > >> cause of the trouble.I hope I can get help from you. > >> > > >> > >> > > >> > >> > > >> Regionserver on the log is as follows: > >> > > >> 2015-10-23 17:28:49,335 INFO [regionserver60020.logRoller] > >> > wal.FSHLog: > >> > > >> moving old hlog file /hbase/WALs/l-hbase30.data.cn8.qunar.com > >> > > >> ,60020,1442810406218/l-hbase30.data.cn8.qunar.com > >> > > %2C60020%2C1442810406218.1445580462689 > >> > > >> whose highest sequenceid is 9071525521 to /hbase/oldWALs/ > >> > > >> l-hbase30.data.cn8.qunar.com%2C60020%2C1442810406218. > 1445580462689 > >> > > >> 2015-10-23 17:33:31,375 WARN [RpcServer.reader=8,port=60020] > >> > > >> ipc.RpcServer: RpcServer.listener,port=60020: count of bytes > read: > >> 0 > >> > > >> java.io.IOException: Connection reset by peer > >> > > >> at sun.nio.ch.FileDispatcherImpl.read0(Native Method) > >> > > >> at sun.nio.ch.SocketDispatcher.re > >> ad(SocketDispatcher.java:39) > >> > > >> at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil. > java:223) > >> > > >> at sun.nio.ch.IOUtil.read(IOUtil.java:197) > >> > > >> at > >> > sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:379) > >> > > >> at > >> > > >> org.apache.hadoop.hbase.ipc.RpcServer.channelRead(RpcServer. > >> java:2368) > >> > > >> at > >> > > >> > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc > >> ess(RpcServer.java:1403) > >> > > >> at > >> > > >> > >> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer > >> ver.java:770) > >> > > >> at > >> > > >> > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL > >> oop(RpcServer.java:563) > >> > > >> at > >> > > >> > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp > >> cServer.java:538) > >> > > >> at > >> > > >> > >> > > > >> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool > >> Executor.java:1145) > >> > > >> at > >> > > >> > >> > > > >> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo > >> lExecutor.java:615) > >> > > >> at java.lang.Thread.run(Thread.java:744) > >> > > >> 2015-10-23 17:33:31,779 WARN [RpcServer.reader=2,port=60020] > >> > > >> ipc.RpcServer: RpcServer.listener,port=60020: count of bytes > read: > >> 0 > >> > > >> java.io.IOException: Connection reset by peer > >> > > >> at sun.nio.ch.FileDispatcherImpl.read0(Native Method) > >> > > >> at sun.nio.ch.SocketDispatcher.re > >> ad(SocketDispatcher.java:39) > >> > > >> at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil. > java:223) > >> > > >> at sun.nio.ch.IOUtil.read(IOUtil.java:197) > >> > > >> at > >> > sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:379) > >> > > >> at > >> > > >> org.apache.hadoop.hbase.ipc.RpcServer.channelRead(RpcServer. > >> java:2368) > >> > > >> at > >> > > >> > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Connection.readAndProc > >> ess(RpcServer.java:1403) > >> > > >> at > >> > > >> > >> > > org.apache.hadoop.hbase.ipc.RpcServer$Listener.doRead(RpcSer > >> ver.java:770) > >> > > >> at > >> > > >> > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.doRunL > >> oop(RpcServer.java:563) > >> > > >> at > >> > > >> > >> > > > >> > org.apache.hadoop.hbase.ipc.RpcServer$Listener$Reader.run(Rp > >> cServer.java:538) > >> > > >> at > >> > > >> > >> > > > >> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool > >> Executor.java:1145) > >> > > >> at > >> > > >> > >> > > > >> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo > >> lExecutor.java:615) > >> > > >> at java.lang.Thread.run(Thread.java:744) > >> > > > >> > > >> > > > > > > > > -- > > Thank you > > Kiran Sarvabhotla > > > > -----Even a correct decision is wrong when it is taken late > > > > > > > -- > Thank you > Kiran Sarvabhotla > > -----Even a correct decision is wrong when it is taken late >