This could indicate swapping during GC. On Thu, Dec 9, 2010 at 12:13 PM, Lance Riedel <lancerie...@gmail.com> wrote:
> Seems reasonable, but having trouble making sense of the GC logs I had > turned on. Basically since there was a full GC a minute before this happens > on that server that lasts less than a second. > > > Example: > So, here is what the last of the GC logs say for that Regionserver > (04.hadoop on 10.100.154.10) (note the last full GC only 1 minutes before @ > 5:03:29.850+0000): > > 2010-12-09T05:02:19.472+0000: 12170.437: [GC [PSYoungGen: > 278229K->5024K(287360K)] 694269K->421693K(721856K), 0.0274840 secs] [Times: > user=0.04 sys=0.00, real=0.03 secs] > 2010-12-09T05:02:55.479+0000: 12206.444: [GC [PSYoungGen: > 281450K->7968K(273344K)] 698120K->429658K(707840K), 0.0503000 secs] [Times: > user=0.10 sys=0.00, real=0.05 secs] > 2010-12-09T05:03:29.800+0000: 12240.766: [GC [PSYoungGen: > 273295K->7712K(262464K)] 694986K->437306K(696960K), 0.0490770 secs] [Times: > user=0.12 sys=0.01, real=0.05 secs] > 2010-12-09T05:03:29.850+0000: 12240.815: [Full GC [PSYoungGen: > 7712K->0K(262464K)] [PSOldGen: 429594K->353152K(429824K)] > 437306K->353152K(692288K) [PSPermGen: 18016K->18016K(21248K)], 0.8960760 > secs] [Times: user=0.89 sys=0.00, real=0.90 secs] > 2010-12-09T05:04:03.345+0000: 12274.310: [GC [PSYoungGen: > 254720K->6944K(251584K)] 607872K->360528K(681408K), 0.0331060 secs] [Times: > user=0.03 sys=0.00, real=0.04 secs] > 2010-12-09T05:04:33.875+0000: 12304.840: [GC [PSYoungGen: > 251552K->7200K(242240K)] 605136K->367800K(672064K), 0.0336420 secs] [Times: > user=0.09 sys=0.01, real=0.03 secs] > Heap > PSYoungGen total 242240K, used 223390K [0x00007fcc640b0000, > 0x00007fcc77a90000, 0x00007fcc78e00000) > eden space 235008K, 91% used > [0x00007fcc640b0000,0x00007fcc713cfa30,0x00007fcc72630000) > from space 7232K, 99% used > [0x00007fcc769b0000,0x00007fcc770b8090,0x00007fcc770c0000) > to space 10048K, 0% used > [0x00007fcc770c0000,0x00007fcc770c0000,0x00007fcc77a90000) > PSOldGen total 429824K, used 360600K [0x00007fcc3a600000, > 0x00007fcc549c0000, 0x00007fcc640b0000) > object space 429824K, 83% used > [0x00007fcc3a600000,0x00007fcc50626060,0x00007fcc549c0000) > PSPermGen total 21248K, used 18131K [0x00007fcc35200000, > 0x00007fcc366c0000, 0x00007fcc3a600000) > object space 21248K, 85% used > [0x00007fcc35200000,0x00007fcc363b4d60,0x00007fcc366c0000) > > Now Dead / End of log > > > That said, I will continue down the path you pointed me on . > > Thanks, > Lance > > On Dec 9, 2010, at 11:40 AM, Jean-Daniel Cryans wrote: > > > Lance, > > > > Both those lines indicate the problem: > > > > IPC Server handler 13 on 60020 took 182416ms > > Client session timed out, have not heard from server in 182936ms > > > > It's very clear that your region servers are suffering from > > pause-of-the-world garbage collection issues. Basically this one GC'ed > > for 3 minutes, which is over the 1 minute timeout to consider a region > > server dead. The rest is just the side effect of that. > > > > This subject comes often on the mailing list, for example: > > http://search-hadoop.com/m/t97q027tr7f2 > > > > Make sure you don't swap, give plenty of RAM to HBase, use LZO > > compression, don't underprovision your machines, etc > > > > J-D > > > > On Thu, Dec 9, 2010 at 11:27 AM, Lance Riedel <lancerie...@gmail.com> > wrote: > >> > >> We have a 6 node cluster, 5 with region serves. 2 of the region servers > have been stable for days, but 3 of them keep crashing. Here are the logs > around around when the crash occurs. (btw, we are shoving approximately the > twitter firehose into hbase via flume) I'm an hbase newbie, but I have been > reading. Not sure what else is needed to help debug. When the problem > occurs, it is concurrent with issues that appear all the way down in the > data node. Also, interesting to note, all 3 servers seem to fail > differently somewhat: > >> > >> Servers (6): > >> 00.hadoop is the master > >> 01.hadoop, 03.hadoop (there is no 02.hadoop - flaky machine that got > taken out) - are stable Region Servers > >> 04-06.hadoop crash at different times, very different logs, Region > Servers > >> > >> > >> Master Attributes: > >> HBase Version 0.89.20100924+28, r > >> Hadoop Version 0.20.2+737, r98c55c28258aa6f42250569bd7fa431ac657bdbd > >> > >> HBase Root Directory hdfs://00.hadoop.****:54310/hbase > >> > >> Load average > >> 103.0 > >> Average number of regions per regionserver. Naive computation. > >> Regions On FS > >> 577 > >> > >> > >> (This is after recently restarting 04-06) > >> > >> > >> Address Start Code Load > >> 01.hadoop:60030 1291268910638 requests=223, regions=118, > usedHeap=519, maxHeap=987 > >> 03.hadoop:60030 1291269219610 requests=23, regions=111, usedHeap=862, > maxHeap=987 > >> 04.hadoop:60030 1291910665912 requests=169, regions=82, usedHeap=194, > maxHeap=888 > >> 05.hadoop:60030 1291909584060 requests=232, regions=110, > usedHeap=477, maxHeap=888 > >> 06.hadoop60030 1291909723787 requests=99, regions=94, usedHeap=394, > maxHeap=888 > >> Total: > >> servers: 5 > >> requests=746, regions=515 > >> > >> > >> > >> Following are various logs around the time of the failures for 04-06 > >> > >> > ************************************************************************************ > >> 04.hadoop > >> > ************************************************************************************ > >> REGION SERVER -> > >> 2010-12-09 05:01:58,640 INFO > org.apache.hadoop.hbase.regionserver.HRegion: compaction completed on region > article,a83858a08f2270d319f75a7b43c756c2453988e7,1291811929868.58f6d9fc80 > >> c78f3ca490b0280b4f1226. in 0sec > >> 2010-12-09 05:04:30,898 DEBUG > org.apache.hadoop.hbase.io.hfile.LruBlockCache: LRU Stats: total=1.46 MB, > free=176.33 MB, max=177.79 MB, blocks=0, accesses=200583, hits=0, hitRatio= > >> 0.00%%, evictions=0, evicted=0, evictedPerRun=NaN > >> 2010-12-09 05:04:59,612 INFO > org.apache.hadoop.hbase.regionserver.wal.SequenceFileLogWriter: Using syncFs > -- HDFS-200 > >> 2010-12-09 05:04:59,618 INFO > org.apache.hadoop.hbase.regionserver.wal.HLog: Roll > /hbase/.logs/04.hadoop.****,60020,1291858770125/10.100.154.103%3A60020.1291870585253, > entries=5318 > >> 6, filesize=63776311. New hlog > /hbase/.logs/04.hadoop.****,60020,1291858770125/10.100.154.103 > %3A60020.1291871099582 > >> 2010-12-09 05:08:02,033 DEBUG > org.apache.hadoop.hbase.regionserver.wal.HLog: Found 1 hlogs to remove out > of total 6; oldest outstanding sequenceid is 32891801 from region article > >> > ,4b2039b791e894dd479b90661ca97087f61645d3,1291813201039.7dff47c63bb5648bb1ee6670c60553a5. > >> 2010-12-09 05:08:02,034 WARN > org.apache.hadoop.hbase.regionserver.wal.HLog: IPC Server handler 13 on > 60020 took 182416ms appending an edit to hlog; editcount=0 > >> 2010-12-09 05:08:02,066 INFO > org.apache.hadoop.hbase.regionserver.wal.HLog: moving old hlog file > /hbase/.logs/04.hadoop.****,60020,1291858770125/10.100.154.103 > %3A60020.12918679476 > >> 65 whose highest sequenceid is 32880227 to /hbase/.oldlogs/ > 10.100.154.103%3A60020.1291867947665 > >> 2010-12-09 05:08:02,066 INFO org.apache.zookeeper.ClientCnxn: Client > session timed out, have not heard from server in 182936ms for sessionid > 0x12ca58c08ff0048, closing socket conn > >> ection and attempting reconnect > >> 2010-12-09 05:08:02,089 FATAL > org.apache.hadoop.hbase.regionserver.HRegionServer: Aborting region server > serverName=04.hadoop.****,60020,1291858770125, load=(requests=24, regions= > >> 64, usedHeap=548, maxHeap=888): Unhandled exception > >> org.apache.hadoop.hbase.YouAreDeadException: > org.apache.hadoop.hbase.YouAreDeadException: Server REPORT rejected; > currently processing 04.hadoop.****,60020,1291858770125 as dead s > >> erver > >> at > org.apache.hadoop.hbase.master.ServerManager.checkIsDead(ServerManager.java:217) > >> at > org.apache.hadoop.hbase.master.ServerManager.regionServerReport(ServerManager.java:271) > >> at > org.apache.hadoop.hbase.master.HMaster.regionServerReport(HMaster.java:744) > >> at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source) > >> at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > >> at java.lang.reflect.Method.invoke(Method.java:597) > >> at > org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:561) > >> at > org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:1039) > >> > >> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native > Method) > >> at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) > >> at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) > at java.lang.reflect.Constructor.newInstance(Constructor.java:513) > >> at > org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:94) > >> at > org.apache.hadoop.hbase.RemoteExceptionHandler.checkThrowable(RemoteExceptionHandler.java:48) > >> at > org.apache.hadoop.hbase.RemoteExceptionHandler.checkIOException(RemoteExceptionHandler.java:66) > >> at > org.apache.hadoop.hbase.regionserver.HRegionServer.doRun(HRegionServer.java:635) > >> at > org.apache.hadoop.hbase.regionserver.HRegionServer.access$000(HRegionServer.java:126) > >> at > org.apache.hadoop.hbase.regionserver.HRegionServer$1.run(HRegionServer.java:518) > >> at > org.apache.hadoop.hbase.regionserver.HRegionServer$1.run(HRegionServer.java:516) > >> at java.security.AccessController.doPrivileged(Native Method) > >> at javax.security.auth.Subject.doAs(Subject.java:337) > >> at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1046) > >> at > org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:516) > >> at java.lang.Thread.run(Thread.java:662) > >> 2010-12-09 05:08:02,090 INFO > org.apache.hadoop.hbase.regionserver.HRegionServer: Dump of metrics: > request=0.0, regions=64, stores=64, storefiles=136, storefileIndexSize=27, > memsto > >> reSize=353, compactionQueueSize=0, usedHeap=549, maxHeap=888, > blockCacheSize=1530552, blockCacheFree=184893160, blockCacheCount=0, > blockCacheHitRatio=0 > >> 2010-12-09 05:08:02,090 INFO org.apache.hadoop.ipc.HBaseServer: Stopping > server on 60020 > >> 2010-12-09 05:08:02,090 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 0 on 60020: exiting2010-12-09 05:08:02,090 INFO > org.apache.hadoop.ipc.HBaseServer: Stopping IPC Server listener on 60020 > >> > >> > >> MASTER HBASE -> > >> > >> 2010-12-09 05:05:06,072 INFO org.apache.hadoop.hbase.master.BaseScanner: > All 1 .META. region(s) scanned > >> 2010-12-09 05:05:40,004 INFO > org.apache.hadoop.hbase.master.ServerManager: > 04.hadoop.jive,60020,1291858770125 znode expired > >> 2010-12-09 05:05:40,005 DEBUG > org.apache.hadoop.hbase.master.ServerManager: > Added=04.hadoop.jive,60020,1291858770125 to dead servers, added shutdown > processing operation > >> 2010-12-09 05:05:40,005 DEBUG > org.apache.hadoop.hbase.master.RegionServerOperationQueue: Processing todo: > ProcessServerShutdown of 04.hadoop.jive,60020,12918587701252010-12-09 > 05:05:40,005 INFO org.apache.hadoop.hbase.master.RegionServerOperation: > Process shutdown of server 04.hadoop.jive,60020,1291858770125: logSplit: > false, rootRescanned: f > >> alse, numberOfMetaRegions: 1, onlineMetaRegions.size(): 1 > >> 2010-12-09 05:05:40,008 INFO > org.apache.hadoop.hbase.regionserver.wal.HLog: Splitting 7 hlog(s) in > hdfs://00.hadoop.jive:54310/hbase/.logs/04.hadoop.jive,60020,12918587701252010-12-09 > 05:05:40,008 DEBUG org.apache.hadoop.hbase.regionserver.wal.HLog: Splitting > hlog 1 of 7: > hdfs://00.hadoop.jive:54310/hbase/.logs/04.hadoop.jive,60020,1291858770125/10.1 > >> 00.154.103%3A60020.1291867947665, length=637577092010-12-09 05:05:40,008 > INFO org.apache.hadoop.hbase.util.FSUtils: Recovering > filehdfs://00.hadoop.jive:54310/hbase/.logs/04.hadoop.jive,60020,1291858770125/ > 10.100.154.103%3A60020 > >> .1291867947665 > >> > >> > >> NAMENODE -> > >> > >> 2010-12-09 05:08:02,471 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* > NameSystem.addStoredBlock: blockMap updated: 10.100.159.13:50010 is added > to blk_1531008743226086399_251615 size 63757709 > >> 2010-12-09 05:08:02,473 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 9 on 54310, call > complete(/hbase/.logs/04.hadoop.jive,60020,1291858770125/10.100.154.103%3A60020.1291867947665, > DFSClient_hb_m_10.194.194.79:60000_1291788452343) from 10.194.194.79:44117: > error: org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No > lease on > /hbase/.logs/04.hadoop.jive,60020,1291858770125/10.100.154.103%3A60020.1291867947665 > File does not exist. [Lease. Holder: > DFSClient_hb_m_10.194.194.79:60000_1291788452343, pendingcreates: 1] > >> org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease > on > /hbase/.logs/04.hadoop.jive,60020,1291858770125/10.100.154.103%3A60020.1291867947665 > File does not exist. [Lease. Holder: > DFSClient_hb_m_10.194.194.79:60000_1291788452343, pendingcreates: 1] > >> at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1488) > >> at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1479) > >> at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFileInternal(FSNamesystem.java:1534) > >> at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFile(FSNamesystem.java:1522) > >> at > org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:610) > >> at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source) > >> at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > >> at java.lang.reflect.Method.invoke(Method.java:597) > >> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:528) > >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1319) > >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1315) > >> at java.security.AccessController.doPrivileged(Native Method) > >> at javax.security.auth.Subject.doAs(Subject.java:396) > >> at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1063) > >> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1313) > >> 2010-12-09 05:08:04,206 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* > NameSystem.addStoredBlock: blockMap updated: 10.99.97.106:50010 is added > to blk_-734280257049179934_251614 size 7330 > >> > >> DATANODE -> > >> > >> 2010-12-09 05:08:02,212 WARN > org.apache.hadoop.hdfs.server.datanode.DataNode: IOException in > BlockReceiver.lastNodeRun: java.io.IOException: Broken pipe > >> at sun.nio.ch.FileDispatcher.write0(Native Method) > >> at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29) > at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100) at > sun.nio.ch.IOUtil.write(IOUtil.java:71) > >> at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334) > >> at > org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55) > at > org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142) > at > org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146) > >> at > org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107) > at java.io.DataOutputStream.writeLong(DataOutputStream.java:207) > >> at > org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.write(DataTransferProtocol.java:133) > at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.lastDataNodeRun(BlockReceiver.java:840) > >> at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:870) > >> at java.lang.Thread.run(Thread.java:662) > >> 2010-12-09 05:08:02,213 WARN > org.apache.hadoop.hdfs.server.datanode.DataNode: checkDiskError: exception: > >> 2010-12-09 05:08:02,213 WARN > org.apache.hadoop.hdfs.server.datanode.DataNode: checkDiskError: exception: > java.io.IOException: Broken pipe > >> at sun.nio.ch.FileDispatcher.write0(Native Method) > >> at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29) > >> at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100) > >> at sun.nio.ch.IOUtil.write(IOUtil.java:71) > >> at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334) > >> at > org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55) > >> at > org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142) > >> at > org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146) > >> at > org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107) > >> at java.io.DataOutputStream.writeLong(DataOutputStream.java:207) > >> at > org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.write(DataTransferProtocol.java:133) > >> at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.lastDataNodeRun(BlockReceiver.java:840) > >> at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:870) > >> at java.lang.Thread.run(Thread.java:662) > >> 2010-12-09 05:08:02,221 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder > blk_-8817504198034990390_251613 0 Exception java.io.IOException: Broken pipe > >> at sun.nio.ch.FileDispatcher.write0(Native Method) > >> at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29) > >> at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100) > >> at sun.nio.ch.IOUtil.write(IOUtil.java:71) > >> at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334) > >> at > org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55) > >> at > org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142) > >> at > org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146) > >> at > org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107) > >> at java.io.DataOutputStream.writeLong(DataOutputStream.java:207) > >> at > org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.write(DataTransferProtocol.java:133) > >> at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.lastDataNodeRun(BlockReceiver.java:840) > >> at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:870) > >> at java.lang.Thread.run(Thread.java:662) > >> > >> 2010-12-09 05:08:02,221 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 0 for block > blk_-8817504198034990390_251613 terminating > >> 2010-12-09 05:08:02,224 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: writeBlock > blk_-8817504198034990390_251613 received exception java.io.IOException: > Connection reset by peer > >> 2010-12-09 05:08:02,224 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: Scheduling block > blk_4505306716668305567_250752 file > /dist1/data/hadoop-data/current/subdir19/blk_4505306716668305567 for > deletion > >> 2010-12-09 05:08:02,226 ERROR > org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration( > 10.100.154.103:50010, > storageID=DS-866555524-10.100.154.103-50010-1291262739506, infoPort=50075, > ipcPort=50020):DataXceiver > >> java.io.IOException: Connection reset by peer > >> at sun.nio.ch.FileDispatcher.read0(Native Method) > >> at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:21) > >> at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:237) > >> at sun.nio.ch.IOUtil.read(IOUtil.java:210) > >> at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:236) > >> at > org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:55) > >> at > org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142) > >> at > org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155) > >> at > org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128) > >> at > java.io.BufferedInputStream.read1(BufferedInputStream.java:256) > >> at java.io.BufferedInputStream.read(BufferedInputStream.java:317) > >> at java.io.DataInputStream.read(DataInputStream.java:132) > >> at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readToBuf(BlockReceiver.java:267) > >> at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readNextPacket(BlockReceiver.java:357) > >> at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:378) > >> at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:534) > >> at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:417) > >> at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:122) > >> 2010-12-09 05:08:02,226 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: Deleted block > blk_1422641301942366074_250732 at file > /dist1/data/hadoop-data/current/subdir19/blk_1422641301942366074 > >> 2010-12-09 05:08:02,227 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: Deleted block > blk_4505306716668305567_250752 at file > /dist1/data/hadoop-data/current/subdir19/blk_4505306716668305567 > >> > >> > >> > >> > ************************************************************************************ > >> 05.hadoop > >> > ************************************************************************************ > >> REGION SERVER -> > >> 2010-12-09 04:29:25,047 DEBUG > org.apache.hadoop.hbase.regionserver.Store: Compaction size of data: 250.5m; > Skipped 1 file(s), size: 206896367 > >> 2010-12-09 04:29:25,047 INFO org.apache.hadoop.hbase.regionserver.Store: > Started compaction of 2 file(s) in data of > article,060c4b75067378c045eead9edec2d4fb7d8f3e5e,1291812563166.95e92e97823e11e57850720a63433a62. > into > hdfs://00.hadoop.****:54310/hbase/article/95e92e97823e11e57850720a63433a62/.tmp, > sequenceid=33050221 > >> 2010-12-09 04:29:26,328 INFO org.apache.hadoop.hbase.regionserver.Store: > Completed compaction of 2 file(s) in data of > article,060c4b75067378c045eead9edec2d4fb7d8f3e5e,1291812563166.95e92e97823e11e57850720a63433a62.; > new storefile is > hdfs://00.hadoop.****:54310/hbase/article/95e92e97823e11e57850720a63433a62/data/2787768957794924068; > store size is 250.5m > >> 2010-12-09 04:29:26,328 INFO > org.apache.hadoop.hbase.regionserver.HRegion: compaction completed on region > article,060c4b75067378c045eead9edec2d4fb7d8f3e5e,1291812563166.95e92e97823e11e57850720a63433a62. > in 1sec > >> 2010-12-09 04:29:26,328 INFO > org.apache.hadoop.hbase.regionserver.HRegion: Starting compaction on region > article,ed0cdc7ae6db73432e2ca47c973d545bee11ddd8,1291813400028.85a8d31d099f409dd36f3bcc4be1eba6. > >> 2010-12-09 04:29:26,330 INFO > org.apache.hadoop.hbase.regionserver.HRegion: compaction completed on region > article,ed0cdc7ae6db73432e2ca47c973d545bee11ddd8,1291813400028.85a8d31d099f409dd36f3bcc4be1eba6. > in 0sec > >> 2010-12-09 04:32:33,081 INFO > org.apache.hadoop.hbase.regionserver.HRegionServer: MSG_REGIONSERVER_STOP > >> 2010-12-09 04:32:33,081 INFO org.apache.hadoop.ipc.HBaseServer: Stopping > server on 60020 > >> 2010-12-09 04:32:33,089 WARN org.apache.hadoop.hdfs.DFSClient: > DFSOutputStream ResponseProcessor exception for block > blk_2012253533858803341_250248java.io.EOFException > >> at java.io.DataInputStream.readFully(DataInputStream.java:180) > >> at java.io.DataInputStream.readLong(DataInputStream.java:399) > >> at > org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.readFields(DataTransferProtocol.java:120) > >> at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2621) > >> > >> 2010-12-09 04:32:33,089 INFO org.apache.zookeeper.ClientCnxn: Unable to > read additional data from server sessionid 0x12ca58c08ff0049, likely server > has closed socket, closing sock > >> et connection and attempting reconnect > >> 2010-12-09 04:32:33,089 WARN > org.apache.hadoop.hbase.regionserver.wal.HLog: IPC Server handler 5 on 60020 > took 75446ms appending an edit to hlog; editcount=50505 > >> 2010-12-09 04:32:33,090 INFO org.apache.hadoop.ipc.HBaseServer: IPC > Server handler 18 on 60020: exiting > >> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.hdfs.DFSClient: Error > Recovery for block blk_2012253533858803341_250248 waiting for responder to > exit. > >> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: IPC > Server handler 8 on 60020: exiting > >> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 1 on 60020: exiting > >> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: IPC > Server handler 10 on 60020: exiting > >> 2010-12-09 04:32:33,140 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 0 on 60020: exiting2010-12-09 04:32:33,140 INFO > org.apache.hadoop.ipc.HBaseServer: PRI IPC Server handler 5 on 60020: > exiting > >> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 8 on 60020: exiting2010-12-09 04:32:33,140 INFO > org.apache.hadoop.ipc.HBaseServer: PRI IPC Server handler 4 on 60020: > exiting > >> 2010-12-09 04:32:33,140 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 3 on 60020: exiting2010-12-09 04:32:33,142 INFO > org.apache.hadoop.ipc.HBaseServer: PRI IPC Server handler 7 on 60020: > exiting > >> 2010-12-09 04:32:33,142 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 9 on 60020: exiting2010-12-09 04:32:33,143 INFO > org.apache.hadoop.hbase.regionserver.HRegionServer: Stopping infoServer > >> 2010-12-09 04:32:33,143 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 2 on 60020: exiting > >> 2010-12-09 04:32:33,143 INFO org.apache.hadoop.ipc.HBaseServer: Stopping > IPC Server Responder > >> 2010-12-09 04:32:33,143 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 6 on 60020: exiting > >> 2010-12-09 04:32:33,144 INFO org.mortbay.log: Stopped > selectchannelconnec...@0.0.0.0:60030 > >> > >> > >> > >> > >> NAMENODE -> > >> 2010-12-09 04:32:08,408 WARN org.apache.hadoop.hdfs.StateChange: DIR* > NameSystem.startFile: failed to create file > /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555 > for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client > 10.194.194.79, because this file is already being created by NN_Recovery on > 10.100.159.13 > >> 2010-12-09 04:32:08,408 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 6 on 54310, call > append(/hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555, > DFSClient_hb_m_10.194.194.79:60000_1291788452343) from 10.194.194.79:59118: > error: org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed > to create file /hba > >> se/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555 > for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client > 10.194.194.79, because this file is already being created by NN_Recovery on > 10.100.159.13 > >> org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to > create file > /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555 > for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client > 10.194.194.79, because this file is already being created by NN_Recovery on > 10.100.159.13 > >> at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:1194) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.appendFile(FSNamesystem.java:1282) > >> at > org.apache.hadoop.hdfs.server.namenode.NameNode.append(NameNode.java:541) > at sun.reflect.GeneratedMethodAccessor37.invoke(Unknown Source) > >> at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > >> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:528) at > org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1319) > >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1315) > at java.security.AccessController.doPrivileged(Native Method) > >> at javax.security.auth.Subject.doAs(Subject.java:396) at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1063) > >> at > org.apache.hadoop.ipc.Server$Handler.run(Server.java:1313)2010-12-09 > 04:32:09,411 WARN org.apache.hadoop.hdfs.StateChange: DIR* > NameSystem.startFile: failed to create file > /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A6 > >> 0020.1291868481555 for DFSClient_hb_m_10.194.194.79:60000_1291788452343 > on client 10.194.194.79, because this file is already being created by > NN_Recovery on 10.100.159.132010-12-09 04:32:09,411 INFO > org.apache.hadoop.ipc.Server: IPC Server handler 9 on 54310, call > append(/hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13 > %3A60020.1291868 > >> 481555, DFSClient_hb_m_10.194.194.79:60000_1291788452343) from > 10.194.194.79:59118: error: > org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to > create file > /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555 > for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client > 10.194.194.79, because this file is > >> already being created by NN_Recovery on > 10.100.159.13org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: > failed to create file /hbase/.logs/05.hadoop.jive,60020,1291862294517/ > 10.100.159.13%3A60020.1291868481555 for DFSClie > >> nt_hb_m_10.194.194.79:60000_1291788452343 on client 10.194.194.79, > because this file is already being created by NN_Recovery on 10.100.159.13 > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:1194) > >> at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.appendFile(FSNamesystem.java:1282) > at > org.apache.hadoop.hdfs.server.namenode.NameNode.append(NameNode.java:541) > >> at sun.reflect.GeneratedMethodAccessor37.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > >> at java.lang.reflect.Method.invoke(Method.java:597) at > org.apache.hadoop.ipc.RPC$Server.call(RPC.java:528) > >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1319) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1315) > >> > >> > >> > >> DATANODE > >> > >> 2010-12-09 04:32:09,534 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder > blk_2012253533858803341_250248 1 : Thread is interrupted. > >> 2010-12-09 04:32:09,534 ERROR > org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration( > 10.100.159.13:50010, > storageID=DS-145997007-10.100.159.13-50010-1291262743182, infoPort=50075, > ipcPort=50020):DataXceiver > >> java.io.IOException: Interrupted receiveBlock > >> at > org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:579) > >> at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:417) > >> at > org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:122) > >> 2010-12-09 04:32:09,534 INFO > org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 1 for block > blk_2012253533858803341_250248 terminating > >> > >> > >> > ************************************************************************************ > >> 06.hadoop > >> > ************************************************************************************ > >> REGION SERVER -> > >> > >> 2010-12-09 06:17:47,215 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:47,430 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:47,549 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:47,557 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:47,563 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:48,619 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:48,619 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:48,619 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:48,619 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:48,619 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:49,703 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:49,703 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:49,703 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:49,705 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> 2010-12-09 06:17:49,729 DEBUG > org.apache.hadoop.hbase.regionserver.HRegionServer: > NotServingRegionException; > article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. > >> > >> A bunch of these ^ > >> > >> Then a little later:: > >> .... > >> > >> 2010-12-09 06:23:23,561 DEBUG > org.apache.hadoop.hbase.regionserver.CompactSplitThread: Compaction > requested for region > article,993143fca3613014990cc6975cacdd62f4147e03,12918133247 > >> 53.19413f33fc660536d9d4f315a1283440. because: > regionserver60020.cacheFlusher > >> 2010-12-09 06:23:27,058 INFO > org.apache.hadoop.hbase.regionserver.wal.SequenceFileLogWriter: Using syncFs > -- HDFS-200 > >> 2010-12-09 06:23:27,080 INFO > org.apache.hadoop.hbase.regionserver.wal.HLog: Roll > /hbase/.logs/06.hadoop.****,60020,1291864620851/10.99.97.106%3A60020.1291875047947, > entries=53315, > >> filesize=63781925. New hlog > /hbase/.logs/06.hadoop.****,60020,1291864620851/10.99.97.106 > %3A60020.1291875807006 > >> 2010-12-09 06:23:27,080 DEBUG > org.apache.hadoop.hbase.regionserver.wal.HLog: Found 1 hlogs to remove out > of total 8; oldest outstanding sequenceid is 33526881 from region article > >> > ,1221b0013c61fd7446cb822f698d0bcafa487342,1291872610788.bb2aa6ae8c694a096a00e7bb91fc46c3. > >> 2010-12-09 06:23:27,080 INFO > org.apache.hadoop.hbase.regionserver.wal.HLog: moving old hlog file > /hbase/.logs/06.hadoop.****,60020,1291864620851/10.99.97.106 > %3A60020.1291872064424 > >> whose highest sequenceid is 33506639 to /hbase/.oldlogs/10.99.97.106 > %3A60020.1291872064424 > >> 2010-12-09 06:25:39,162 WARN > org.apache.hadoop.hbase.regionserver.wal.HLog: IPC Server handler 12 on > 60020 took 119034ms appending an edit to hlog; editcount=3683 > >> 2010-12-09 06:25:39,162 WARN org.apache.hadoop.hdfs.DFSClient: > DataStreamer Exception: java.io.IOException: Broken pipe > >> at sun.nio.ch.FileDispatcher.write0(Native Method) > >> at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29) > >> at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100) > >> at sun.nio.ch.IOUtil.write(IOUtil.java:71) > >> at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334) > >> at > org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55) > >> at > org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142) > >> at > org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146) > >> at > org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107) > >> at > java.io.BufferedOutputStream.write(BufferedOutputStream.java:105) > >> at java.io.DataOutputStream.write(DataOutputStream.java:90) > >> at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2509) > >> 2010-12-09 06:25:39,163 WARN org.apache.hadoop.hdfs.DFSClient: > DFSOutputStream ResponseProcessor exception for block > blk_1141147913164105086_258331java.io.EOFException > >> at java.io.DataInputStream.readFully(DataInputStream.java:180) > >> at java.io.DataInputStream.readLong(DataInputStream.java:399) > >> at > org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.readFields(DataTransferProtocol.java:120) > >> at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2621) > >> > >> 2010-12-09 06:25:39,163 INFO org.apache.hadoop.hdfs.DFSClient: Error > Recovery for block blk_1141147913164105086_258331 waiting for responder to > exit. > >> 2010-12-09 06:25:39,164 WARN org.apache.hadoop.hdfs.DFSClient: > DFSOutputStream ResponseProcessor exception for block > blk_-7076574401277381592_258324java.io.EOFException > >> at java.io.DataInputStream.readFully(DataInputStream.java:180) > >> at java.io.DataInputStream.readLong(DataInputStream.java:399) > >> at > org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.readFields(DataTransferProtocol.java:120) > >> at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2621) > >> > >> 2010-12-09 06:25:39,164 WARN org.apache.hadoop.hdfs.DFSClient: Error > Recovery for block blk_-7076574401277381592_258324 bad datanode[0] > 10.99.97.106:50010 > >> 2010-12-09 06:25:39,164 INFO org.apache.zookeeper.ClientCnxn: Unable to > read additional data from server sessionid 0x12ca58c08ff004b, likely server > has closed socket, closing socket connection and attempting reconnect > >> 2010-12-09 06:25:39,164 WARN org.apache.hadoop.hdfs.DFSClient: Error > Recovery for block blk_-7076574401277381592_258324 in pipeline > 10.99.97.106:50010, 10.100.154.103:50010: bad datanode 10.99.97.106:50010 > >> 2010-12-09 06:25:39,165 WARN org.apache.hadoop.hdfs.DFSClient: Error > Recovery for block blk_1141147913164105086_258331 bad datanode[0] > 10.99.97.106:50010 > >> 2010-12-09 06:25:39,167 INFO org.apache.zookeeper.ClientCnxn: Unable to > read additional data from server sessionid 0x12ca58c08ff004c, likely server > has closed socket, closing socket connection and attempting reconnect > >> 2010-12-09 06:25:39,169 INFO org.apache.hadoop.ipc.Client: Retrying > connect to server: /10.100.154.103:50020. Already tried 0 time(s). > >> 2010-12-09 06:25:39,169 WARN org.apache.hadoop.hdfs.DFSClient: Error > Recovery for block blk_1141147913164105086_258331 in pipeline > 10.99.97.106:50010, 10.100.159.13:50010: bad datanode 10.99.97.106:50010 > >> 2010-12-09 06:25:39,209 INFO > org.apache.hadoop.hbase.regionserver.HRegionServer: MSG_REGIONSERVER_STOP > >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: Stopping > server on 60020 > >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC > Server handler 3 on 60020: exiting > >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC > Server handler 9 on 60020: exiting > >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC > Server handler 24 on 60020: exiting > >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 1 on 60020: exiting > >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC > Server handler 15 on 60020: exiting > >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 0 on 60020: exiting > >> 2010-12-09 06:25:39,214 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 3 on 60020: exiting > >> 2010-12-09 06:25:39,214 INFO org.apache.hadoop.ipc.HBaseServer: Stopping > IPC Server listener on 60020 > >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 2 on 60020: exiting > >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 7 on 60020: exiting > >> 2010-12-09 06:25:39,213 INFO org.apache.hadoop.ipc.HBaseServer: IPC > Server handler 21 on 60020: exiting > >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 4 on 60020: exiting > >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: Stopping > IPC Server Responder > >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 5 on 60020: exiting > >> 2010-12-09 06:25:39,216 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 8 on 60020: exiting > >> 2010-12-09 06:25:39,216 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 9 on 60020: exiting > >> 2010-12-09 06:25:39,216 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC > Server handler 6 on 60020: exiting > >> 2010-12-09 06:25:39,216 INFO > org.apache.hadoop.hbase.regionserver.HRegionServer: Stopping infoServer > >> 2010-12-09 06:25:39,223 WARN org.apache.hadoop.hdfs.DFSClient: Failed > recovery attempt #0 from primary datanode 10.100.159.13:50010 > >> > >> > >> > >> > >