Seems reasonable, but having trouble making sense of the GC logs I had turned on. Basically since there was a full GC a minute before this happens on that server that lasts less than a second.
Example: So, here is what the last of the GC logs say for that Regionserver (04.hadoop on 10.100.154.10) (note the last full GC only 1 minutes before @ 5:03:29.850+0000): 2010-12-09T05:02:19.472+0000: 12170.437: [GC [PSYoungGen: 278229K->5024K(287360K)] 694269K->421693K(721856K), 0.0274840 secs] [Times: user=0.04 sys=0.00, real=0.03 secs] 2010-12-09T05:02:55.479+0000: 12206.444: [GC [PSYoungGen: 281450K->7968K(273344K)] 698120K->429658K(707840K), 0.0503000 secs] [Times: user=0.10 sys=0.00, real=0.05 secs] 2010-12-09T05:03:29.800+0000: 12240.766: [GC [PSYoungGen: 273295K->7712K(262464K)] 694986K->437306K(696960K), 0.0490770 secs] [Times: user=0.12 sys=0.01, real=0.05 secs] 2010-12-09T05:03:29.850+0000: 12240.815: [Full GC [PSYoungGen: 7712K->0K(262464K)] [PSOldGen: 429594K->353152K(429824K)] 437306K->353152K(692288K) [PSPermGen: 18016K->18016K(21248K)], 0.8960760 secs] [Times: user=0.89 sys=0.00, real=0.90 secs] 2010-12-09T05:04:03.345+0000: 12274.310: [GC [PSYoungGen: 254720K->6944K(251584K)] 607872K->360528K(681408K), 0.0331060 secs] [Times: user=0.03 sys=0.00, real=0.04 secs] 2010-12-09T05:04:33.875+0000: 12304.840: [GC [PSYoungGen: 251552K->7200K(242240K)] 605136K->367800K(672064K), 0.0336420 secs] [Times: user=0.09 sys=0.01, real=0.03 secs] Heap PSYoungGen total 242240K, used 223390K [0x00007fcc640b0000, 0x00007fcc77a90000, 0x00007fcc78e00000) eden space 235008K, 91% used [0x00007fcc640b0000,0x00007fcc713cfa30,0x00007fcc72630000) from space 7232K, 99% used [0x00007fcc769b0000,0x00007fcc770b8090,0x00007fcc770c0000) to space 10048K, 0% used [0x00007fcc770c0000,0x00007fcc770c0000,0x00007fcc77a90000) PSOldGen total 429824K, used 360600K [0x00007fcc3a600000, 0x00007fcc549c0000, 0x00007fcc640b0000) object space 429824K, 83% used [0x00007fcc3a600000,0x00007fcc50626060,0x00007fcc549c0000) PSPermGen total 21248K, used 18131K [0x00007fcc35200000, 0x00007fcc366c0000, 0x00007fcc3a600000) object space 21248K, 85% used [0x00007fcc35200000,0x00007fcc363b4d60,0x00007fcc366c0000) Now Dead / End of log That said, I will continue down the path you pointed me on . Thanks, Lance On Dec 9, 2010, at 11:40 AM, Jean-Daniel Cryans wrote: > Lance, > > Both those lines indicate the problem: > > IPC Server handler 13 on 60020 took 182416ms > Client session timed out, have not heard from server in 182936ms > > It's very clear that your region servers are suffering from > pause-of-the-world garbage collection issues. Basically this one GC'ed > for 3 minutes, which is over the 1 minute timeout to consider a region > server dead. The rest is just the side effect of that. > > This subject comes often on the mailing list, for example: > http://search-hadoop.com/m/t97q027tr7f2 > > Make sure you don't swap, give plenty of RAM to HBase, use LZO > compression, don't underprovision your machines, etc > > J-D > > On Thu, Dec 9, 2010 at 11:27 AM, Lance Riedel <lancerie...@gmail.com> wrote: >> >> We have a 6 node cluster, 5 with region serves. 2 of the region servers >> have been stable for days, but 3 of them keep crashing. Here are the logs >> around around when the crash occurs. (btw, we are shoving approximately the >> twitter firehose into hbase via flume) I'm an hbase newbie, but I have been >> reading. Not sure what else is needed to help debug. When the problem >> occurs, it is concurrent with issues that appear all the way down in the >> data node. Also, interesting to note, all 3 servers seem to fail >> differently somewhat: >> >> Servers (6): >> 00.hadoop is the master >> 01.hadoop, 03.hadoop (there is no 02.hadoop - flaky machine that got taken >> out) - are stable Region Servers >> 04-06.hadoop crash at different times, very different logs, Region Servers >> >> >> Master Attributes: >> HBase Version 0.89.20100924+28, r >> Hadoop Version 0.20.2+737, r98c55c28258aa6f42250569bd7fa431ac657bdbd >> >> HBase Root Directory hdfs://00.hadoop.****:54310/hbase >> >> Load average >> 103.0 >> Average number of regions per regionserver. Naive computation. >> Regions On FS >> 577 >> >> >> (This is after recently restarting 04-06) >> >> >> Address Start Code Load >> 01.hadoop:60030 1291268910638 requests=223, regions=118, usedHeap=519, >> maxHeap=987 >> 03.hadoop:60030 1291269219610 requests=23, regions=111, usedHeap=862, >> maxHeap=987 >> 04.hadoop:60030 1291910665912 requests=169, regions=82, usedHeap=194, >> maxHeap=888 >> 05.hadoop:60030 1291909584060 requests=232, regions=110, usedHeap=477, >> maxHeap=888 >> 06.hadoop60030 1291909723787 requests=99, regions=94, usedHeap=394, >> maxHeap=888 >> Total: >> servers: 5 >> requests=746, regions=515 >> >> >> >> Following are various logs around the time of the failures for 04-06 >> >> ************************************************************************************ >> 04.hadoop >> ************************************************************************************ >> REGION SERVER -> >> 2010-12-09 05:01:58,640 INFO org.apache.hadoop.hbase.regionserver.HRegion: >> compaction completed on region >> article,a83858a08f2270d319f75a7b43c756c2453988e7,1291811929868.58f6d9fc80 >> c78f3ca490b0280b4f1226. in 0sec >> 2010-12-09 05:04:30,898 DEBUG >> org.apache.hadoop.hbase.io.hfile.LruBlockCache: LRU Stats: total=1.46 MB, >> free=176.33 MB, max=177.79 MB, blocks=0, accesses=200583, hits=0, hitRatio= >> 0.00%%, evictions=0, evicted=0, evictedPerRun=NaN >> 2010-12-09 05:04:59,612 INFO >> org.apache.hadoop.hbase.regionserver.wal.SequenceFileLogWriter: Using syncFs >> -- HDFS-200 >> 2010-12-09 05:04:59,618 INFO org.apache.hadoop.hbase.regionserver.wal.HLog: >> Roll >> /hbase/.logs/04.hadoop.****,60020,1291858770125/10.100.154.103%3A60020.1291870585253, >> entries=5318 >> 6, filesize=63776311. New hlog >> /hbase/.logs/04.hadoop.****,60020,1291858770125/10.100.154.103%3A60020.1291871099582 >> 2010-12-09 05:08:02,033 DEBUG org.apache.hadoop.hbase.regionserver.wal.HLog: >> Found 1 hlogs to remove out of total 6; oldest outstanding sequenceid is >> 32891801 from region article >> ,4b2039b791e894dd479b90661ca97087f61645d3,1291813201039.7dff47c63bb5648bb1ee6670c60553a5. >> 2010-12-09 05:08:02,034 WARN org.apache.hadoop.hbase.regionserver.wal.HLog: >> IPC Server handler 13 on 60020 took 182416ms appending an edit to hlog; >> editcount=0 >> 2010-12-09 05:08:02,066 INFO org.apache.hadoop.hbase.regionserver.wal.HLog: >> moving old hlog file >> /hbase/.logs/04.hadoop.****,60020,1291858770125/10.100.154.103%3A60020.12918679476 >> 65 whose highest sequenceid is 32880227 to >> /hbase/.oldlogs/10.100.154.103%3A60020.1291867947665 >> 2010-12-09 05:08:02,066 INFO org.apache.zookeeper.ClientCnxn: Client session >> timed out, have not heard from server in 182936ms for sessionid >> 0x12ca58c08ff0048, closing socket conn >> ection and attempting reconnect >> 2010-12-09 05:08:02,089 FATAL >> org.apache.hadoop.hbase.regionserver.HRegionServer: Aborting region server >> serverName=04.hadoop.****,60020,1291858770125, load=(requests=24, regions= >> 64, usedHeap=548, maxHeap=888): Unhandled exception >> org.apache.hadoop.hbase.YouAreDeadException: >> org.apache.hadoop.hbase.YouAreDeadException: Server REPORT rejected; >> currently processing 04.hadoop.****,60020,1291858770125 as dead s >> erver >> at >> org.apache.hadoop.hbase.master.ServerManager.checkIsDead(ServerManager.java:217) >> at >> org.apache.hadoop.hbase.master.ServerManager.regionServerReport(ServerManager.java:271) >> at >> org.apache.hadoop.hbase.master.HMaster.regionServerReport(HMaster.java:744) >> at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) >> at java.lang.reflect.Method.invoke(Method.java:597) >> at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:561) >> at >> org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:1039) >> >> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >> Method) >> at >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) >> at >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) >> at java.lang.reflect.Constructor.newInstance(Constructor.java:513) >> at >> org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:94) >> at >> org.apache.hadoop.hbase.RemoteExceptionHandler.checkThrowable(RemoteExceptionHandler.java:48) >> at >> org.apache.hadoop.hbase.RemoteExceptionHandler.checkIOException(RemoteExceptionHandler.java:66) >> at >> org.apache.hadoop.hbase.regionserver.HRegionServer.doRun(HRegionServer.java:635) >> at >> org.apache.hadoop.hbase.regionserver.HRegionServer.access$000(HRegionServer.java:126) >> at >> org.apache.hadoop.hbase.regionserver.HRegionServer$1.run(HRegionServer.java:518) >> at >> org.apache.hadoop.hbase.regionserver.HRegionServer$1.run(HRegionServer.java:516) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:337) >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1046) >> at >> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:516) >> at java.lang.Thread.run(Thread.java:662) >> 2010-12-09 05:08:02,090 INFO >> org.apache.hadoop.hbase.regionserver.HRegionServer: Dump of metrics: >> request=0.0, regions=64, stores=64, storefiles=136, storefileIndexSize=27, >> memsto >> reSize=353, compactionQueueSize=0, usedHeap=549, maxHeap=888, >> blockCacheSize=1530552, blockCacheFree=184893160, blockCacheCount=0, >> blockCacheHitRatio=0 >> 2010-12-09 05:08:02,090 INFO org.apache.hadoop.ipc.HBaseServer: Stopping >> server on 60020 >> 2010-12-09 05:08:02,090 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 0 on 60020: exiting2010-12-09 05:08:02,090 INFO >> org.apache.hadoop.ipc.HBaseServer: Stopping IPC Server listener on 60020 >> >> >> MASTER HBASE -> >> >> 2010-12-09 05:05:06,072 INFO org.apache.hadoop.hbase.master.BaseScanner: All >> 1 .META. region(s) scanned >> 2010-12-09 05:05:40,004 INFO org.apache.hadoop.hbase.master.ServerManager: >> 04.hadoop.jive,60020,1291858770125 znode expired >> 2010-12-09 05:05:40,005 DEBUG org.apache.hadoop.hbase.master.ServerManager: >> Added=04.hadoop.jive,60020,1291858770125 to dead servers, added shutdown >> processing operation >> 2010-12-09 05:05:40,005 DEBUG >> org.apache.hadoop.hbase.master.RegionServerOperationQueue: Processing todo: >> ProcessServerShutdown of 04.hadoop.jive,60020,12918587701252010-12-09 >> 05:05:40,005 INFO org.apache.hadoop.hbase.master.RegionServerOperation: >> Process shutdown of server 04.hadoop.jive,60020,1291858770125: logSplit: >> false, rootRescanned: f >> alse, numberOfMetaRegions: 1, onlineMetaRegions.size(): 1 >> 2010-12-09 05:05:40,008 INFO org.apache.hadoop.hbase.regionserver.wal.HLog: >> Splitting 7 hlog(s) in >> hdfs://00.hadoop.jive:54310/hbase/.logs/04.hadoop.jive,60020,12918587701252010-12-09 >> 05:05:40,008 DEBUG org.apache.hadoop.hbase.regionserver.wal.HLog: Splitting >> hlog 1 of 7: >> hdfs://00.hadoop.jive:54310/hbase/.logs/04.hadoop.jive,60020,1291858770125/10.1 >> 00.154.103%3A60020.1291867947665, length=637577092010-12-09 05:05:40,008 >> INFO org.apache.hadoop.hbase.util.FSUtils: Recovering >> filehdfs://00.hadoop.jive:54310/hbase/.logs/04.hadoop.jive,60020,1291858770125/10.100.154.103%3A60020 >> .1291867947665 >> >> >> NAMENODE -> >> >> 2010-12-09 05:08:02,471 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* >> NameSystem.addStoredBlock: blockMap updated: 10.100.159.13:50010 is added to >> blk_1531008743226086399_251615 size 63757709 >> 2010-12-09 05:08:02,473 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 9 on 54310, call >> complete(/hbase/.logs/04.hadoop.jive,60020,1291858770125/10.100.154.103%3A60020.1291867947665, >> DFSClient_hb_m_10.194.194.79:60000_1291788452343) from 10.194.194.79:44117: >> error: org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No >> lease on >> /hbase/.logs/04.hadoop.jive,60020,1291858770125/10.100.154.103%3A60020.1291867947665 >> File does not exist. [Lease. Holder: >> DFSClient_hb_m_10.194.194.79:60000_1291788452343, pendingcreates: 1] >> org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease on >> /hbase/.logs/04.hadoop.jive,60020,1291858770125/10.100.154.103%3A60020.1291867947665 >> File does not exist. [Lease. Holder: >> DFSClient_hb_m_10.194.194.79:60000_1291788452343, pendingcreates: 1] >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1488) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1479) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFileInternal(FSNamesystem.java:1534) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFile(FSNamesystem.java:1522) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:610) >> at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) >> at java.lang.reflect.Method.invoke(Method.java:597) >> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:528) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1319) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1315) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:396) >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1063) >> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1313) >> 2010-12-09 05:08:04,206 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* >> NameSystem.addStoredBlock: blockMap updated: 10.99.97.106:50010 is added to >> blk_-734280257049179934_251614 size 7330 >> >> DATANODE -> >> >> 2010-12-09 05:08:02,212 WARN >> org.apache.hadoop.hdfs.server.datanode.DataNode: IOException in >> BlockReceiver.lastNodeRun: java.io.IOException: Broken pipe >> at sun.nio.ch.FileDispatcher.write0(Native Method) >> at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29) >> at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100) at >> sun.nio.ch.IOUtil.write(IOUtil.java:71) >> at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334) >> at >> org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55) >> at >> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142) >> at >> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146) >> at >> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107) >> at java.io.DataOutputStream.writeLong(DataOutputStream.java:207) >> at >> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.write(DataTransferProtocol.java:133) >> at >> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.lastDataNodeRun(BlockReceiver.java:840) >> at >> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:870) >> at java.lang.Thread.run(Thread.java:662) >> 2010-12-09 05:08:02,213 WARN >> org.apache.hadoop.hdfs.server.datanode.DataNode: checkDiskError: exception: >> 2010-12-09 05:08:02,213 WARN >> org.apache.hadoop.hdfs.server.datanode.DataNode: checkDiskError: exception: >> java.io.IOException: Broken pipe >> at sun.nio.ch.FileDispatcher.write0(Native Method) >> at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29) >> at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100) >> at sun.nio.ch.IOUtil.write(IOUtil.java:71) >> at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334) >> at >> org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55) >> at >> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142) >> at >> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146) >> at >> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107) >> at java.io.DataOutputStream.writeLong(DataOutputStream.java:207) >> at >> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.write(DataTransferProtocol.java:133) >> at >> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.lastDataNodeRun(BlockReceiver.java:840) >> at >> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:870) >> at java.lang.Thread.run(Thread.java:662) >> 2010-12-09 05:08:02,221 INFO >> org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder >> blk_-8817504198034990390_251613 0 Exception java.io.IOException: Broken pipe >> at sun.nio.ch.FileDispatcher.write0(Native Method) >> at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29) >> at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100) >> at sun.nio.ch.IOUtil.write(IOUtil.java:71) >> at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334) >> at >> org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55) >> at >> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142) >> at >> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146) >> at >> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107) >> at java.io.DataOutputStream.writeLong(DataOutputStream.java:207) >> at >> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.write(DataTransferProtocol.java:133) >> at >> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.lastDataNodeRun(BlockReceiver.java:840) >> at >> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:870) >> at java.lang.Thread.run(Thread.java:662) >> >> 2010-12-09 05:08:02,221 INFO >> org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 0 for block >> blk_-8817504198034990390_251613 terminating >> 2010-12-09 05:08:02,224 INFO >> org.apache.hadoop.hdfs.server.datanode.DataNode: writeBlock >> blk_-8817504198034990390_251613 received exception java.io.IOException: >> Connection reset by peer >> 2010-12-09 05:08:02,224 INFO >> org.apache.hadoop.hdfs.server.datanode.DataNode: Scheduling block >> blk_4505306716668305567_250752 file >> /dist1/data/hadoop-data/current/subdir19/blk_4505306716668305567 for deletion >> 2010-12-09 05:08:02,226 ERROR >> org.apache.hadoop.hdfs.server.datanode.DataNode: >> DatanodeRegistration(10.100.154.103:50010, >> storageID=DS-866555524-10.100.154.103-50010-1291262739506, infoPort=50075, >> ipcPort=50020):DataXceiver >> java.io.IOException: Connection reset by peer >> at sun.nio.ch.FileDispatcher.read0(Native Method) >> at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:21) >> at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:237) >> at sun.nio.ch.IOUtil.read(IOUtil.java:210) >> at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:236) >> at >> org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:55) >> at >> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142) >> at >> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155) >> at >> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128) >> at java.io.BufferedInputStream.read1(BufferedInputStream.java:256) >> at java.io.BufferedInputStream.read(BufferedInputStream.java:317) >> at java.io.DataInputStream.read(DataInputStream.java:132) >> at >> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readToBuf(BlockReceiver.java:267) >> at >> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readNextPacket(BlockReceiver.java:357) >> at >> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:378) >> at >> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:534) >> at >> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:417) >> at >> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:122) >> 2010-12-09 05:08:02,226 INFO >> org.apache.hadoop.hdfs.server.datanode.DataNode: Deleted block >> blk_1422641301942366074_250732 at file >> /dist1/data/hadoop-data/current/subdir19/blk_1422641301942366074 >> 2010-12-09 05:08:02,227 INFO >> org.apache.hadoop.hdfs.server.datanode.DataNode: Deleted block >> blk_4505306716668305567_250752 at file >> /dist1/data/hadoop-data/current/subdir19/blk_4505306716668305567 >> >> >> >> ************************************************************************************ >> 05.hadoop >> ************************************************************************************ >> REGION SERVER -> >> 2010-12-09 04:29:25,047 DEBUG org.apache.hadoop.hbase.regionserver.Store: >> Compaction size of data: 250.5m; Skipped 1 file(s), size: 206896367 >> 2010-12-09 04:29:25,047 INFO org.apache.hadoop.hbase.regionserver.Store: >> Started compaction of 2 file(s) in data of >> article,060c4b75067378c045eead9edec2d4fb7d8f3e5e,1291812563166.95e92e97823e11e57850720a63433a62. >> into >> hdfs://00.hadoop.****:54310/hbase/article/95e92e97823e11e57850720a63433a62/.tmp, >> sequenceid=33050221 >> 2010-12-09 04:29:26,328 INFO org.apache.hadoop.hbase.regionserver.Store: >> Completed compaction of 2 file(s) in data of >> article,060c4b75067378c045eead9edec2d4fb7d8f3e5e,1291812563166.95e92e97823e11e57850720a63433a62.; >> new storefile is >> hdfs://00.hadoop.****:54310/hbase/article/95e92e97823e11e57850720a63433a62/data/2787768957794924068; >> store size is 250.5m >> 2010-12-09 04:29:26,328 INFO org.apache.hadoop.hbase.regionserver.HRegion: >> compaction completed on region >> article,060c4b75067378c045eead9edec2d4fb7d8f3e5e,1291812563166.95e92e97823e11e57850720a63433a62. >> in 1sec >> 2010-12-09 04:29:26,328 INFO org.apache.hadoop.hbase.regionserver.HRegion: >> Starting compaction on region >> article,ed0cdc7ae6db73432e2ca47c973d545bee11ddd8,1291813400028.85a8d31d099f409dd36f3bcc4be1eba6. >> 2010-12-09 04:29:26,330 INFO org.apache.hadoop.hbase.regionserver.HRegion: >> compaction completed on region >> article,ed0cdc7ae6db73432e2ca47c973d545bee11ddd8,1291813400028.85a8d31d099f409dd36f3bcc4be1eba6. >> in 0sec >> 2010-12-09 04:32:33,081 INFO >> org.apache.hadoop.hbase.regionserver.HRegionServer: MSG_REGIONSERVER_STOP >> 2010-12-09 04:32:33,081 INFO org.apache.hadoop.ipc.HBaseServer: Stopping >> server on 60020 >> 2010-12-09 04:32:33,089 WARN org.apache.hadoop.hdfs.DFSClient: >> DFSOutputStream ResponseProcessor exception for block >> blk_2012253533858803341_250248java.io.EOFException >> at java.io.DataInputStream.readFully(DataInputStream.java:180) >> at java.io.DataInputStream.readLong(DataInputStream.java:399) >> at >> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.readFields(DataTransferProtocol.java:120) >> at >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2621) >> >> 2010-12-09 04:32:33,089 INFO org.apache.zookeeper.ClientCnxn: Unable to read >> additional data from server sessionid 0x12ca58c08ff0049, likely server has >> closed socket, closing sock >> et connection and attempting reconnect >> 2010-12-09 04:32:33,089 WARN org.apache.hadoop.hbase.regionserver.wal.HLog: >> IPC Server handler 5 on 60020 took 75446ms appending an edit to hlog; >> editcount=50505 >> 2010-12-09 04:32:33,090 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server >> handler 18 on 60020: exiting >> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.hdfs.DFSClient: Error >> Recovery for block blk_2012253533858803341_250248 waiting for responder to >> exit. >> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server >> handler 8 on 60020: exiting >> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 1 on 60020: exiting >> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server >> handler 10 on 60020: exiting >> 2010-12-09 04:32:33,140 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 0 on 60020: exiting2010-12-09 04:32:33,140 INFO >> org.apache.hadoop.ipc.HBaseServer: PRI IPC Server handler 5 on 60020: exiting >> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 8 on 60020: exiting2010-12-09 04:32:33,140 INFO >> org.apache.hadoop.ipc.HBaseServer: PRI IPC Server handler 4 on 60020: exiting >> 2010-12-09 04:32:33,140 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 3 on 60020: exiting2010-12-09 04:32:33,142 INFO >> org.apache.hadoop.ipc.HBaseServer: PRI IPC Server handler 7 on 60020: exiting >> 2010-12-09 04:32:33,142 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 9 on 60020: exiting2010-12-09 04:32:33,143 INFO >> org.apache.hadoop.hbase.regionserver.HRegionServer: Stopping infoServer >> 2010-12-09 04:32:33,143 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 2 on 60020: exiting >> 2010-12-09 04:32:33,143 INFO org.apache.hadoop.ipc.HBaseServer: Stopping IPC >> Server Responder >> 2010-12-09 04:32:33,143 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 6 on 60020: exiting >> 2010-12-09 04:32:33,144 INFO org.mortbay.log: Stopped >> selectchannelconnec...@0.0.0.0:60030 >> >> >> >> >> NAMENODE -> >> 2010-12-09 04:32:08,408 WARN org.apache.hadoop.hdfs.StateChange: DIR* >> NameSystem.startFile: failed to create file >> /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555 >> for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client >> 10.194.194.79, because this file is already being created by NN_Recovery on >> 10.100.159.13 >> 2010-12-09 04:32:08,408 INFO org.apache.hadoop.ipc.Server: IPC Server >> handler 6 on 54310, call >> append(/hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555, >> DFSClient_hb_m_10.194.194.79:60000_1291788452343) from 10.194.194.79:59118: >> error: org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed >> to create file /hba >> se/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555 >> for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client >> 10.194.194.79, because this file is already being created by NN_Recovery on >> 10.100.159.13 >> org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to >> create file >> /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555 >> for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client >> 10.194.194.79, because this file is already being created by NN_Recovery on >> 10.100.159.13 >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:1194) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.appendFile(FSNamesystem.java:1282) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.append(NameNode.java:541) >> at sun.reflect.GeneratedMethodAccessor37.invoke(Unknown Source) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) >> at java.lang.reflect.Method.invoke(Method.java:597) >> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:528) at >> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1319) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1315) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:396) at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1063) >> at >> org.apache.hadoop.ipc.Server$Handler.run(Server.java:1313)2010-12-09 >> 04:32:09,411 WARN org.apache.hadoop.hdfs.StateChange: DIR* >> NameSystem.startFile: failed to create file >> /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A6 >> 0020.1291868481555 for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on >> client 10.194.194.79, because this file is already being created by >> NN_Recovery on 10.100.159.132010-12-09 04:32:09,411 INFO >> org.apache.hadoop.ipc.Server: IPC Server handler 9 on 54310, call >> append(/hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868 >> 481555, DFSClient_hb_m_10.194.194.79:60000_1291788452343) from >> 10.194.194.79:59118: error: >> org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to >> create file >> /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555 >> for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client >> 10.194.194.79, because this file is >> already being created by NN_Recovery on >> 10.100.159.13org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: >> failed to create file >> /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555 >> for DFSClie >> nt_hb_m_10.194.194.79:60000_1291788452343 on client 10.194.194.79, because >> this file is already being created by NN_Recovery on 10.100.159.13 at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:1194) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.appendFile(FSNamesystem.java:1282) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNode.append(NameNode.java:541) >> at sun.reflect.GeneratedMethodAccessor37.invoke(Unknown Source) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) >> at java.lang.reflect.Method.invoke(Method.java:597) at >> org.apache.hadoop.ipc.RPC$Server.call(RPC.java:528) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1319) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1315) >> >> >> >> DATANODE >> >> 2010-12-09 04:32:09,534 INFO >> org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder >> blk_2012253533858803341_250248 1 : Thread is interrupted. >> 2010-12-09 04:32:09,534 ERROR >> org.apache.hadoop.hdfs.server.datanode.DataNode: >> DatanodeRegistration(10.100.159.13:50010, >> storageID=DS-145997007-10.100.159.13-50010-1291262743182, infoPort=50075, >> ipcPort=50020):DataXceiver >> java.io.IOException: Interrupted receiveBlock >> at >> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:579) >> at >> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:417) >> at >> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:122) >> 2010-12-09 04:32:09,534 INFO >> org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 1 for block >> blk_2012253533858803341_250248 terminating >> >> >> ************************************************************************************ >> 06.hadoop >> ************************************************************************************ >> REGION SERVER -> >> >> 2010-12-09 06:17:47,215 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:47,430 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:47,549 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:47,557 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:47,563 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:48,619 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:48,619 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:48,619 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:48,619 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:48,619 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:49,703 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:49,703 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:49,703 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:49,705 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> 2010-12-09 06:17:49,729 DEBUG >> org.apache.hadoop.hbase.regionserver.HRegionServer: >> NotServingRegionException; >> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f. >> >> A bunch of these ^ >> >> Then a little later:: >> .... >> >> 2010-12-09 06:23:23,561 DEBUG >> org.apache.hadoop.hbase.regionserver.CompactSplitThread: Compaction >> requested for region >> article,993143fca3613014990cc6975cacdd62f4147e03,12918133247 >> 53.19413f33fc660536d9d4f315a1283440. because: regionserver60020.cacheFlusher >> 2010-12-09 06:23:27,058 INFO >> org.apache.hadoop.hbase.regionserver.wal.SequenceFileLogWriter: Using syncFs >> -- HDFS-200 >> 2010-12-09 06:23:27,080 INFO org.apache.hadoop.hbase.regionserver.wal.HLog: >> Roll >> /hbase/.logs/06.hadoop.****,60020,1291864620851/10.99.97.106%3A60020.1291875047947, >> entries=53315, >> filesize=63781925. New hlog >> /hbase/.logs/06.hadoop.****,60020,1291864620851/10.99.97.106%3A60020.1291875807006 >> 2010-12-09 06:23:27,080 DEBUG org.apache.hadoop.hbase.regionserver.wal.HLog: >> Found 1 hlogs to remove out of total 8; oldest outstanding sequenceid is >> 33526881 from region article >> ,1221b0013c61fd7446cb822f698d0bcafa487342,1291872610788.bb2aa6ae8c694a096a00e7bb91fc46c3. >> 2010-12-09 06:23:27,080 INFO org.apache.hadoop.hbase.regionserver.wal.HLog: >> moving old hlog file >> /hbase/.logs/06.hadoop.****,60020,1291864620851/10.99.97.106%3A60020.1291872064424 >> whose highest sequenceid is 33506639 to >> /hbase/.oldlogs/10.99.97.106%3A60020.1291872064424 >> 2010-12-09 06:25:39,162 WARN org.apache.hadoop.hbase.regionserver.wal.HLog: >> IPC Server handler 12 on 60020 took 119034ms appending an edit to hlog; >> editcount=3683 >> 2010-12-09 06:25:39,162 WARN org.apache.hadoop.hdfs.DFSClient: DataStreamer >> Exception: java.io.IOException: Broken pipe >> at sun.nio.ch.FileDispatcher.write0(Native Method) >> at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29) >> at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100) >> at sun.nio.ch.IOUtil.write(IOUtil.java:71) >> at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334) >> at >> org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55) >> at >> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142) >> at >> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146) >> at >> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107) >> at java.io.BufferedOutputStream.write(BufferedOutputStream.java:105) >> at java.io.DataOutputStream.write(DataOutputStream.java:90) >> at >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2509) >> 2010-12-09 06:25:39,163 WARN org.apache.hadoop.hdfs.DFSClient: >> DFSOutputStream ResponseProcessor exception for block >> blk_1141147913164105086_258331java.io.EOFException >> at java.io.DataInputStream.readFully(DataInputStream.java:180) >> at java.io.DataInputStream.readLong(DataInputStream.java:399) >> at >> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.readFields(DataTransferProtocol.java:120) >> at >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2621) >> >> 2010-12-09 06:25:39,163 INFO org.apache.hadoop.hdfs.DFSClient: Error >> Recovery for block blk_1141147913164105086_258331 waiting for responder to >> exit. >> 2010-12-09 06:25:39,164 WARN org.apache.hadoop.hdfs.DFSClient: >> DFSOutputStream ResponseProcessor exception for block >> blk_-7076574401277381592_258324java.io.EOFException >> at java.io.DataInputStream.readFully(DataInputStream.java:180) >> at java.io.DataInputStream.readLong(DataInputStream.java:399) >> at >> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.readFields(DataTransferProtocol.java:120) >> at >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2621) >> >> 2010-12-09 06:25:39,164 WARN org.apache.hadoop.hdfs.DFSClient: Error >> Recovery for block blk_-7076574401277381592_258324 bad datanode[0] >> 10.99.97.106:50010 >> 2010-12-09 06:25:39,164 INFO org.apache.zookeeper.ClientCnxn: Unable to read >> additional data from server sessionid 0x12ca58c08ff004b, likely server has >> closed socket, closing socket connection and attempting reconnect >> 2010-12-09 06:25:39,164 WARN org.apache.hadoop.hdfs.DFSClient: Error >> Recovery for block blk_-7076574401277381592_258324 in pipeline >> 10.99.97.106:50010, 10.100.154.103:50010: bad datanode 10.99.97.106:50010 >> 2010-12-09 06:25:39,165 WARN org.apache.hadoop.hdfs.DFSClient: Error >> Recovery for block blk_1141147913164105086_258331 bad datanode[0] >> 10.99.97.106:50010 >> 2010-12-09 06:25:39,167 INFO org.apache.zookeeper.ClientCnxn: Unable to read >> additional data from server sessionid 0x12ca58c08ff004c, likely server has >> closed socket, closing socket connection and attempting reconnect >> 2010-12-09 06:25:39,169 INFO org.apache.hadoop.ipc.Client: Retrying connect >> to server: /10.100.154.103:50020. Already tried 0 time(s). >> 2010-12-09 06:25:39,169 WARN org.apache.hadoop.hdfs.DFSClient: Error >> Recovery for block blk_1141147913164105086_258331 in pipeline >> 10.99.97.106:50010, 10.100.159.13:50010: bad datanode 10.99.97.106:50010 >> 2010-12-09 06:25:39,209 INFO >> org.apache.hadoop.hbase.regionserver.HRegionServer: MSG_REGIONSERVER_STOP >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: Stopping >> server on 60020 >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server >> handler 3 on 60020: exiting >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server >> handler 9 on 60020: exiting >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server >> handler 24 on 60020: exiting >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 1 on 60020: exiting >> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server >> handler 15 on 60020: exiting >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 0 on 60020: exiting >> 2010-12-09 06:25:39,214 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 3 on 60020: exiting >> 2010-12-09 06:25:39,214 INFO org.apache.hadoop.ipc.HBaseServer: Stopping IPC >> Server listener on 60020 >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 2 on 60020: exiting >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 7 on 60020: exiting >> 2010-12-09 06:25:39,213 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server >> handler 21 on 60020: exiting >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 4 on 60020: exiting >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: Stopping IPC >> Server Responder >> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 5 on 60020: exiting >> 2010-12-09 06:25:39,216 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 8 on 60020: exiting >> 2010-12-09 06:25:39,216 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 9 on 60020: exiting >> 2010-12-09 06:25:39,216 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC >> Server handler 6 on 60020: exiting >> 2010-12-09 06:25:39,216 INFO >> org.apache.hadoop.hbase.regionserver.HRegionServer: Stopping infoServer >> 2010-12-09 06:25:39,223 WARN org.apache.hadoop.hdfs.DFSClient: Failed >> recovery attempt #0 from primary datanode 10.100.159.13:50010 >> >> >> >>