I came across just probably the culprit -  I just discovered that the machines 
that were dying had OS swap turned OFF.

So, I added swap to those machines, but also reduced the memory configs for 
tasks. (was at 1 gig, now at 512m).  Seems to be stable right now, but the 
nightly longer term jobs run at night, so we will see how they do.

This is an ec2 c1.xlarge, so the memory is at 8 gigs.  I had 4 mappers per 
node, which was fine when all they had on them were Tasktrackers and datanodes. 
But with the hbase regionservers, the machines could definitely get 
over-allocated.

Lance

On Dec 9, 2010, at 3:57 PM, Ted Dunning wrote:

> This could indicate swapping during GC.
> 
> On Thu, Dec 9, 2010 at 12:13 PM, Lance Riedel <lancerie...@gmail.com> wrote:
> 
>> Seems reasonable, but having trouble making sense of the GC logs I had
>> turned on.  Basically since there was a full GC a minute before this happens
>> on that server that lasts less than a second.
>> 
>> 
>> Example:
>> So, here is what the last of the GC logs say for that Regionserver
>> (04.hadoop on 10.100.154.10) (note the last full GC only 1 minutes before @
>> 5:03:29.850+0000):
>> 
>> 2010-12-09T05:02:19.472+0000: 12170.437: [GC [PSYoungGen:
>> 278229K->5024K(287360K)] 694269K->421693K(721856K), 0.0274840 secs] [Times:
>> user=0.04 sys=0.00, real=0.03 secs]
>> 2010-12-09T05:02:55.479+0000: 12206.444: [GC [PSYoungGen:
>> 281450K->7968K(273344K)] 698120K->429658K(707840K), 0.0503000 secs] [Times:
>> user=0.10 sys=0.00, real=0.05 secs]
>> 2010-12-09T05:03:29.800+0000: 12240.766: [GC [PSYoungGen:
>> 273295K->7712K(262464K)] 694986K->437306K(696960K), 0.0490770 secs] [Times:
>> user=0.12 sys=0.01, real=0.05 secs]
>> 2010-12-09T05:03:29.850+0000: 12240.815: [Full GC [PSYoungGen:
>> 7712K->0K(262464K)] [PSOldGen: 429594K->353152K(429824K)]
>> 437306K->353152K(692288K) [PSPermGen: 18016K->18016K(21248K)], 0.8960760
>> secs] [Times: user=0.89 sys=0.00, real=0.90 secs]
>> 2010-12-09T05:04:03.345+0000: 12274.310: [GC [PSYoungGen:
>> 254720K->6944K(251584K)] 607872K->360528K(681408K), 0.0331060 secs] [Times:
>> user=0.03 sys=0.00, real=0.04 secs]
>> 2010-12-09T05:04:33.875+0000: 12304.840: [GC [PSYoungGen:
>> 251552K->7200K(242240K)] 605136K->367800K(672064K), 0.0336420 secs] [Times:
>> user=0.09 sys=0.01, real=0.03 secs]
>> Heap
>> PSYoungGen      total 242240K, used 223390K [0x00007fcc640b0000,
>> 0x00007fcc77a90000, 0x00007fcc78e00000)
>> eden space 235008K, 91% used
>> [0x00007fcc640b0000,0x00007fcc713cfa30,0x00007fcc72630000)
>> from space 7232K, 99% used
>> [0x00007fcc769b0000,0x00007fcc770b8090,0x00007fcc770c0000)
>> to   space 10048K, 0% used
>> [0x00007fcc770c0000,0x00007fcc770c0000,0x00007fcc77a90000)
>> PSOldGen        total 429824K, used 360600K [0x00007fcc3a600000,
>> 0x00007fcc549c0000, 0x00007fcc640b0000)
>> object space 429824K, 83% used
>> [0x00007fcc3a600000,0x00007fcc50626060,0x00007fcc549c0000)
>> PSPermGen       total 21248K, used 18131K [0x00007fcc35200000,
>> 0x00007fcc366c0000, 0x00007fcc3a600000)
>> object space 21248K, 85% used
>> [0x00007fcc35200000,0x00007fcc363b4d60,0x00007fcc366c0000)
>> 
>> Now Dead / End of log
>> 
>> 
>> That said, I will continue down the path you pointed me on .
>> 
>> Thanks,
>> Lance
>> 
>> On Dec 9, 2010, at 11:40 AM, Jean-Daniel Cryans wrote:
>> 
>>> Lance,
>>> 
>>> Both those lines indicate the problem:
>>> 
>>> IPC Server handler 13 on 60020 took 182416ms
>>> Client session timed out, have not heard from server in 182936ms
>>> 
>>> It's very clear that your region servers are suffering from
>>> pause-of-the-world garbage collection issues. Basically this one GC'ed
>>> for 3 minutes, which is over the 1 minute timeout to consider a region
>>> server dead. The rest is just the side effect of that.
>>> 
>>> This subject comes often on the mailing list, for example:
>>> http://search-hadoop.com/m/t97q027tr7f2
>>> 
>>> Make sure you don't swap, give plenty of RAM to HBase, use LZO
>>> compression, don't underprovision your machines, etc
>>> 
>>> J-D
>>> 
>>> On Thu, Dec 9, 2010 at 11:27 AM, Lance Riedel <lancerie...@gmail.com>
>> wrote:
>>>> 
>>>> We have a 6 node cluster, 5 with region serves.  2 of the region servers
>> have been stable for days, but 3 of them keep crashing.  Here are the logs
>> around around when the crash occurs. (btw, we are shoving approximately the
>> twitter firehose into hbase via flume) I'm an hbase newbie, but I have been
>> reading. Not sure what else is needed to help debug. When the problem
>> occurs, it is concurrent with issues that appear all the way down in the
>> data node.  Also, interesting to note, all 3 servers seem to fail
>> differently somewhat:
>>>> 
>>>> Servers (6):
>>>> 00.hadoop is the master
>>>> 01.hadoop, 03.hadoop (there is no 02.hadoop - flaky machine that got
>> taken out) - are stable Region Servers
>>>> 04-06.hadoop crash at different times, very different logs, Region
>> Servers
>>>> 
>>>> 
>>>> Master Attributes:
>>>> HBase Version 0.89.20100924+28, r
>>>> Hadoop Version 0.20.2+737, r98c55c28258aa6f42250569bd7fa431ac657bdbd
>>>> 
>>>> HBase Root Directory hdfs://00.hadoop.****:54310/hbase
>>>> 
>>>> Load average
>>>> 103.0
>>>> Average number of regions per regionserver. Naive computation.
>>>> Regions On FS
>>>> 577
>>>> 
>>>> 
>>>> (This is after recently restarting 04-06)
>>>> 
>>>> 
>>>> Address   Start Code    Load
>>>> 01.hadoop:60030  1291268910638   requests=223, regions=118,
>> usedHeap=519, maxHeap=987
>>>> 03.hadoop:60030  1291269219610   requests=23, regions=111, usedHeap=862,
>> maxHeap=987
>>>> 04.hadoop:60030  1291910665912   requests=169, regions=82, usedHeap=194,
>> maxHeap=888
>>>> 05.hadoop:60030  1291909584060   requests=232, regions=110,
>> usedHeap=477, maxHeap=888
>>>> 06.hadoop60030   1291909723787   requests=99, regions=94, usedHeap=394,
>> maxHeap=888
>>>> Total:
>>>> servers: 5
>>>> requests=746, regions=515
>>>> 
>>>> 
>>>> 
>>>> Following are various logs around the time of the failures for 04-06
>>>> 
>>>> 
>> ************************************************************************************
>>>> 04.hadoop
>>>> 
>> ************************************************************************************
>>>> REGION SERVER ->
>>>> 2010-12-09 05:01:58,640 INFO
>> org.apache.hadoop.hbase.regionserver.HRegion: compaction completed on region
>> article,a83858a08f2270d319f75a7b43c756c2453988e7,1291811929868.58f6d9fc80
>>>> c78f3ca490b0280b4f1226. in 0sec
>>>> 2010-12-09 05:04:30,898 DEBUG
>> org.apache.hadoop.hbase.io.hfile.LruBlockCache: LRU Stats: total=1.46 MB,
>> free=176.33 MB, max=177.79 MB, blocks=0, accesses=200583, hits=0, hitRatio=
>>>> 0.00%%, evictions=0, evicted=0, evictedPerRun=NaN
>>>> 2010-12-09 05:04:59,612 INFO
>> org.apache.hadoop.hbase.regionserver.wal.SequenceFileLogWriter: Using syncFs
>> -- HDFS-200
>>>> 2010-12-09 05:04:59,618 INFO
>> org.apache.hadoop.hbase.regionserver.wal.HLog: Roll
>> /hbase/.logs/04.hadoop.****,60020,1291858770125/10.100.154.103%3A60020.1291870585253,
>> entries=5318
>>>> 6, filesize=63776311. New hlog
>> /hbase/.logs/04.hadoop.****,60020,1291858770125/10.100.154.103
>> %3A60020.1291871099582
>>>> 2010-12-09 05:08:02,033 DEBUG
>> org.apache.hadoop.hbase.regionserver.wal.HLog: Found 1 hlogs to remove  out
>> of total 6; oldest outstanding sequenceid is 32891801 from region article
>>>> 
>> ,4b2039b791e894dd479b90661ca97087f61645d3,1291813201039.7dff47c63bb5648bb1ee6670c60553a5.
>>>> 2010-12-09 05:08:02,034 WARN
>> org.apache.hadoop.hbase.regionserver.wal.HLog: IPC Server handler 13 on
>> 60020 took 182416ms appending an edit to hlog; editcount=0
>>>> 2010-12-09 05:08:02,066 INFO
>> org.apache.hadoop.hbase.regionserver.wal.HLog: moving old hlog file
>> /hbase/.logs/04.hadoop.****,60020,1291858770125/10.100.154.103
>> %3A60020.12918679476
>>>> 65 whose highest sequenceid is 32880227 to /hbase/.oldlogs/
>> 10.100.154.103%3A60020.1291867947665
>>>> 2010-12-09 05:08:02,066 INFO org.apache.zookeeper.ClientCnxn: Client
>> session timed out, have not heard from server in 182936ms for sessionid
>> 0x12ca58c08ff0048, closing socket conn
>>>> ection and attempting reconnect
>>>> 2010-12-09 05:08:02,089 FATAL
>> org.apache.hadoop.hbase.regionserver.HRegionServer: Aborting region server
>> serverName=04.hadoop.****,60020,1291858770125, load=(requests=24, regions=
>>>> 64, usedHeap=548, maxHeap=888): Unhandled exception
>>>> org.apache.hadoop.hbase.YouAreDeadException:
>> org.apache.hadoop.hbase.YouAreDeadException: Server REPORT rejected;
>> currently processing 04.hadoop.****,60020,1291858770125 as dead s
>>>> erver
>>>>       at
>> org.apache.hadoop.hbase.master.ServerManager.checkIsDead(ServerManager.java:217)
>>>>       at
>> org.apache.hadoop.hbase.master.ServerManager.regionServerReport(ServerManager.java:271)
>>>>       at
>> org.apache.hadoop.hbase.master.HMaster.regionServerReport(HMaster.java:744)
>>>>       at sun.reflect.GeneratedMethodAccessor2.invoke(Unknown Source)
>>>>       at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>>>       at java.lang.reflect.Method.invoke(Method.java:597)
>>>>       at
>> org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:561)
>>>>       at
>> org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:1039)
>>>> 
>>>>       at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>> Method)
>>>>       at
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>>>>       at
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>>       at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>>>>       at
>> org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:94)
>>>>       at
>> org.apache.hadoop.hbase.RemoteExceptionHandler.checkThrowable(RemoteExceptionHandler.java:48)
>>>>       at
>> org.apache.hadoop.hbase.RemoteExceptionHandler.checkIOException(RemoteExceptionHandler.java:66)
>>>>       at
>> org.apache.hadoop.hbase.regionserver.HRegionServer.doRun(HRegionServer.java:635)
>>>>       at
>> org.apache.hadoop.hbase.regionserver.HRegionServer.access$000(HRegionServer.java:126)
>>>>       at
>> org.apache.hadoop.hbase.regionserver.HRegionServer$1.run(HRegionServer.java:518)
>>>>       at
>> org.apache.hadoop.hbase.regionserver.HRegionServer$1.run(HRegionServer.java:516)
>>>>       at java.security.AccessController.doPrivileged(Native Method)
>>>>       at javax.security.auth.Subject.doAs(Subject.java:337)
>>>>       at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1046)
>>>>       at
>> org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:516)
>>>>       at java.lang.Thread.run(Thread.java:662)
>>>> 2010-12-09 05:08:02,090 INFO
>> org.apache.hadoop.hbase.regionserver.HRegionServer: Dump of metrics:
>> request=0.0, regions=64, stores=64, storefiles=136, storefileIndexSize=27,
>> memsto
>>>> reSize=353, compactionQueueSize=0, usedHeap=549, maxHeap=888,
>> blockCacheSize=1530552, blockCacheFree=184893160, blockCacheCount=0,
>> blockCacheHitRatio=0
>>>> 2010-12-09 05:08:02,090 INFO org.apache.hadoop.ipc.HBaseServer: Stopping
>> server on 60020
>>>> 2010-12-09 05:08:02,090 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 0 on 60020: exiting2010-12-09 05:08:02,090 INFO
>> org.apache.hadoop.ipc.HBaseServer: Stopping IPC Server listener on 60020
>>>> 
>>>> 
>>>> MASTER HBASE ->
>>>> 
>>>> 2010-12-09 05:05:06,072 INFO org.apache.hadoop.hbase.master.BaseScanner:
>> All 1 .META. region(s) scanned
>>>> 2010-12-09 05:05:40,004 INFO
>> org.apache.hadoop.hbase.master.ServerManager:
>> 04.hadoop.jive,60020,1291858770125 znode expired
>>>> 2010-12-09 05:05:40,005 DEBUG
>> org.apache.hadoop.hbase.master.ServerManager:
>> Added=04.hadoop.jive,60020,1291858770125 to dead servers, added shutdown
>> processing operation
>>>> 2010-12-09 05:05:40,005 DEBUG
>> org.apache.hadoop.hbase.master.RegionServerOperationQueue: Processing todo:
>> ProcessServerShutdown of 04.hadoop.jive,60020,12918587701252010-12-09
>> 05:05:40,005 INFO org.apache.hadoop.hbase.master.RegionServerOperation:
>> Process shutdown of server 04.hadoop.jive,60020,1291858770125: logSplit:
>> false, rootRescanned: f
>>>> alse, numberOfMetaRegions: 1, onlineMetaRegions.size(): 1
>>>> 2010-12-09 05:05:40,008 INFO
>> org.apache.hadoop.hbase.regionserver.wal.HLog: Splitting 7 hlog(s) in
>> hdfs://00.hadoop.jive:54310/hbase/.logs/04.hadoop.jive,60020,12918587701252010-12-09
>> 05:05:40,008 DEBUG org.apache.hadoop.hbase.regionserver.wal.HLog: Splitting
>> hlog 1 of 7:
>> hdfs://00.hadoop.jive:54310/hbase/.logs/04.hadoop.jive,60020,1291858770125/10.1
>>>> 00.154.103%3A60020.1291867947665, length=637577092010-12-09 05:05:40,008
>> INFO org.apache.hadoop.hbase.util.FSUtils: Recovering
>> filehdfs://00.hadoop.jive:54310/hbase/.logs/04.hadoop.jive,60020,1291858770125/
>> 10.100.154.103%3A60020
>>>> .1291867947665
>>>> 
>>>> 
>>>> NAMENODE ->
>>>> 
>>>> 2010-12-09 05:08:02,471 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
>> NameSystem.addStoredBlock: blockMap updated: 10.100.159.13:50010 is added
>> to blk_1531008743226086399_251615 size 63757709
>>>> 2010-12-09 05:08:02,473 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 9 on 54310, call
>> complete(/hbase/.logs/04.hadoop.jive,60020,1291858770125/10.100.154.103%3A60020.1291867947665,
>> DFSClient_hb_m_10.194.194.79:60000_1291788452343) from 10.194.194.79:44117:
>> error: org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No
>> lease on 
>> /hbase/.logs/04.hadoop.jive,60020,1291858770125/10.100.154.103%3A60020.1291867947665
>> File does not exist. [Lease.  Holder:
>> DFSClient_hb_m_10.194.194.79:60000_1291788452343, pendingcreates: 1]
>>>> org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease
>> on 
>> /hbase/.logs/04.hadoop.jive,60020,1291858770125/10.100.154.103%3A60020.1291867947665
>> File does not exist. [Lease.  Holder:
>> DFSClient_hb_m_10.194.194.79:60000_1291788452343, pendingcreates: 1]
>>>>       at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1488)
>>>>       at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1479)
>>>>       at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFileInternal(FSNamesystem.java:1534)
>>>>       at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.completeFile(FSNamesystem.java:1522)
>>>>       at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.complete(NameNode.java:610)
>>>>       at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source)
>>>>       at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>>>       at java.lang.reflect.Method.invoke(Method.java:597)
>>>>       at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:528)
>>>>       at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1319)
>>>>       at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1315)
>>>>       at java.security.AccessController.doPrivileged(Native Method)
>>>>       at javax.security.auth.Subject.doAs(Subject.java:396)
>>>>       at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1063)
>>>>       at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1313)
>>>> 2010-12-09 05:08:04,206 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
>> NameSystem.addStoredBlock: blockMap updated: 10.99.97.106:50010 is added
>> to blk_-734280257049179934_251614 size 7330
>>>> 
>>>> DATANODE ->
>>>> 
>>>> 2010-12-09 05:08:02,212 WARN
>> org.apache.hadoop.hdfs.server.datanode.DataNode: IOException in
>> BlockReceiver.lastNodeRun: java.io.IOException: Broken pipe
>>>>       at sun.nio.ch.FileDispatcher.write0(Native Method)
>>>>       at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29)
>>     at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100)        at
>> sun.nio.ch.IOUtil.write(IOUtil.java:71)
>>>>       at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334)
>>>>       at
>> org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55)
>>       at
>> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142)
>>       at
>> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
>>>>       at
>> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
>>       at java.io.DataOutputStream.writeLong(DataOutputStream.java:207)
>>>>       at
>> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.write(DataTransferProtocol.java:133)
>>       at
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.lastDataNodeRun(BlockReceiver.java:840)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:870)
>>>>       at java.lang.Thread.run(Thread.java:662)
>>>> 2010-12-09 05:08:02,213 WARN
>> org.apache.hadoop.hdfs.server.datanode.DataNode: checkDiskError: exception:
>>>> 2010-12-09 05:08:02,213 WARN
>> org.apache.hadoop.hdfs.server.datanode.DataNode: checkDiskError: exception:
>> java.io.IOException: Broken pipe
>>>>       at sun.nio.ch.FileDispatcher.write0(Native Method)
>>>>       at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29)
>>>>       at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100)
>>>>       at sun.nio.ch.IOUtil.write(IOUtil.java:71)
>>>>       at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334)
>>>>       at
>> org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55)
>>>>       at
>> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142)
>>>>       at
>> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
>>>>       at
>> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
>>>>       at java.io.DataOutputStream.writeLong(DataOutputStream.java:207)
>>>>       at
>> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.write(DataTransferProtocol.java:133)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.lastDataNodeRun(BlockReceiver.java:840)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:870)
>>>>       at java.lang.Thread.run(Thread.java:662)
>>>> 2010-12-09 05:08:02,221 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder
>> blk_-8817504198034990390_251613 0 Exception java.io.IOException: Broken pipe
>>>>       at sun.nio.ch.FileDispatcher.write0(Native Method)
>>>>       at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29)
>>>>       at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100)
>>>>       at sun.nio.ch.IOUtil.write(IOUtil.java:71)
>>>>       at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334)
>>>>       at
>> org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55)
>>>>       at
>> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142)
>>>>       at
>> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
>>>>       at
>> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
>>>>       at java.io.DataOutputStream.writeLong(DataOutputStream.java:207)
>>>>       at
>> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.write(DataTransferProtocol.java:133)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.lastDataNodeRun(BlockReceiver.java:840)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:870)
>>>>       at java.lang.Thread.run(Thread.java:662)
>>>> 
>>>> 2010-12-09 05:08:02,221 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 0 for block
>> blk_-8817504198034990390_251613 terminating
>>>> 2010-12-09 05:08:02,224 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: writeBlock
>> blk_-8817504198034990390_251613 received exception java.io.IOException:
>> Connection reset by peer
>>>> 2010-12-09 05:08:02,224 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Scheduling block
>> blk_4505306716668305567_250752 file
>> /dist1/data/hadoop-data/current/subdir19/blk_4505306716668305567 for
>> deletion
>>>> 2010-12-09 05:08:02,226 ERROR
>> org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
>> 10.100.154.103:50010,
>> storageID=DS-866555524-10.100.154.103-50010-1291262739506, infoPort=50075,
>> ipcPort=50020):DataXceiver
>>>> java.io.IOException: Connection reset by peer
>>>>       at sun.nio.ch.FileDispatcher.read0(Native Method)
>>>>       at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:21)
>>>>       at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:237)
>>>>       at sun.nio.ch.IOUtil.read(IOUtil.java:210)
>>>>       at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:236)
>>>>       at
>> org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:55)
>>>>       at
>> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142)
>>>>       at
>> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:155)
>>>>       at
>> org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:128)
>>>>       at
>> java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
>>>>       at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
>>>>       at java.io.DataInputStream.read(DataInputStream.java:132)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readToBuf(BlockReceiver.java:267)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.readNextPacket(BlockReceiver.java:357)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:378)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:534)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:417)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:122)
>>>> 2010-12-09 05:08:02,226 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Deleted block
>> blk_1422641301942366074_250732 at file
>> /dist1/data/hadoop-data/current/subdir19/blk_1422641301942366074
>>>> 2010-12-09 05:08:02,227 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Deleted block
>> blk_4505306716668305567_250752 at file
>> /dist1/data/hadoop-data/current/subdir19/blk_4505306716668305567
>>>> 
>>>> 
>>>> 
>>>> 
>> ************************************************************************************
>>>> 05.hadoop
>>>> 
>> ************************************************************************************
>>>> REGION SERVER ->
>>>> 2010-12-09 04:29:25,047 DEBUG
>> org.apache.hadoop.hbase.regionserver.Store: Compaction size of data: 250.5m;
>> Skipped 1 file(s), size: 206896367
>>>> 2010-12-09 04:29:25,047 INFO org.apache.hadoop.hbase.regionserver.Store:
>> Started compaction of 2 file(s) in data of
>> article,060c4b75067378c045eead9edec2d4fb7d8f3e5e,1291812563166.95e92e97823e11e57850720a63433a62.
>> into
>> hdfs://00.hadoop.****:54310/hbase/article/95e92e97823e11e57850720a63433a62/.tmp,
>> sequenceid=33050221
>>>> 2010-12-09 04:29:26,328 INFO org.apache.hadoop.hbase.regionserver.Store:
>> Completed compaction of 2 file(s) in data of
>> article,060c4b75067378c045eead9edec2d4fb7d8f3e5e,1291812563166.95e92e97823e11e57850720a63433a62.;
>> new storefile is
>> hdfs://00.hadoop.****:54310/hbase/article/95e92e97823e11e57850720a63433a62/data/2787768957794924068;
>> store size is 250.5m
>>>> 2010-12-09 04:29:26,328 INFO
>> org.apache.hadoop.hbase.regionserver.HRegion: compaction completed on region
>> article,060c4b75067378c045eead9edec2d4fb7d8f3e5e,1291812563166.95e92e97823e11e57850720a63433a62.
>> in 1sec
>>>> 2010-12-09 04:29:26,328 INFO
>> org.apache.hadoop.hbase.regionserver.HRegion: Starting compaction on region
>> article,ed0cdc7ae6db73432e2ca47c973d545bee11ddd8,1291813400028.85a8d31d099f409dd36f3bcc4be1eba6.
>>>> 2010-12-09 04:29:26,330 INFO
>> org.apache.hadoop.hbase.regionserver.HRegion: compaction completed on region
>> article,ed0cdc7ae6db73432e2ca47c973d545bee11ddd8,1291813400028.85a8d31d099f409dd36f3bcc4be1eba6.
>> in 0sec
>>>> 2010-12-09 04:32:33,081 INFO
>> org.apache.hadoop.hbase.regionserver.HRegionServer: MSG_REGIONSERVER_STOP
>>>> 2010-12-09 04:32:33,081 INFO org.apache.hadoop.ipc.HBaseServer: Stopping
>> server on 60020
>>>> 2010-12-09 04:32:33,089 WARN org.apache.hadoop.hdfs.DFSClient:
>> DFSOutputStream ResponseProcessor exception  for block
>> blk_2012253533858803341_250248java.io.EOFException
>>>>       at java.io.DataInputStream.readFully(DataInputStream.java:180)
>>>>       at java.io.DataInputStream.readLong(DataInputStream.java:399)
>>>>       at
>> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.readFields(DataTransferProtocol.java:120)
>>>>       at
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2621)
>>>> 
>>>> 2010-12-09 04:32:33,089 INFO org.apache.zookeeper.ClientCnxn: Unable to
>> read additional data from server sessionid 0x12ca58c08ff0049, likely server
>> has closed socket, closing sock
>>>> et connection and attempting reconnect
>>>> 2010-12-09 04:32:33,089 WARN
>> org.apache.hadoop.hbase.regionserver.wal.HLog: IPC Server handler 5 on 60020
>> took 75446ms appending an edit to hlog; editcount=50505
>>>> 2010-12-09 04:32:33,090 INFO org.apache.hadoop.ipc.HBaseServer: IPC
>> Server handler 18 on 60020: exiting
>>>> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.hdfs.DFSClient: Error
>> Recovery for block blk_2012253533858803341_250248 waiting for responder to
>> exit.
>>>> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: IPC
>> Server handler 8 on 60020: exiting
>>>> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 1 on 60020: exiting
>>>> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: IPC
>> Server handler 10 on 60020: exiting
>>>> 2010-12-09 04:32:33,140 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 0 on 60020: exiting2010-12-09 04:32:33,140 INFO
>> org.apache.hadoop.ipc.HBaseServer: PRI IPC Server handler 5 on 60020:
>> exiting
>>>> 2010-12-09 04:32:33,139 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 8 on 60020: exiting2010-12-09 04:32:33,140 INFO
>> org.apache.hadoop.ipc.HBaseServer: PRI IPC Server handler 4 on 60020:
>> exiting
>>>> 2010-12-09 04:32:33,140 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 3 on 60020: exiting2010-12-09 04:32:33,142 INFO
>> org.apache.hadoop.ipc.HBaseServer: PRI IPC Server handler 7 on 60020:
>> exiting
>>>> 2010-12-09 04:32:33,142 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 9 on 60020: exiting2010-12-09 04:32:33,143 INFO
>> org.apache.hadoop.hbase.regionserver.HRegionServer: Stopping infoServer
>>>> 2010-12-09 04:32:33,143 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 2 on 60020: exiting
>>>> 2010-12-09 04:32:33,143 INFO org.apache.hadoop.ipc.HBaseServer: Stopping
>> IPC Server Responder
>>>> 2010-12-09 04:32:33,143 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 6 on 60020: exiting
>>>> 2010-12-09 04:32:33,144 INFO org.mortbay.log: Stopped
>> selectchannelconnec...@0.0.0.0:60030
>>>> 
>>>> 
>>>> 
>>>> 
>>>> NAMENODE ->
>>>> 2010-12-09 04:32:08,408 WARN org.apache.hadoop.hdfs.StateChange: DIR*
>> NameSystem.startFile: failed to create file
>> /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555
>> for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client
>> 10.194.194.79, because this file is already being created by NN_Recovery on
>> 10.100.159.13
>>>> 2010-12-09 04:32:08,408 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 6 on 54310, call
>> append(/hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555,
>> DFSClient_hb_m_10.194.194.79:60000_1291788452343) from 10.194.194.79:59118:
>> error: org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed
>> to create file /hba
>>>> se/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555
>> for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client
>> 10.194.194.79, because this file is already being created by NN_Recovery on
>> 10.100.159.13
>>>> org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to
>> create file 
>> /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555
>> for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client
>> 10.194.194.79, because this file is already being created by NN_Recovery on
>> 10.100.159.13
>>>>       at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:1194)
>>       at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.appendFile(FSNamesystem.java:1282)
>>>>       at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.append(NameNode.java:541)
>>     at sun.reflect.GeneratedMethodAccessor37.invoke(Unknown Source)
>>>>       at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>       at java.lang.reflect.Method.invoke(Method.java:597)
>>>>       at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:528)        at
>> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1319)
>>>>       at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1315)
>>     at java.security.AccessController.doPrivileged(Native Method)
>>>>       at javax.security.auth.Subject.doAs(Subject.java:396)        at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1063)
>>>>       at
>> org.apache.hadoop.ipc.Server$Handler.run(Server.java:1313)2010-12-09
>> 04:32:09,411 WARN org.apache.hadoop.hdfs.StateChange: DIR*
>> NameSystem.startFile: failed to create file
>> /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A6
>>>> 0020.1291868481555 for DFSClient_hb_m_10.194.194.79:60000_1291788452343
>> on client 10.194.194.79, because this file is already being created by
>> NN_Recovery on 10.100.159.132010-12-09 04:32:09,411 INFO
>> org.apache.hadoop.ipc.Server: IPC Server handler 9 on 54310, call
>> append(/hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13
>> %3A60020.1291868
>>>> 481555, DFSClient_hb_m_10.194.194.79:60000_1291788452343) from
>> 10.194.194.79:59118: error:
>> org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException: failed to
>> create file 
>> /hbase/.logs/05.hadoop.jive,60020,1291862294517/10.100.159.13%3A60020.1291868481555
>> for DFSClient_hb_m_10.194.194.79:60000_1291788452343 on client
>> 10.194.194.79, because this file is
>>>> already being created by NN_Recovery on
>> 10.100.159.13org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException:
>> failed to create file /hbase/.logs/05.hadoop.jive,60020,1291862294517/
>> 10.100.159.13%3A60020.1291868481555 for DFSClie
>>>> nt_hb_m_10.194.194.79:60000_1291788452343 on client 10.194.194.79,
>> because this file is already being created by NN_Recovery on 10.100.159.13
>>     at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:1194)
>>>>       at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.appendFile(FSNamesystem.java:1282)
>>       at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.append(NameNode.java:541)
>>>>       at sun.reflect.GeneratedMethodAccessor37.invoke(Unknown Source)
>>     at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>>>       at java.lang.reflect.Method.invoke(Method.java:597)        at
>> org.apache.hadoop.ipc.RPC$Server.call(RPC.java:528)
>>>>       at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1319)
>>     at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1315)
>>>> 
>>>> 
>>>> 
>>>> DATANODE
>>>> 
>>>> 2010-12-09 04:32:09,534 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder
>> blk_2012253533858803341_250248 1 : Thread is interrupted.
>>>> 2010-12-09 04:32:09,534 ERROR
>> org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
>> 10.100.159.13:50010,
>> storageID=DS-145997007-10.100.159.13-50010-1291262743182, infoPort=50075,
>> ipcPort=50020):DataXceiver
>>>> java.io.IOException: Interrupted receiveBlock
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:579)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:417)
>>>>       at
>> org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:122)
>>>> 2010-12-09 04:32:09,534 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: PacketResponder 1 for block
>> blk_2012253533858803341_250248 terminating
>>>> 
>>>> 
>>>> 
>> ************************************************************************************
>>>> 06.hadoop
>>>> 
>> ************************************************************************************
>>>> REGION SERVER ->
>>>> 
>>>> 2010-12-09 06:17:47,215 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:47,430 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:47,549 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:47,557 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:47,563 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:48,619 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:48,619 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:48,619 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:48,619 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:48,619 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:49,703 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:49,703 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:49,703 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:49,705 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 2010-12-09 06:17:49,729 DEBUG
>> org.apache.hadoop.hbase.regionserver.HRegionServer:
>> NotServingRegionException;
>> article,8d36892f14dfa6d17d08900743960d0b0973fc96,1291812431639.cce2b650859704eb184df8fb5c5eee6f.
>>>> 
>>>> A bunch of these ^
>>>> 
>>>> Then a little later::
>>>> ....
>>>> 
>>>> 2010-12-09 06:23:23,561 DEBUG
>> org.apache.hadoop.hbase.regionserver.CompactSplitThread: Compaction
>> requested for region
>> article,993143fca3613014990cc6975cacdd62f4147e03,12918133247
>>>> 53.19413f33fc660536d9d4f315a1283440. because:
>> regionserver60020.cacheFlusher
>>>> 2010-12-09 06:23:27,058 INFO
>> org.apache.hadoop.hbase.regionserver.wal.SequenceFileLogWriter: Using syncFs
>> -- HDFS-200
>>>> 2010-12-09 06:23:27,080 INFO
>> org.apache.hadoop.hbase.regionserver.wal.HLog: Roll
>> /hbase/.logs/06.hadoop.****,60020,1291864620851/10.99.97.106%3A60020.1291875047947,
>> entries=53315,
>>>> filesize=63781925. New hlog
>> /hbase/.logs/06.hadoop.****,60020,1291864620851/10.99.97.106
>> %3A60020.1291875807006
>>>> 2010-12-09 06:23:27,080 DEBUG
>> org.apache.hadoop.hbase.regionserver.wal.HLog: Found 1 hlogs to remove  out
>> of total 8; oldest outstanding sequenceid is 33526881 from region article
>>>> 
>> ,1221b0013c61fd7446cb822f698d0bcafa487342,1291872610788.bb2aa6ae8c694a096a00e7bb91fc46c3.
>>>> 2010-12-09 06:23:27,080 INFO
>> org.apache.hadoop.hbase.regionserver.wal.HLog: moving old hlog file
>> /hbase/.logs/06.hadoop.****,60020,1291864620851/10.99.97.106
>> %3A60020.1291872064424
>>>> whose highest sequenceid is 33506639 to /hbase/.oldlogs/10.99.97.106
>> %3A60020.1291872064424
>>>> 2010-12-09 06:25:39,162 WARN
>> org.apache.hadoop.hbase.regionserver.wal.HLog: IPC Server handler 12 on
>> 60020 took 119034ms appending an edit to hlog; editcount=3683
>>>> 2010-12-09 06:25:39,162 WARN org.apache.hadoop.hdfs.DFSClient:
>> DataStreamer Exception: java.io.IOException: Broken pipe
>>>>       at sun.nio.ch.FileDispatcher.write0(Native Method)
>>>>       at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:29)
>>>>       at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:100)
>>>>       at sun.nio.ch.IOUtil.write(IOUtil.java:71)
>>>>       at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:334)
>>>>       at
>> org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:55)
>>>>       at
>> org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142)
>>>>       at
>> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:146)
>>>>       at
>> org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:107)
>>>>       at
>> java.io.BufferedOutputStream.write(BufferedOutputStream.java:105)
>>>>       at java.io.DataOutputStream.write(DataOutputStream.java:90)
>>>>       at
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2509)
>>>> 2010-12-09 06:25:39,163 WARN org.apache.hadoop.hdfs.DFSClient:
>> DFSOutputStream ResponseProcessor exception  for block
>> blk_1141147913164105086_258331java.io.EOFException
>>>>       at java.io.DataInputStream.readFully(DataInputStream.java:180)
>>>>       at java.io.DataInputStream.readLong(DataInputStream.java:399)
>>>>       at
>> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.readFields(DataTransferProtocol.java:120)
>>>>       at
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2621)
>>>> 
>>>> 2010-12-09 06:25:39,163 INFO org.apache.hadoop.hdfs.DFSClient: Error
>> Recovery for block blk_1141147913164105086_258331 waiting for responder to
>> exit.
>>>> 2010-12-09 06:25:39,164 WARN org.apache.hadoop.hdfs.DFSClient:
>> DFSOutputStream ResponseProcessor exception  for block
>> blk_-7076574401277381592_258324java.io.EOFException
>>>>       at java.io.DataInputStream.readFully(DataInputStream.java:180)
>>>>       at java.io.DataInputStream.readLong(DataInputStream.java:399)
>>>>       at
>> org.apache.hadoop.hdfs.protocol.DataTransferProtocol$PipelineAck.readFields(DataTransferProtocol.java:120)
>>>>       at
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2621)
>>>> 
>>>> 2010-12-09 06:25:39,164 WARN org.apache.hadoop.hdfs.DFSClient: Error
>> Recovery for block blk_-7076574401277381592_258324 bad datanode[0]
>> 10.99.97.106:50010
>>>> 2010-12-09 06:25:39,164 INFO org.apache.zookeeper.ClientCnxn: Unable to
>> read additional data from server sessionid 0x12ca58c08ff004b, likely server
>> has closed socket, closing socket connection and attempting reconnect
>>>> 2010-12-09 06:25:39,164 WARN org.apache.hadoop.hdfs.DFSClient: Error
>> Recovery for block blk_-7076574401277381592_258324 in pipeline
>> 10.99.97.106:50010, 10.100.154.103:50010: bad datanode 10.99.97.106:50010
>>>> 2010-12-09 06:25:39,165 WARN org.apache.hadoop.hdfs.DFSClient: Error
>> Recovery for block blk_1141147913164105086_258331 bad datanode[0]
>> 10.99.97.106:50010
>>>> 2010-12-09 06:25:39,167 INFO org.apache.zookeeper.ClientCnxn: Unable to
>> read additional data from server sessionid 0x12ca58c08ff004c, likely server
>> has closed socket, closing socket connection and attempting reconnect
>>>> 2010-12-09 06:25:39,169 INFO org.apache.hadoop.ipc.Client: Retrying
>> connect to server: /10.100.154.103:50020. Already tried 0 time(s).
>>>> 2010-12-09 06:25:39,169 WARN org.apache.hadoop.hdfs.DFSClient: Error
>> Recovery for block blk_1141147913164105086_258331 in pipeline
>> 10.99.97.106:50010, 10.100.159.13:50010: bad datanode 10.99.97.106:50010
>>>> 2010-12-09 06:25:39,209 INFO
>> org.apache.hadoop.hbase.regionserver.HRegionServer: MSG_REGIONSERVER_STOP
>>>> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: Stopping
>> server on 60020
>>>> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC
>> Server handler 3 on 60020: exiting
>>>> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC
>> Server handler 9 on 60020: exiting
>>>> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC
>> Server handler 24 on 60020: exiting
>>>> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 1 on 60020: exiting
>>>> 2010-12-09 06:25:39,210 INFO org.apache.hadoop.ipc.HBaseServer: IPC
>> Server handler 15 on 60020: exiting
>>>> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 0 on 60020: exiting
>>>> 2010-12-09 06:25:39,214 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 3 on 60020: exiting
>>>> 2010-12-09 06:25:39,214 INFO org.apache.hadoop.ipc.HBaseServer: Stopping
>> IPC Server listener on 60020
>>>> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 2 on 60020: exiting
>>>> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 7 on 60020: exiting
>>>> 2010-12-09 06:25:39,213 INFO org.apache.hadoop.ipc.HBaseServer: IPC
>> Server handler 21 on 60020: exiting
>>>> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 4 on 60020: exiting
>>>> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: Stopping
>> IPC Server Responder
>>>> 2010-12-09 06:25:39,215 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 5 on 60020: exiting
>>>> 2010-12-09 06:25:39,216 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 8 on 60020: exiting
>>>> 2010-12-09 06:25:39,216 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 9 on 60020: exiting
>>>> 2010-12-09 06:25:39,216 INFO org.apache.hadoop.ipc.HBaseServer: PRI IPC
>> Server handler 6 on 60020: exiting
>>>> 2010-12-09 06:25:39,216 INFO
>> org.apache.hadoop.hbase.regionserver.HRegionServer: Stopping infoServer
>>>> 2010-12-09 06:25:39,223 WARN org.apache.hadoop.hdfs.DFSClient: Failed
>> recovery attempt #0 from primary datanode 10.100.159.13:50010
>>>> 
>>>> 
>>>> 
>>>> 
>> 
>> 

Reply via email to