Hey,

Looks like DFS errors.  The "bad firstAck" is a sign of a datanode
problem.  Perhaps your ulimit is too low - 2048 xciever count but only
1024 sockets?  Sounds suspicious to me.

On Thu, Apr 1, 2010 at 1:19 PM, Oded Rosen <o...@legolas-media.com> wrote:
> **Hi all,
>
> I have a problem with a massive HBase loading job.
> It is from raw files to hbase, through some mapreduce processing +
> manipulating (so loading direcly to files will not be easy).
>
> After some dozen million successful writes - a few hours of load - some of
> the regionservers start to die - one by one - until the whole cluster is
> kaput.
> The hbase master sees a "znode expired" error each time a regionserver
> falls. The regionserver errors are attached.
>
> Current configurations:
> Four nodes - one namenode+master, three datanodes+regionservers.
> dfs.datanode.max.xcievers: 2047
> ulimit: 1024
> servers: fedora
> hadoop-0.20, hbase-0.20, hdfs (private servers, not on ec2 or anything).
>
>
> *The specific errors from the regionserver log (from <IP6>, see comment):*
>
> 2010-04-01 11:36:00,224 WARN org.apache.hadoop.hdfs.DFSClient:
> DFSOutputStream ResponseProcessor exception  for block
> blk_7621973847448611459_244908java.io.IOException: Bad response 1 for block
> blk_7621973847448611459_244908 from datanode <IP2>:50010
>    at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$ResponseProcessor.run(DFSClient.java:2423)
>
> *after that, some of this appears:*
>
> 2010-04-01 11:36:20,602 INFO org.apache.hadoop.hdfs.DFSClient: Exception in
> createBlockOutputStream java.io.IOException: Bad connect ack with
> firstBadLink <IP2>:50010
> 2010-04-01 11:36:20,602 INFO org.apache.hadoop.hdfs.DFSClient: Abandoning
> block blk_4280490438976631008_245009
>
> *and the FATAL:*
>
> 2010-04-01 11:36:32,634 FATAL org.apache.hadoop.hbase.regionserver.HLog:
> Could not append. Requesting close of hlog
> java.io.IOException: Bad connect ack with firstBadLink <IP2>:50010
>    at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.createBlockOutputStream(DFSClient.java:2872)
>    at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2795)
>    at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2078)
>    at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2264)
>
> *this FATAL error appears many times until this one kicks in:*
>
> 2010-04-01 11:38:57,281 FATAL
> org.apache.hadoop.hbase.regionserver.MemStoreFlusher: Replay of hlog
> required. Forcing server shutdown
> org.apache.hadoop.hbase.DroppedSnapshotException: region: .META.,,1
>    at
> org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:977)
>    at
> org.apache.hadoop.hbase.regionserver.HRegion.flushcache(HRegion.java:846)
>    at
> org.apache.hadoop.hbase.regionserver.MemStoreFlusher.flushRegion(MemStoreFlusher.java:241)
>    at
> org.apache.hadoop.hbase.regionserver.MemStoreFlusher.run(MemStoreFlusher.java:149)
> Caused by: java.io.IOException: Bad connect ack with firstBadLink
> <IP2>:50010
>    at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.createBlockOutputStream(DFSClient.java:2872)
>    at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.nextBlockOutputStream(DFSClient.java:2795)
>    at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.access$2000(DFSClient.java:2078)
>    at
> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2264)
>
> *(then the regionserver starts closing itself)*
>
> The regionserver on <IP6> was shut down, but problems are corellated with
> <IP2> (notice the ip in the error msgs). <IP2> was also considered a dead
> node after these errors, according to the hadoop namenode web ui.
> I think this is an hdfs failure, rather then hbase/zookeeper (although it is
> probably because of hbase high load...).
>
> On the datanodes, once in a while I had:
>
> 2010-04-01 11:24:59,265 ERROR
> org.apache.hadoop.hdfs.server.datanode.DataNode:
> DatanodeRegistration(<IP2>:50010,
> storageID=DS-1822315410-<IP2>-50010-1266860406782, infoPort=50075,
> ipcPort=50020):DataXceiver
>
> but these errors occured at different times, and not even around crashes. No
> fatal errors found on the datanode log (but it still crashed).
>
> I haven't seen this exact error on the web (only similar ones);
> This guy (http://osdir.com/ml/hbase-user-hadoop-apache/2009-02/msg00186.html)
> had a similar problem, but not exactly the same.
>
> Any ideas?
> thanks,
>
> --
> Oded
>

Reply via email to