Kazuki-san,

Setting the ZK timeout to a large value will stop the expirations but may not 
provide sufficiently fast failure detection for your use case of course.

However if even Ganglia stops working during a large mapreduce job, I think you 
need to question the adequacy of the network hardware.

   - Andy

> From: Kazuki Ohta <kazuki.o...@gmail.com>
> Subject: massive zk expirations under heavy network load
> To: user@hbase.apache.org
> Cc: kazuki.o...@gmail.com
> Date: Wednesday, April 20, 2011, 11:41 AM
> Hi,
> 
> I'm now using CDH3u0 at 16nodes cluster (hdp0-hdp15).
> The configuraiton is below.
> 
> hdp0: zk + master + region + nn + dn + jt + tt
> hdp1: zk + master + region + snn + dn + tt
> hdp2: zk + region + dn + tt
> hdp3 to hdp15: region + dn + tt
> 
> Usually, it works really well. But once the user throws MapReduce
> job which requires massive network transfer in the shuffle phase,
> the master got the zk session timeout exception, and fails-over to
> another master.
> 
> The problem is that shuffle network transfer dominates the switch,
> and important zk packets are not transferred properly at
> that time.
> 
> Even ganglia monitoring seems to stop at that time. And mr task
> attempts also got zk session timeouts and dies altogether (about
> 100 tasks dies at the same time. input and output are both hbase).
> 
> This is the potential problem running MapReduce job alongside
> with HBase. Does anyone know any good solution for this
> phenomenon?


Reply via email to