Re: How to find out whether a node is Overloaded from Cpu utilization ?

2012-01-18 Thread John Meagher
The problem I've run into more than memory is having the system CPU time get out of control. My guess is that the threshold for what is considered overloaded is going to be dependent on your system setup, what you're running on it, and what bounds your jobs. On Tue, Jan 17, 2012 at 22:06,

RE: How to find out whether a node is Overloaded from Cpu utilization ?

2012-01-18 Thread Bill Brune
Hi, Memory loading in most linux distro's is not readily available from top or the usual suspects, in fact looking at top is rather misleading. Linux can run just fine with committed memory greater than 100%, what you want to look at is the % of committed memory relative to the total memery.

RE: How to find out whether a node is Overloaded from Cpu utilization ?

2012-01-17 Thread Bill Brune
Hi, The significant factor in cluster loading is memory, not CPU. Hadoop views the cluster only with respect to memory and cares not about CPU utilization or Disk saturation. If you run too many TaskTrackers, you risk memory overcommit where the Linux OOM will come out of the closet and

RE: How to find out whether a node is Overloaded from Cpu utilization ?

2012-01-17 Thread ArunKumar
Guys ! So can i say that if memory usage is more than say 90 % the node is overloaded. If so, what can be that threshold percent value or how can we find it ? Arun -- View this message in context:

Re: How to find out whether a node is Overloaded from Cpu utilization ?

2012-01-16 Thread Amandeep Khurana
Arun, I don't think you'll hear a fixed number. Having said that, I have seen CPU being pegged at 95% during jobs and the cluster working perfectly fine. On the slaves, if you have nothing else going on, Hadoop only has TaskTrackers and DataNodes. Those two daemons are relatively light weight in