Go ahead and open a MR jira (would appreciate a patch too! ;) ). thanks, Arun
On Dec 20, 2011, at 2:55 PM, Patai Sangbutsarakum wrote: > Thanks again Arun, you save me again.. :-) > > This is a great starting point. for CPU and possibly Mem. > > For the IOPS, just would like to ask if the tasknode/datanode collect the > number > or we should dig into OS level.. like /proc/PID_OF_tt/io > ^hope this make sense > > -P > > On Tue, Dec 20, 2011 at 1:22 PM, Arun C Murthy <a...@hortonworks.com> wrote: >> Take a look at the JobHistory files produced for each job. >> >> With 0.20.205 you get CPU (slot millis). >> With 0.23 (alpha quality) you get CPU and JVM metrics (GC etc.). I believe >> you also get Memory, but not IOPS. >> >> Arun >> >> On Dec 20, 2011, at 1:11 PM, Patai Sangbutsarakum wrote: >> >>> Thanks for reply, but I don't think metric exposed to Ganglia would be >>> what i am really looking for.. >>> >>> what i am looking for is some kind of these (but not limit to) >>> >>> Job_xxxx_yyyy >>> CPU time: 10204 sec. <--aggregate from all tasknodes >>> IOPS: 2344 <-- aggregated from all datanode >>> MEM: 30G <-- aggregated >>> >>> etc, >>> >>> Job_aaa_bbb >>> CPU time: >>> IOPS: >>> MEM: >>> >>> Sorry for ambiguous question. >>> Thanks >>> >>> On Tue, Dec 20, 2011 at 12:47 PM, He Chen <airb...@gmail.com> wrote: >>>> You may need Ganglia. It is a cluster monitoring software. >>>> >>>> On Tue, Dec 20, 2011 at 2:44 PM, Patai Sangbutsarakum < >>>> silvianhad...@gmail.com> wrote: >>>> >>>>> Hi Hadoopers, >>>>> >>>>> We're running Hadoop 0.20 CentOS5.5. I am finding the way to collect >>>>> CPU time, memory usage, IOPS of each hadoop Job. >>>>> What would be the good starting point ? document ? api ? >>>>> >>>>> Thanks in advance >>>>> -P >>>>> >>