Cool, is there any documentation on how to use the JMX stuff to get monitoring data?
Cheers, Ivan On 10/17/11 6:04 PM, "Rajiv Chittajallu" <raj...@yahoo-inc.com> wrote: >If you are running > 0.20.204 >http://phanpy-nn1.hadoop.apache.org:50070/jmx?qry=Hadoop:service=NameNode, >name=NameNodeInfo > > >ivan.nov...@emc.com wrote on 10/17/11 at 09:18:20 -0700: >>Hi Harsh, >> >>I need access to the data programatically for system automation, and >>hence >>I do not want a monitoring tool but access to the raw data. >> >>I am more than happy to use an exposed function or client program and not >>an internal API. >> >>So i am still a bit confused... What is the simplest way to get at this >>raw disk usage data programmatically? Is there a HDFS equivalent of du >>and df, or are you suggesting to just run that on the linux OS (which is >>perfectly doable). >> >>Cheers, >>Ivan >> >> >>On 10/17/11 9:05 AM, "Harsh J" <ha...@cloudera.com> wrote: >> >>>Uma/Ivan, >>> >>>The DistributedFileSystem class explicitly is _not_ meant for public >>>consumption, it is an internal one. Additionally, that method has been >>>deprecated. >>> >>>What you need is FileSystem#getStatus() if you want the summarized >>>report via code. >>> >>>A job, that possibly runs "du" or "df", is a good idea if you >>>guarantee perfect homogeneity of path names in your cluster. >>> >>>But I wonder, why won't using a general monitoring tool (such as >>>nagios) for this purpose cut it? What's the end goal here? >>> >>>P.s. I'd moved this conversation to hdfs-user@ earlier on, but now I >>>see it being cross posted into mr-user, common-user, and common-dev -- >>>Why? >>> >>>On Mon, Oct 17, 2011 at 9:25 PM, Uma Maheswara Rao G 72686 >>><mahesw...@huawei.com> wrote: >>>> We can write the simple program and you can call this API. >>>> >>>> Make sure Hadoop jars presents in your class path. >>>> Just for more clarification, DN will send their stats as parts of >>>>hertbeats, So, NN will maintain all the statistics about the diskspace >>>>usage for the complete filesystem and etc... This api will give you >>>>that >>>>stats. >>>> >>>> Regards, >>>> Uma >>>> >>>> ----- Original Message ----- >>>> From: ivan.nov...@emc.com >>>> Date: Monday, October 17, 2011 9:07 pm >>>> Subject: Re: Is there a good way to see how full hdfs is >>>> To: common-user@hadoop.apache.org, mapreduce-u...@hadoop.apache.org >>>> Cc: common-...@hadoop.apache.org >>>> >>>>> So is there a client program to call this? >>>>> >>>>> Can one write their own simple client to call this method from all >>>>> diskson the cluster? >>>>> >>>>> How about a map reduce job to collect from all disks on the cluster? >>>>> >>>>> On 10/15/11 4:51 AM, "Uma Maheswara Rao G 72686" >>>>> <mahesw...@huawei.com>wrote: >>>>> >>>>> >/** Return the disk usage of the filesystem, including total >>>>> capacity,> * used space, and remaining space */ >>>>> > public DiskStatus getDiskStatus() throws IOException { >>>>> > return dfs.getDiskStatus(); >>>>> > } >>>>> > >>>>> >DistributedFileSystem has the above API from java API side. >>>>> > >>>>> >Regards, >>>>> >Uma >>>>> > >>>>> >----- Original Message ----- >>>>> >From: wd <w...@wdicc.com> >>>>> >Date: Saturday, October 15, 2011 4:16 pm >>>>> >Subject: Re: Is there a good way to see how full hdfs is >>>>> >To: mapreduce-u...@hadoop.apache.org >>>>> > >>>>> >> hadoop dfsadmin -report >>>>> >> >>>>> >> On Sat, Oct 15, 2011 at 8:16 AM, Steve Lewis >>>>> >> <lordjoe2...@gmail.com> wrote: >>>>> >> > We have a small cluster with HDFS running on only 8 nodes - I >>>>> >> believe that >>>>> >> > the partition assigned to hdfs might be getting full and >>>>> >> > wonder if the web tools or java api havew a way to look at free >>>>> >> space on >>>>> >> > hdfs >>>>> >> > >>>>> >> > -- >>>>> >> > Steven M. Lewis PhD >>>>> >> > 4221 105th Ave NE >>>>> >> > Kirkland, WA 98033 >>>>> >> > 206-384-1340 (cell) >>>>> >> > Skype lordjoe_com >>>>> >> > >>>>> >> > >>>>> >> > >>>>> >> >>>>> > >>>>> >>>>> >>>> >>> >>> >>> >>>-- >>>Harsh J >>> >>