Re: Monitoring Hadoop Cluster

2012-05-10 Thread Manu S
Thanks a lot Junyong On Fri, May 11, 2012 at 11:15 AM, JunYong Li wrote: > Each has its own merits. > http://developer.yahoo.com/hadoop/tutorial/module7.html#monitoring > > 2012/5/11 Manu S > > > Hi All, > > > > Which is the best monitoring tool for Hadoop cluster monitoring? Ganglia > or > > N

Re: loading a class at job runtime

2012-05-10 Thread JunYong Li
i have encounter CLassNotFound mant times and the reason is the class dose really not exits in the classpath. the computer can't go wrong, so calm down, look it over 2012/5/11 Pat Ferrel > Thanks, that's where I started but #3, which has several methods I tried, > did not work and the reference

Re: Exception closing file

2012-05-10 Thread JunYong Li
The client is done writing data to the given filename, and would like to complete it. The function returns whether the file has been closed successfully. If the function returns false, the caller should try again. A call to complete() will not return true until all the file's blocks have been repli

Re: Monitoring Hadoop Cluster

2012-05-10 Thread JunYong Li
Each has its own merits. http://developer.yahoo.com/hadoop/tutorial/module7.html#monitoring 2012/5/11 Manu S > Hi All, > > Which is the best monitoring tool for Hadoop cluster monitoring? Ganglia or > Nagios? > > Thanks, > Manu S > -- Regards Junyong

Re: Monitoring Hadoop Cluster

2012-05-10 Thread Manu S
Cool :) Thanks for your quick reply. On Fri, May 11, 2012 at 11:07 AM, v v wrote: > Ganglia won't send alerts, but it is a very good monitoring tool for > viewing the metrics (hadoop has support for emitting metrics to > ganglia). > > If you need alerting (eg, dfs namespace > 90%) then you will

Exception closing file

2012-05-10 Thread Aleksandr Elbakyan
Hello, I am running the job, when in two similar size clusters, in one it is running fine on other gives fallowing exceptions. I checked the safe mode is off. Not sure what do I need fix to avoid this? 2012-05-10 00:38:02,637 ERROR org.apache.hadoop.hdfs.DFSClient: Exception closing file /us

Re: reducer out of memory?

2012-05-10 Thread Yang
thanks, let me run more of this with the settings provided later in this thread and provide the details On Wed, May 9, 2012 at 10:12 PM, Harsh J wrote: > Can you share your job details (or a sample reducer code) and also > share your exact error? > > If you are holding reducer provided values/key

Re: reducer out of memory?

2012-05-10 Thread Yang
thanks, let me try this On Wed, May 9, 2012 at 11:27 PM, Zizon Qiu wrote: > try setting a lower value for mapred.job.shuffle.input.buffer.percent . > the reducer used it to decide whether use in-memory shuffle. > the default value is 0.7,meaning 70% of the "memory" are used as shuffle > buffer.

Re: High load on datanode startup

2012-05-10 Thread Raj Vishwanathan
Darrell Are the new dn,nn and mapred directories on the same physical disk? Nothing on NFS , correct? Could you be having some hardware issue? Any clue in /var/log/messages or dmesg? A non responsive system indicates a CPU that is really busy either doing something or waiting for something and

Re: loading a class at job runtime

2012-05-10 Thread Pat Ferrel
Thanks, that's where I started but #3, which has several methods I tried, did not work and the reference claims they are not currently supported in cloudera's implementation. Not sure why so I don't know if it's a hadoop thing or not. I ended up putting the classes in the job jar (#2 of cloude

Re: SQL analysis

2012-05-10 Thread Shi Yu
If the analysis you mention is to create "view" of multiple tables. Once your data is sorted by the keys in HDFS. You could try Map Side join or Reducer Side join in Hadoop to generate the "view" of your data (same keys of multiple data sets are combined). There are many code samples web, play

How to start up datanode with kerberos?

2012-05-10 Thread shixing
Hi,all: Now I want to setup the security with hbase by kerberos. As I know, the hbase's ugi is based on the hadoop UserGroupInformation without parameter "hadoop.job.ugi" after 0.20.2. So when I use the cdh3u3, the ugi can be generated by two authentication : simple or kerberos. Firstly

RE: SQL analysis

2012-05-10 Thread karanveer.singh
I am more worried about the analysis assuming this data is in HDFS. -Original Message- From: Shi Yu [mailto:sh...@uchicago.edu] Sent: 10 May 2012 18:58 To: common-user@hadoop.apache.org Subject: RE: SQL analysis Flume might be suitable for your case. https://cwiki.apache.org/FLUME/

RE: SQL analysis

2012-05-10 Thread Shi Yu
Flume might be suitable for your case. https://cwiki.apache.org/FLUME/ Shi

RE: SQL analysis

2012-05-10 Thread karanveer.singh
Our focus as of now is on batch queries. And keen to explore on the approach to get to the path analysis. Regards, Karanveer -Original Message- From: Shi Yu [mailto:sh...@uchicago.edu] Sent: 10 May 2012 17:02 To: common-user@hadoop.apache.org Subject: Re: SQL analysis It depends on y

Re: SQL analysis

2012-05-10 Thread Shi Yu
It depends on your use case, for example, query only or you have requirement of real time insert and update. The solutions can be different. You might need consider HBase, Cassandra or tools like Flume.

Re: job Killing itself upon some condition

2012-05-10 Thread tousif
On Thursday 10 May 2012 04:25 PM, Devaraj k wrote: Hi Tousif, You can kill the Running Job using the killJob() client API. If you want to kill the job itself, you can get the job id using task attempt id from map() or reduce() functions, and you can invoke the killJob() API based on you

RE: job Killing itself upon some condition

2012-05-10 Thread Devaraj k
Hi Tousif, You can kill the Running Job using the killJob() client API. If you want to kill the job itself, you can get the job id using task attempt id from map() or reduce() functions, and you can invoke the killJob() API based on your condition. Thanks Devaraj __

Re: High load on datanode startup

2012-05-10 Thread Darrell Taylor
On Thu, May 10, 2012 at 9:33 AM, Todd Lipcon wrote: > That's real weird.. > > If you can reproduce this after a reboot, I'd recommend letting the DN > run for a minute, and then capturing a "jstack " as well as > the output of "top -H -p -b -n 5" and send it to the list. What I did after the r

Re: Consistent Checksum error using SequenceFileInputFormat against /content & /parse_text folders output by Nutch.

2012-05-10 Thread Ali Safdar Kureishy
Actually, the ChecksumError happens every time, but it can happen at different points of the execution...sometimes at the beginning, and sometimes at the tail end of the map phase. Hoping to hear from someone with a workaround... Regards, Safdar On Thu, May 10, 2012 at 7:59 AM, Ali Safdar Kurei

job Killing itself upon some condition

2012-05-10 Thread tousif
Hi, i have scheduled a map reduce job and i want this map reduce job to kill itself upon some condition. overall it should fail. Here I'm not using run method of tool runner.

Re: DataNodeRegistration problem

2012-05-10 Thread Harsh J
Fourie, Good to know. Just for the record, do you remember the version of OpenJDK that was installed? So that we can mark that as unusable in our docs/wiki. On Thu, May 10, 2012 at 12:22 PM, Fourie Joubert wrote: > Hi > > Yes - that was indeed the problem... > > I cleaned up the Java's on all th

Re: High load on datanode startup

2012-05-10 Thread Todd Lipcon
That's real weird.. If you can reproduce this after a reboot, I'd recommend letting the DN run for a minute, and then capturing a "jstack " as well as the output of "top -H -p -b -n 5" and send it to the list. What JVM/JDK are you using? What OS version? -Todd On Wed, May 9, 2012 at 11:57 PM,

SQL analysis

2012-05-10 Thread karanveer.singh
We are looking at doing some initial analysis on SQL text info within the query runs to come up with some kind of path output to depict how various tables are linked to each other. For example. A 'view' might be a join from two table's top of the hierarchy and in turn might be creating some new