Re: Accessing Job counters displayed in WEB GUI in Hadoop Code

2011-12-11 Thread Harsh J
Hej again, You can get task data via the completion events call from your client: http://hadoop.apache.org/common/docs/r0.20.1/api/org/apache/hadoop/mapreduce/Job.html#getTaskCompletionEvents(int) They should carry the data you seek. On Mon, Dec 12, 2011 at 4:47 AM, W.P. McNeill wrote: > You ca

Re: HDFS Backup nodes

2011-12-11 Thread M. C. Srivas
You are out of luck if you don't want to use NFS, and yet want redundancy for the NN. Even the new "NN HA" work being done by the community will require NFS ... and the NFS itself needs to be HA. But if you use a Netapp, then the likelihood of the Netapp crashing is lower than the likelihood of a

RE: Grouping nodes into different racks in Hadoop Cluster

2011-12-11 Thread Devaraj K
Hi Arun, You can enable rack awareness for your hadoop cluster by configuring the "topology.script.file.name" property. Please go through this link for more details about rack awareness. http://hadoop.apache.org/common/docs/r0.19.2/cluster_setup.html#Hadoop+Rack+ Awareness Devaraj K

Hadoop-Snappy is integrated into Hadoop Common (JUN 2011).

2011-12-11 Thread Jinyan Xu
This project is integrated into Hadoop Common (JUN 2011). Hadoop-Snappy can be used as an add-on for recent (released) versions of Hadoop that do not provide Snappy Codec support yet. Hadoop-Snappy is being kept in synch with Hadoop Common. what's this meaning then how to enable snappy on

RE: Accessing Job counters displayed in WEB GUI in Hadoop Code

2011-12-11 Thread Devaraj K
Hi Arun, You can get the Counters object from the job after completion and then find the counter whichever you want in Counters object using findCounter api. Please find the sample snippet for accessing the counter after Job completion Configuration conf = new Configuration(); Cluster

Re: Namenode does not start and generates no error messages

2011-12-11 Thread Hemanth Makkapati
You need to format the namenode before you start its daemon. Try "*./hadoop namenode -format* " and then run start-dfs.sh On Sun, Dec 11, 2011 at 6:25 PM, W.P. McNeill wrote: > I am trying to run Hadoop as single-node cluster on OS X 10.7 (Lion), > Hadoop 0.20.203. The namenode does not start a

Namenode does not start and generates no error messages

2011-12-11 Thread W.P. McNeill
I am trying to run Hadoop as single-node cluster on OS X 10.7 (Lion), Hadoop 0.20.203. The namenode does not start and gives no indication of what is wrong. > start-dfs.sh starting namenode, logging to /tmp/hadoop/logs/hadoop-williammcneill-namenode-William-McNeills-MacBook.local.out localhost: st

Re: Accessing Job counters displayed in WEB GUI in Hadoop Code

2011-12-11 Thread W.P. McNeill
You can read counter values from the Job.getCounters API (http://hadoop.apache.org/common/docs/r0.20.1/api/org/apache/hadoop/mapreduce/Job.html). I'm not sure about the other information like execution times. I've been wondering that myself. On 12/10/11, ArunKumar wrote: > Hai guys ! > > Can i ac

Re: Running a job continuously

2011-12-11 Thread Inder Pall
have you looked at kafka. it provides a streaming view of data stream. flume at the moment is getting rewritten as flume ng On Dec 6, 2011 4:28 PM, "Praveen Sripati" wrote: > If the requirement is for real time data processing, using Flume > will not suffice as there is a time lag between the col