Reposting - Hadoop log timestamps & file timestamps not same as system time.

2011-01-10 Thread Ravi Phulari
Hello Friends, I am seeing Hadoop log timestamps & file timestamps not same as system time. I found that this problem was discussed on on mailing list earlier but there was no solution posted. http://www.mail-archive.com/hdfs-user@hadoop.apache.org/msg00310.html I am wondering what is causing

Re: hadoop single user setup

2011-01-10 Thread Mapred Learn
Thanks. I tried start-dfs.sh and namenode service starts but after sometime I see Shutdown message in namenode log: 2011-01-10 14:52:47,568 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 on 9000: starting^M 2011-01-10 14:52:47,568 INFO org.apache.hadoop.ipc.Server: IPC Server handler 5 on

Re: hadoop single user setup

2011-01-10 Thread Aaron Eng
When you run the Hadoop CLI command it spawns a java process which in turn tries to connect to the namenode service. In this case, your client is trying to reach the namenode at "localhost" on TCP port 9000. That connection is failing. The likely reason is that your namenode service is not runnin

hadoop single user setup

2011-01-10 Thread Mapred Learn
hi, I am a newbie and am trying to setup hadoop in single user setup on my windows 7 machine. I followed steps at: http://hadoop.apache.org/common/docs/current/single_node_setup.html#L... but i keep on getting error: $ bin/

Re: Choosing number of map/reduce slots (with hyperthreading)

2011-01-10 Thread Adam Phelps
I'll attempt some tests on this later this week and report back once I've done so. - Adam On 1/10/11 12:06 AM, Eric wrote: With hyperthreading, the cpu tries to prevent being idle by running that extra thread when it has some cycles left. It can do so cheaply, since hyperthreading is much fast

Hadoop log timestamps & file timestamps not same as system time.

2011-01-10 Thread Ravi Phulari
Hello Friends, I am seeing Hadoop log timestamps & file timestamps not same as system time. I found that this problem was discussed on on mailing list earlier but there was no solution posted. http://www.mail-archive.com/hdfs-user@hadoop.apache.org/msg00310.html I am wondering what is causing

map reduce block size

2011-01-10 Thread sharath maddineni
Hi all, Can any please tell me how to find/change the Block for the Terasort implementation in Hadoop .21 Thanks,Sharath

Re: how to read map_0.out

2011-01-10 Thread Owen O'Malley
The intermediate files are called IFiles. The format is trivial and you can read the code to see it. The only tricky bit is that you effectively have N IFiles concatenated together (one per a reduce). -- Owen

Re: JobClient equivalent in new mapreduce api

2011-01-10 Thread Sarthak Dudhara
ok. sure. Thats useful. I will try that out. Sarthak Dudhara On Sun, Jan 9, 2011 at 10:25 AM, Harsh J wrote: > On Sun, Jan 9, 2011 at 10:57 PM, Sarthak Dudhara > wrote: > > Hi Harsh, > > Thanks for the reply... > > I am currently using the job.submit(). Is there a way to track the job > > st

how to read map_0.out

2011-01-10 Thread Ferdy Galema
I stopped a job that was running very slowly, it was running in it's reduce (phase:reduce) part. However, I still want it's output and I cannot run this job again. So I have to stick with the intermediate files. I have a 30GB file map_0.out (found in reducer jobcache) and I want to read it's c

Re: Choosing number of map/reduce slots (with hyperthreading)

2011-01-10 Thread Eric
With hyperthreading, the cpu tries to prevent being idle by running that extra thread when it has some cycles left. It can do so cheaply, since hyperthreading is much faster than context switching. So as Arun suggests, it probably won't hurt as long as you have enough memory in your nodes. Your cpu