Re: [ mapreduce ] Re: hadoop-1.0.0 and errors with log.index

2012-01-31 Thread Marcin Cylke
On 31.01.2012 19:32, Arun C Murthy wrote: Anything in TaskTracker logs ? Actually I've found something like this in jobs log. When running wordcount I got it in a file: ${result_dir}/_logs/history/job_201201311241_0008_1328055139152_hdfs_word+count MapAttempt TASK_TYPE="SETUP" TASKID="task_

Subscription Request

2012-01-31 Thread Ruchita Rathi

Re: Best practices for hadoop shuffling/tunning ?

2012-01-31 Thread Arun C Murthy
Moving to mapreduce-user@, bcc common-user@. Please use project specific lists. Your io.sort.mb is too high. You only have 1G of heap for the map. Reduce parallel copies is too high too. On Jan 30, 2012, at 4:50 AM, praveenesh kumar wrote: > Hey guys, > > Just wanted to ask, are there any sort

Re: Reduce shuffle data transfer takes excessively long

2012-01-31 Thread Robert Evans
If just changing the buffer to 4k makes a big difference could you at a minimum file a JIRA to change that buffer size? I know that it is not a final fix but it sure seems like a very nice Band-Aid to put on until we can get to the root of the issues. --Bobby Evans On 1/27/12 9:23 PM, "Sven G

Re: hadoop-1.0.0 and errors with log.index

2012-01-31 Thread Koji Noguchi
On our cluster, it usually happen when jvm crash with invalid jvm params or jni crashing at init phase. stderr/stdout files are created but log.index does not exist when this happens. We should fix this. Koji On 1/31/12 10:49 AM, "Markus Jelsma" wrote: > Yes, the stacktrace in my previous m

Re: hadoop-1.0.0 and errors with log.index

2012-01-31 Thread Markus Jelsma
Yes, the stacktrace in my previous message is from the task tracker. It seems to happen when there is no data locality for the mapper and it needs to get it from some other datanode. The number of failures is the same as the number of rack-local mappers. > Anything in TaskTracker logs ? > > On

Re: hadoop-1.0.0 and errors with log.index

2012-01-31 Thread Arun C Murthy
Anything in TaskTracker logs ? On Jan 31, 2012, at 10:18 AM, Markus Jelsma wrote: > In our case, which seems to be the same problem, the web UI does not show > anything useful except the first line of the stack trace: > > 2012-01-03 21:16:27,256 WARN org.apache.hadoop.mapred.TaskLog: Failed to

Re: hadoop-1.0.0 and errors with log.index

2012-01-31 Thread Markus Jelsma
In our case, which seems to be the same problem, the web UI does not show anything useful except the first line of the stack trace: 2012-01-03 21:16:27,256 WARN org.apache.hadoop.mapred.TaskLog: Failed to retrieve stdout log for task: attempt_201201031651_0008_m_000233_0 Only the task tracker lo

Re: hadoop-1.0.0 and errors with log.index

2012-01-31 Thread Arun C Murthy
Actually, all that is telling you is that the task failed and the job-client couldn't display the logs. Can you check the JT web-ui and see why the task failed ? If you don't see anything there, you can try see the TaskTracker logs on the node on which the task ran. Arun On Jan 31, 2012, at 3

Re: Problem with Symlinks

2012-01-31 Thread Harsh J
Glad to know you fixed it! I reckon 1.0 changed much scripts and tarball structure and its probably HADOOP_PREFIX instead now. If you feel this may help others as well, please also file a JIRA at https://issues.apache.org/jira/browse/HADOOP with a patch. Thanks! On Tue, Jan 31, 2012 at 9:57 PM, M

Re: Problem with Symlinks

2012-01-31 Thread Mohamed Riadh Trad
HADOOP_HOME is deprecated in hadoop 1.0. I solved my problem by changing the libexec/hadoop-config.sh: this="${BASH_SOURCE-$0}" #echo $this -common_bin=$(cd -P -- "$(dirname -- "$this")" && pwd -P) to +common_bin=$(cd -L -- "$(dirname -- "$this")" && pwd -L) Bests, Trad Mohamed Riadh, M.Sc,

Re: Why it don't print anything into part-00000 file

2012-01-31 Thread Steve Lewis
1 run locally ether problem is small and until you can run a small local problem do not mover to the cluster. 2 make sure ether mapper writes something set a breakpoint at the write call. 3 make sure the reducer reads something. 4 add write call s in the reducers setup and cleanup calls until somet

Re: [ mapreduce ] Re: hadoop-1.0.0 and errors with log.index

2012-01-31 Thread Marcin Cylke
On 31/01/12 12:48, Markus Jelsma wrote: I've seen that the number of related failures is almost always the same as the number of rack-local mappers. Do you see this as well? Yes, it seems that way. Marcin

Re: hadoop-1.0.0 and errors with log.index

2012-01-31 Thread Markus Jelsma
I've seen that the number of related failures is almost always the same as the number of rack-local mappers. Do you see this as well? On Tuesday 31 January 2012 12:21:44 Marcin Cylke wrote: > Hi > > I've upgraded my hadoop cluster to version 1.0.0. The upgrade process > went relatively smoothly

Re: Why it don't print anything into part-00000 file

2012-01-31 Thread Luiz Antonio Falaguasta Barbosa
Ronald, It is the directory where _SUCCESS and part-0 are generated, but they are empty. In my Eclipse, I put in 'Run Configurations' the following, in the tab Arguments: input.small out.ivory.small (as the file where is words and the directory where Hadoop should write results, respectively)

hadoop-1.0.0 and errors with log.index

2012-01-31 Thread Marcin Cylke
Hi I've upgraded my hadoop cluster to version 1.0.0. The upgrade process went relatively smoothly but it rendered the cluster inoperable due to errors in jobtrackers operation: # in job output Error reading task outputhttp://hadoop4:50060/tasklog?plaintext=true&attemptid=attempt_201201311241_000