RE: Task process exit with nonzero status of 134

2011-12-27 Thread anthony garnier

Alex
I don't know why but now I got other error code : 

On Jobtracker : 
java.lang.Throwable: Child Error
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
Caused by: java.io.IOException: Task process exit with nonzero status of 1.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)

On taskTracker : 

2011-12-27 15:14:50,663 INFO org.apache.hadoop.mapred.DefaultTaskController: 
Output from DefaultTaskController's launchTask follows:
2011-12-27 15:14:50,663 INFO org.apache.hadoop.mapred.TaskController:
2011-12-27 15:14:50,664 INFO org.apache.hadoop.mapred.JvmManager: JVM Not 
killed jvm_201112271514_0001_m_-1820714542 but just removed
2011-12-27 15:14:50,664 INFO org.apache.hadoop.mapred.JvmManager: JVM : 
jvm_201112271514_0001_m_-1820714542 exited with exit code 1. Number of tasks it 
ran: 0
2011-12-27 15:14:50,666 WARN org.apache.hadoop.mapred.TaskRunner: 
attempt_201112271514_0001_m_02_0 : Child Error
java.io.IOException: Task process exit with nonzero status of 1.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
2011-12-27 15:14:53,676 INFO org.apache.hadoop.mapred.TaskTracker: addFreeSlot 
: current free slots : 2


From: sokar6...@hotmail.com
To: wget.n...@googlemail.com; common-user@hadoop.apache.org
Subject: RE: Task process exit with nonzero status of 134
Date: Tue, 27 Dec 2011 13:54:22 +







Alex

-XX:+UseCompressedOops option is the default in 1.6.0_24 and above on 64 bit 
JVMs (http://wiki.apache.org/hadoop/HadoopJavaVersions)
Anyway, I tested it but same result.
Is it wise to test hadoop with the new jdk7_2 ?

Anthony


 Date: Tue, 27 Dec 2011 13:47:03 +0100
 Subject: Re: Task process exit with nonzero status of 134
 From: wget.n...@googlemail.com

 To: sokar6...@hotmail.com
 
 Anthony,
 
 134 depends mostly on JRE (Bug) or defect RAM. _30 is the newest
 update, could be a bug inside. Can you test SE 7u2?
 Todd mentioned in a older post to use -XX:+UseCompressedOops
 (hadoop-env.sh). Another option could be to take a closer look at
 garbage collection with compressed option.
 
 - Alex
 
 On Tue, Dec 27, 2011 at 1:20 PM, anthony garnier sokar6...@hotmail.com 
 wrote:
  Alex,
 
  Memory available on namenode / Jobtracker :
  Tasks: 435 total,   1 running, 434 sleeping,   0 stopped,   0 zombie
  Cpu(s):  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,
  0.0%st
  Mem: 15360M total,11609M used, 3750M free,  311M buffers
  Swap: 2047M total,1M used, 2046M free, 8833M cached
 
  On datanode / Tasktracker :
  top - 13:15:27 up 6 days, 21:11,  1 user,  load average: 0.03, 0.28, 0.26
  Tasks: 377 total,   1 running, 376 sleeping,   0 stopped,   0 zombie
  Cpu(s):  0.0%us,  0.0%sy,  0.0%ni, 99.9%id,  0.0%wa,  0.0%hi,  0.0%si,
  0.0%st
  Mem: 72373M total, 4321M used,68051M free,  348M buffers
  Swap: 2047M total,0M used, 2047M free, 2771M cached
 
  src/mapred/mapred-default.xml :
  property
namemapred.child.java.opts/name
value-Xmx200m/value
  /property
 
 
  So there should be enough memory
 
  Anthony
 
 
  Date: Tue, 27 Dec 2011 11:58:46 +0100
 
  Subject: Re: Task process exit with nonzero status of 134
  From: wget.n...@googlemail.com
  To: sokar6...@hotmail.com
 
 
  Anthony,
 
  How much memory you have available? Did the system going into swap?
 
  - Check mapred.map.child.java.opts (mapred.xml) for given MaxSize (xmx).
  - what says top -Hc?
 
  - Alex
 
  On Tue, Dec 27, 2011 at 11:49 AM, anthony garnier sokar6...@hotmail.com
  wrote:
   Hi,
  
   I got Nothing in the dmesg
   I've checked the Tasktracker and this is what I got :
  
   /
   STARTUP_MSG: Starting TaskTracker
   STARTUP_MSG:   host = ylal2960.inetpsa.com/10.68.217.86
   STARTUP_MSG:   args = []
   STARTUP_MSG:   version = 0.20.203.0
   STARTUP_MSG:   build =
  
   http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-security-203
   -r 1099333; compiled by 'oom' on Wed May  4 07:57:50 PDT 2011
   /
   2011-12-23 15:11:02,275 INFO
   org.apache.hadoop.metrics2.impl.MetricsConfig:
   loaded properties from hadoop-metrics2.properties
   2011-12-23 15:11:02,330 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   MetricsSystem,sub=Stats registered.
   2011-12-23 15:11:02,331 INFO
   org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
   period
   at 10 second(s).
   2011-12-23 15:11:02,331 INFO
   org.apache.hadoop.metrics2.impl.MetricsSystemImpl: TaskTracker metrics
   system started
   2011-12-23 15:11:02,597 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   ugi
   registered.
   2011-12-23 15:11:02,738 INFO org.mortbay.log: Logging to
   org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
   org.mortbay.log.Slf4jLog
   2011-12-23 15:11:02,803 INFO

RE: Task process exit with nonzero status of 134

2011-12-27 Thread anthony garnier

Alex

java -version -Xmx1024m is running fine

I was also able to create directory in /users/hadoop-0.20.203.0/log/userlogs

But anyway I've doe a fresh install with the 0.20.205.0 version and the 
mapreduce job completed successfully !

Thank you for your help.

Anthony

 Date: Tue, 27 Dec 2011 17:06:28 +0100
 Subject: Re: Task process exit with nonzero status of 134
 From: wget.n...@googlemail.com
 To: sokar6...@hotmail.com
 
 Anthoney,
 
 That could be when you try to allocate some dedicated space and it is
 not available or present on your machine.
 Did  -Xmx has an dedicated memory value that you can provide to JVM on
 your machine (e.g. -Xmx1024m)?
 
 Second - filesystem full (mostly hadoop/userlogs). Ext3 has an limit
 of 32k links per node. Try to create a directory in (mkdir
 hadoop/userlogs/123), if you have too much directories an error will
 occur.
 
 - Alex
 
 
 
 
 
 On Tue, Dec 27, 2011 at 3:18 PM, anthony garnier sokar6...@hotmail.com 
 wrote:
  Alex
  I don't know why but now I got other error code :
 
  On Jobtracker :
  java.lang.Throwable: Child Error
  at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:271)
  Caused by: java.io.IOException: Task process exit with nonzero status of 1.
  at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
 
  On taskTracker :
 
  2011-12-27 15:14:50,663 INFO org.apache.hadoop.mapred.DefaultTaskController:
  Output from DefaultTaskController's launchTask follows:
  2011-12-27 15:14:50,663 INFO org.apache.hadoop.mapred.TaskController:
  2011-12-27 15:14:50,664 INFO org.apache.hadoop.mapred.JvmManager: JVM Not
  killed jvm_201112271514_0001_m_-1820714542 but just removed
  2011-12-27 15:14:50,664 INFO org.apache.hadoop.mapred.JvmManager: JVM :
  jvm_201112271514_0001_m_-1820714542 exited with exit code 1. Number of tasks
  it ran: 0
  2011-12-27 15:14:50,666 WARN org.apache.hadoop.mapred.TaskRunner:
  attempt_201112271514_0001_m_02_0 : Child Error
  java.io.IOException: Task process exit with nonzero status of 1.
  at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)
  2011-12-27 15:14:53,676 INFO org.apache.hadoop.mapred.TaskTracker:
  addFreeSlot : current free slots : 2
 
 
  
  From: sokar6...@hotmail.com
  To: wget.n...@googlemail.com; common-user@hadoop.apache.org
  Subject: RE: Task process exit with nonzero status of 134
  Date: Tue, 27 Dec 2011 13:54:22 +
 
 
  Alex
 
  -XX:+UseCompressedOops option is the default in 1.6.0_24 and above on 64 bit
  JVMs (http://wiki.apache.org/hadoop/HadoopJavaVersions)
  Anyway, I tested it but same result.
  Is it wise to test hadoop with the new jdk7_2 ?
 
  Anthony
 
 
  Date: Tue, 27 Dec 2011 13:47:03 +0100
  Subject: Re: Task process exit with nonzero status of 134
  From: wget.n...@googlemail.com
 
  To: sokar6...@hotmail.com
 
  Anthony,
 
  134 depends mostly on JRE (Bug) or defect RAM. _30 is the newest
  update, could be a bug inside. Can you test SE 7u2?
  Todd mentioned in a older post to use -XX:+UseCompressedOops
  (hadoop-env.sh). Another option could be to take a closer look at
  garbage collection with compressed option.
 
  - Alex
 
  On Tue, Dec 27, 2011 at 1:20 PM, anthony garnier sokar6...@hotmail.com
  wrote:
   Alex,
  
   Memory available on namenode / Jobtracker :
   Tasks: 435 total,   1 running, 434 sleeping,   0 stopped,   0 zombie
   Cpu(s):  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,
   0.0%st
   Mem: 15360M total,11609M used, 3750M free,  311M buffers
   Swap: 2047M total,1M used, 2046M free, 8833M cached
  
   On datanode / Tasktracker :
   top - 13:15:27 up 6 days, 21:11,  1 user,  load average: 0.03, 0.28,
   0.26
   Tasks: 377 total,   1 running, 376 sleeping,   0 stopped,   0 zombie
   Cpu(s):  0.0%us,  0.0%sy,  0.0%ni, 99.9%id,  0.0%wa,  0.0%hi,  0.0%si,
   0.0%st
   Mem: 72373M total, 4321M used,68051M free,  348M buffers
   Swap: 2047M total,0M used, 2047M free, 2771M cached
  
   src/mapred/mapred-default.xml :
   property
 namemapred.child.java.opts/name
 value-Xmx200m/value
   /property
  
  
   So there should be enough memory
  
   Anthony
  
  
   Date: Tue, 27 Dec 2011 11:58:46 +0100
  
   Subject: Re: Task process exit with nonzero status of 134
   From: wget.n...@googlemail.com
   To: sokar6...@hotmail.com
  
  
   Anthony,
  
   How much memory you have available? Did the system going into swap?
  
   - Check mapred.map.child.java.opts (mapred.xml) for given MaxSize
   (xmx).
   - what says top -Hc?
  
   - Alex
  
   On Tue, Dec 27, 2011 at 11:49 AM, anthony garnier
   sokar6...@hotmail.com
   wrote:
Hi,
   
I got Nothing in the dmesg
I've checked the Tasktracker and this is what I got :
   
/
STARTUP_MSG: Starting TaskTracker
STARTUP_MSG:   host = ylal2960.inetpsa.com/10.68.217.86

Re: Task process exit with nonzero status of 134

2011-12-27 Thread alo alt
Anthony,

personally I haven't tested yet, jdk7 have bugs already. Was only a
hint to see if the error occurs.
I would focus on memory issues, the installed RAM are okay? No errors?
My next step would be to downgrade into one JDK earlier to check for a
bug. Did you update the OS before?

- Alex

On Tue, Dec 27, 2011 at 2:54 PM, anthony garnier sokar6...@hotmail.com wrote:

 Alex

 -XX:+UseCompressedOops option is the default in 1.6.0_24 and above on 64 bit 
 JVMs (http://wiki.apache.org/hadoop/HadoopJavaVersions)
 Anyway, I tested it but same result.
 Is it wise to test hadoop with the new jdk7_2 ?

 Anthony


 Date: Tue, 27 Dec 2011 13:47:03 +0100
 Subject: Re: Task process exit with nonzero status of 134
 From: wget.n...@googlemail.com

 To: sokar6...@hotmail.com

 Anthony,

 134 depends mostly on JRE (Bug) or defect RAM. _30 is the newest
 update, could be a bug inside. Can you test SE 7u2?
 Todd mentioned in a older post to use -XX:+UseCompressedOops
 (hadoop-env.sh). Another option could be to take a closer look at
 garbage collection with compressed option.

 - Alex

 On Tue, Dec 27, 2011 at 1:20 PM, anthony garnier sokar6...@hotmail.com 
 wrote:
  Alex,
 
  Memory available on namenode / Jobtracker :
  Tasks: 435 total,   1 running, 434 sleeping,   0 stopped,   0 zombie
  Cpu(s):  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,
  0.0%st
  Mem:     15360M total,    11609M used,     3750M free,      311M buffers
  Swap:     2047M total,        1M used,     2046M free,     8833M cached
 
  On datanode / Tasktracker :
  top - 13:15:27 up 6 days, 21:11,  1 user,  load average: 0.03, 0.28, 0.26
  Tasks: 377 total,   1 running, 376 sleeping,   0 stopped,   0 zombie
  Cpu(s):  0.0%us,  0.0%sy,  0.0%ni, 99.9%id,  0.0%wa,  0.0%hi,  0.0%si,
  0.0%st
  Mem:     72373M total,     4321M used,    68051M free,      348M buffers
  Swap:     2047M total,        0M used,     2047M free,     2771M cached
 
  src/mapred/mapred-default.xml :
  property
    namemapred.child.java.opts/name
    value-Xmx200m/value
  /property
 
 
  So there should be enough memory
 
  Anthony
 
 
  Date: Tue, 27 Dec 2011 11:58:46 +0100
 
  Subject: Re: Task process exit with nonzero status of 134
  From: wget.n...@googlemail.com
  To: sokar6...@hotmail.com
 
 
  Anthony,
 
  How much memory you have available? Did the system going into swap?
 
  - Check mapred.map.child.java.opts (mapred.xml) for given MaxSize (xmx).
  - what says top -Hc?
 
  - Alex
 
  On Tue, Dec 27, 2011 at 11:49 AM, anthony garnier sokar6...@hotmail.com
  wrote:
   Hi,
  
   I got Nothing in the dmesg
   I've checked the Tasktracker and this is what I got :
  
   /
   STARTUP_MSG: Starting TaskTracker
   STARTUP_MSG:   host = ylal2960.inetpsa.com/10.68.217.86
   STARTUP_MSG:   args = []
   STARTUP_MSG:   version = 0.20.203.0
   STARTUP_MSG:   build =
  
   http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-security-203
   -r 1099333; compiled by 'oom' on Wed May  4 07:57:50 PDT 2011
   /
   2011-12-23 15:11:02,275 INFO
   org.apache.hadoop.metrics2.impl.MetricsConfig:
   loaded properties from hadoop-metrics2.properties
   2011-12-23 15:11:02,330 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   MetricsSystem,sub=Stats registered.
   2011-12-23 15:11:02,331 INFO
   org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
   period
   at 10 second(s).
   2011-12-23 15:11:02,331 INFO
   org.apache.hadoop.metrics2.impl.MetricsSystemImpl: TaskTracker metrics
   system started
   2011-12-23 15:11:02,597 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   ugi
   registered.
   2011-12-23 15:11:02,738 INFO org.mortbay.log: Logging to
   org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
   org.mortbay.log.Slf4jLog
   2011-12-23 15:11:02,803 INFO org.apache.hadoop.http.HttpServer: Added
   global
   filtersafety
   (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
   2011-12-23 15:11:02,827 INFO org.apache.hadoop.mapred.TaskLogsTruncater:
   Initializing logs' truncater with mapRetainSize=-1 and
   reduceRetainSize=-1
   2011-12-23 15:11:02,832 INFO org.apache.hadoop.mapred.TaskTracker:
   Starting
   tasktracker with owner as root
   2011-12-23 15:11:02,870 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   jvm
   registered.
   2011-12-23 15:11:02,871 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   TaskTrackerMetrics registered.
   2011-12-23 15:11:02,897 INFO org.apache.hadoop.ipc.Server: Starting
   SocketReader
   2011-12-23 15:11:02,900 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   RpcDetailedActivityForPort58709 registered.
   2011-12-23 15:11:02,900 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source

RE: Task process exit with nonzero status of 134

2011-12-27 Thread anthony garnier

Alex

-XX:+UseCompressedOops option is the default in 1.6.0_24 and above on 64 bit 
JVMs (http://wiki.apache.org/hadoop/HadoopJavaVersions)
Anyway, I tested it but same result.
Is it wise to test hadoop with the new jdk7_2 ?

Anthony


 Date: Tue, 27 Dec 2011 13:47:03 +0100
 Subject: Re: Task process exit with nonzero status of 134
 From: wget.n...@googlemail.com

 To: sokar6...@hotmail.com
 
 Anthony,
 
 134 depends mostly on JRE (Bug) or defect RAM. _30 is the newest
 update, could be a bug inside. Can you test SE 7u2?
 Todd mentioned in a older post to use -XX:+UseCompressedOops
 (hadoop-env.sh). Another option could be to take a closer look at
 garbage collection with compressed option.
 
 - Alex
 
 On Tue, Dec 27, 2011 at 1:20 PM, anthony garnier sokar6...@hotmail.com 
 wrote:
  Alex,
 
  Memory available on namenode / Jobtracker :
  Tasks: 435 total,   1 running, 434 sleeping,   0 stopped,   0 zombie
  Cpu(s):  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,
  0.0%st
  Mem: 15360M total,11609M used, 3750M free,  311M buffers
  Swap: 2047M total,1M used, 2046M free, 8833M cached
 
  On datanode / Tasktracker :
  top - 13:15:27 up 6 days, 21:11,  1 user,  load average: 0.03, 0.28, 0.26
  Tasks: 377 total,   1 running, 376 sleeping,   0 stopped,   0 zombie
  Cpu(s):  0.0%us,  0.0%sy,  0.0%ni, 99.9%id,  0.0%wa,  0.0%hi,  0.0%si,
  0.0%st
  Mem: 72373M total, 4321M used,68051M free,  348M buffers
  Swap: 2047M total,0M used, 2047M free, 2771M cached
 
  src/mapred/mapred-default.xml :
  property
namemapred.child.java.opts/name
value-Xmx200m/value
  /property
 
 
  So there should be enough memory
 
  Anthony
 
 
  Date: Tue, 27 Dec 2011 11:58:46 +0100
 
  Subject: Re: Task process exit with nonzero status of 134
  From: wget.n...@googlemail.com
  To: sokar6...@hotmail.com
 
 
  Anthony,
 
  How much memory you have available? Did the system going into swap?
 
  - Check mapred.map.child.java.opts (mapred.xml) for given MaxSize (xmx).
  - what says top -Hc?
 
  - Alex
 
  On Tue, Dec 27, 2011 at 11:49 AM, anthony garnier sokar6...@hotmail.com
  wrote:
   Hi,
  
   I got Nothing in the dmesg
   I've checked the Tasktracker and this is what I got :
  
   /
   STARTUP_MSG: Starting TaskTracker
   STARTUP_MSG:   host = ylal2960.inetpsa.com/10.68.217.86
   STARTUP_MSG:   args = []
   STARTUP_MSG:   version = 0.20.203.0
   STARTUP_MSG:   build =
  
   http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-security-203
   -r 1099333; compiled by 'oom' on Wed May  4 07:57:50 PDT 2011
   /
   2011-12-23 15:11:02,275 INFO
   org.apache.hadoop.metrics2.impl.MetricsConfig:
   loaded properties from hadoop-metrics2.properties
   2011-12-23 15:11:02,330 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   MetricsSystem,sub=Stats registered.
   2011-12-23 15:11:02,331 INFO
   org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
   period
   at 10 second(s).
   2011-12-23 15:11:02,331 INFO
   org.apache.hadoop.metrics2.impl.MetricsSystemImpl: TaskTracker metrics
   system started
   2011-12-23 15:11:02,597 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   ugi
   registered.
   2011-12-23 15:11:02,738 INFO org.mortbay.log: Logging to
   org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
   org.mortbay.log.Slf4jLog
   2011-12-23 15:11:02,803 INFO org.apache.hadoop.http.HttpServer: Added
   global
   filtersafety
   (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
   2011-12-23 15:11:02,827 INFO org.apache.hadoop.mapred.TaskLogsTruncater:
   Initializing logs' truncater with mapRetainSize=-1 and
   reduceRetainSize=-1
   2011-12-23 15:11:02,832 INFO org.apache.hadoop.mapred.TaskTracker:
   Starting
   tasktracker with owner as root
   2011-12-23 15:11:02,870 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   jvm
   registered.
   2011-12-23 15:11:02,871 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   TaskTrackerMetrics registered.
   2011-12-23 15:11:02,897 INFO org.apache.hadoop.ipc.Server: Starting
   SocketReader
   2011-12-23 15:11:02,900 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   RpcDetailedActivityForPort58709 registered.
   2011-12-23 15:11:02,900 INFO
   org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
   RpcActivityForPort58709 registered.
   2011-12-23 15:11:02,903 INFO org.apache.hadoop.ipc.Server: IPC Server
   Responder: starting
   2011-12-23 15:11:02,904 INFO org.apache.hadoop.ipc.Server: IPC Server
   listener on 58709: starting
   2011-12-23 15:11:02,904 INFO org.apache.hadoop.ipc.Server: IPC Server
   handler 0 on 58709: starting
   2011-12-23 15:11:02,904 INFO

Re: Task process exit with nonzero status of 134

2011-12-27 Thread anthony garnier
 with retainTimeStamp:1324735862967
2011-12-23 15:11:03,264 INFO org.apache.hadoop.mapred.UserLogCleaner: Adding 
job_201112231312_0002 for user-log deletion with retainTimeStamp:1324735862967
2011-12-23 15:11:03,264 INFO org.apache.hadoop.mapred.UserLogCleaner: Adding 
job_201112231142_0002 for user-log deletion with retainTimeStamp:1324735862967
2011-12-23 15:11:03,264 INFO org.apache.hadoop.mapred.UserLogCleaner: Adding 
job_201112230933_0002 for user-log deletion with retainTimeStamp:1324735862967
2011-12-23 15:11:03,264 INFO org.apache.hadoop.mapred.UserLogCleaner: Adding 
job_201112231357_0001 for user-log deletion with retainTimeStamp:1324735862967
2011-12-23 15:11:03,264 INFO org.apache.hadoop.mapred.UserLogCleaner: Adding 
job_201112231052_0002 for user-log deletion with retainTimeStamp:1324735862967
2011-12-23 15:11:03,264 INFO org.apache.hadoop.mapred.UserLogCleaner: Adding 
job_201112231426_0001 for user-log deletion with retainTimeStamp:1324735862967
2011-12-23 15:11:51,501 INFO org.apache.hadoop.mapred.TaskTracker: 
LaunchTaskAction (registerTask): attempt_201112231511_0001_m_02_0 task's 
state:UNASSIGNED
2011-12-23 15:11:51,505 INFO org.apache.hadoop.mapred.TaskTracker: Trying to 
launch : attempt_201112231511_0001_m_02_0 which needs 1 slots
2011-12-23 15:11:51,505 INFO org.apache.hadoop.mapred.TaskTracker: In 
TaskLauncher, current free slots : 2 and trying to launch 
attempt_201112231511_0001_m_02_0 which needs 1 slots
2011-12-23 15:11:51,633 WARN org.apache.hadoop.conf.Configuration: 
/users/hadoop/mapred/local/ttprivate/taskTracker/root/jobcache/job_201112231511_0001/job.xml:a
 attempt to override final parameter: hadoop.tmp.dir;  Ignoring.
2011-12-23 15:11:51,635 WARN org.apache.hadoop.conf.Configuration: 
/users/hadoop/mapred/local/ttprivate/taskTracker/root/jobcache/job_201112231511_0001/job.xml:a
 attempt to override final parameter: dfs.data.dir;  Ignoring.
2011-12-23 15:11:51,636 WARN org.apache.hadoop.conf.Configuration: 
/users/hadoop/mapred/local/ttprivate/taskTracker/root/jobcache/job_201112231511_0001/job.xml:a
 attempt to override final parameter: dfs.name.dir;  Ignoring.
2011-12-23 15:11:51,685 INFO org.apache.hadoop.mapred.JobLocalizer: 
Initializing user root on this TT.
2011-12-23 15:11:51,720 WARN org.apache.hadoop.conf.Configuration: 
/users/hadoop/mapred/local/ttprivate/taskTracker/root/jobcache/job_201112231511_0001/job.xml:a
 attempt to override final parameter: hadoop.tmp.dir;  Ignoring.
2011-12-23 15:11:51,722 WARN org.apache.hadoop.conf.Configuration: 
/users/hadoop/mapred/local/ttprivate/taskTracker/root/jobcache/job_201112231511_0001/job.xml:a
 attempt to override final parameter: dfs.data.dir;  Ignoring.
2011-12-23 15:11:51,723 WARN org.apache.hadoop.conf.Configuration: 
/users/hadoop/mapred/local/ttprivate/taskTracker/root/jobcache/job_201112231511_0001/job.xml:a
 attempt to override final parameter: dfs.name.dir;  Ignoring.
2011-12-23 15:11:51,845 WARN org.apache.hadoop.conf.Configuration: 
/users/hadoop/mapred/local/taskTracker/root/jobcache/job_201112231511_0001/job.xml:a
 attempt to override final parameter: hadoop.tmp.dir;  Ignoring.
2011-12-23 15:11:51,847 WARN org.apache.hadoop.conf.Configuration: 
/users/hadoop/mapred/local/taskTracker/root/jobcache/job_201112231511_0001/job.xml:a
 attempt to override final parameter: dfs.data.dir;  Ignoring.
2011-12-23 15:11:51,847 WARN org.apache.hadoop.conf.Configuration: 
/users/hadoop/mapred/local/taskTracker/root/jobcache/job_201112231511_0001/job.xml:a
 attempt to override final parameter: dfs.name.dir;  Ignoring.
2011-12-23 15:11:51,858 INFO org.apache.hadoop.mapred.JvmManager: In JvmRunner 
constructed JVM ID: jvm_201112231511_0001_m_-1653678851
2011-12-23 15:11:51,859 INFO org.apache.hadoop.mapred.JvmManager: JVM Runner 
jvm_201112231511_0001_m_-1653678851 spawned.
2011-12-23 15:11:51,861 INFO org.apache.hadoop.mapred.TaskController: Writing 
commands to 
/users/hadoop/mapred/local/ttprivate/taskTracker/root/jobcache/job_201112231511_0001/attempt_201112231511_0001_m_02_0/taskjvm.sh
2011-12-23 15:11:52,427 INFO org.apache.hadoop.mapred.TaskTracker: JVM with ID: 
jvm_201112231511_0001_m_-1653678851 given task: 
attempt_201112231511_0001_m_02_0
2011-12-23 15:11:52,675 WARN org.apache.hadoop.mapred.DefaultTaskController: 
Exit code from task is : 134
2011-12-23 15:11:52,675 INFO org.apache.hadoop.mapred.DefaultTaskController: 
Output from DefaultTaskController's launchTask follows:
2011-12-23 15:11:52,675 INFO org.apache.hadoop.mapred.TaskController:
2011-12-23 15:11:52,679 INFO org.apache.hadoop.mapred.JvmManager: JVM : 
jvm_201112231511_0001_m_-1653678851 exited with exit code 134. Number of tasks 
it ran: 0
2011-12-23 15:11:52,681 WARN org.apache.hadoop.mapred.TaskRunner: 
attempt_201112231511_0001_m_02_0 : Child Error
java.io.IOException: Task process exit with nonzero status of 134.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)



I also checked the logs of the task :

logs

RE: Task process exit with nonzero status of 134

2011-12-27 Thread anthony garnier

Alex,

Memory available on namenode / Jobtracker : 
Tasks: 435 total,   1 running, 434 sleeping,   0 stopped,   0 zombie
Cpu(s):  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Mem: 15360M total,11609M used, 3750M free,  311M buffers
Swap: 2047M total,1M used, 2046M free, 8833M cached

On datanode / Tasktracker : 
top - 13:15:27 up 6 days, 21:11,  1 user,  load average: 0.03, 0.28, 0.26
Tasks: 377 total,   1 running, 376 sleeping,   0 stopped,   0 zombie
Cpu(s):  0.0%us,  0.0%sy,  0.0%ni, 99.9%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Mem: 72373M total, 4321M used,68051M free,  348M buffers
Swap: 2047M total,0M used, 2047M free, 2771M cached

src/mapred/mapred-default.xml :
property
  namemapred.child.java.opts/name
  value-Xmx200m/value
/property


So there should be enough memory 

Anthony


 Date: Tue, 27 Dec 2011 11:58:46 +0100
 Subject: Re: Task process exit with nonzero status of 134
 From: wget.n...@googlemail.com
 To: sokar6...@hotmail.com
 
 Anthony,
 
 How much memory you have available? Did the system going into swap?
 
 - Check mapred.map.child.java.opts (mapred.xml) for given MaxSize (xmx).
 - what says top -Hc?
 
 - Alex
 
 On Tue, Dec 27, 2011 at 11:49 AM, anthony garnier sokar6...@hotmail.com 
 wrote:
  Hi,
 
  I got Nothing in the dmesg
  I've checked the Tasktracker and this is what I got :
 
  /
  STARTUP_MSG: Starting TaskTracker
  STARTUP_MSG:   host = ylal2960.inetpsa.com/10.68.217.86
  STARTUP_MSG:   args = []
  STARTUP_MSG:   version = 0.20.203.0
  STARTUP_MSG:   build =
  http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-security-203
  -r 1099333; compiled by 'oom' on Wed May  4 07:57:50 PDT 2011
  /
  2011-12-23 15:11:02,275 INFO org.apache.hadoop.metrics2.impl.MetricsConfig:
  loaded properties from hadoop-metrics2.properties
  2011-12-23 15:11:02,330 INFO
  org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
  MetricsSystem,sub=Stats registered.
  2011-12-23 15:11:02,331 INFO
  org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period
  at 10 second(s).
  2011-12-23 15:11:02,331 INFO
  org.apache.hadoop.metrics2.impl.MetricsSystemImpl: TaskTracker metrics
  system started
  2011-12-23 15:11:02,597 INFO
  org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi
  registered.
  2011-12-23 15:11:02,738 INFO org.mortbay.log: Logging to
  org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
  org.mortbay.log.Slf4jLog
  2011-12-23 15:11:02,803 INFO org.apache.hadoop.http.HttpServer: Added global
  filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
  2011-12-23 15:11:02,827 INFO org.apache.hadoop.mapred.TaskLogsTruncater:
  Initializing logs' truncater with mapRetainSize=-1 and reduceRetainSize=-1
  2011-12-23 15:11:02,832 INFO org.apache.hadoop.mapred.TaskTracker: Starting
  tasktracker with owner as root
  2011-12-23 15:11:02,870 INFO
  org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source jvm
  registered.
  2011-12-23 15:11:02,871 INFO
  org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
  TaskTrackerMetrics registered.
  2011-12-23 15:11:02,897 INFO org.apache.hadoop.ipc.Server: Starting
  SocketReader
  2011-12-23 15:11:02,900 INFO
  org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
  RpcDetailedActivityForPort58709 registered.
  2011-12-23 15:11:02,900 INFO
  org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source
  RpcActivityForPort58709 registered.
  2011-12-23 15:11:02,903 INFO org.apache.hadoop.ipc.Server: IPC Server
  Responder: starting
  2011-12-23 15:11:02,904 INFO org.apache.hadoop.ipc.Server: IPC Server
  listener on 58709: starting
  2011-12-23 15:11:02,904 INFO org.apache.hadoop.ipc.Server: IPC Server
  handler 0 on 58709: starting
  2011-12-23 15:11:02,904 INFO org.apache.hadoop.ipc.Server: IPC Server
  handler 1 on 58709: starting
  2011-12-23 15:11:02,904 INFO org.apache.hadoop.ipc.Server: IPC Server
  handler 2 on 58709: starting
  2011-12-23 15:11:02,905 INFO org.apache.hadoop.mapred.TaskTracker:
  TaskTracker up at: localhost/127.0.0.1:58709
  2011-12-23 15:11:02,905 INFO org.apache.hadoop.mapred.TaskTracker: Starting
  tracker tracker_ylal2960.inetpsa.com:localhost/127.0.0.1:58709
  2011-12-23 15:11:02,908 INFO org.apache.hadoop.ipc.Server: IPC Server
  handler 3 on 58709: starting
  2011-12-23 15:11:02,958 INFO org.apache.hadoop.mapred.TaskTracker:  Using
  MemoryCalculatorPlugin :
  org.apache.hadoop.util.LinuxMemoryCalculatorPlugin@7054f9f1
  2011-12-23 15:11:02,960 INFO org.apache.hadoop.mapred.TaskTracker: Starting
  thread: Map-events fetcher for all reduce tasks on
  tracker_ylal2960.inetpsa.com:localhost/127.0.0.1:58709
  2011-12-23 15:11:02,964 INFO

Task process exit with nonzero status of 134

2011-12-23 Thread anthony garnier

Hi folks,

I've just done a fresh install of Hadoop, Namenode and datanode are up, 
Task/job Tracker also up, but when I run the Map reduce worcount exemple I got 
this error on Task tracker: 

2011-12-23 15:11:52,679 INFO org.apache.hadoop.mapred.JvmManager: JVM : 
jvm_201112231511_0001_m_-1653678851 exited with exit code 134. Number of tasks 
it ran: 0
2011-12-23 15:11:52,681 WARN org.apache.hadoop.mapred.TaskRunner: 
attempt_201112231511_0001_m_02_0 : Child Error
java.io.IOException: Task process exit with nonzero status of 134.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)

And job tracker stuck : 
# hadoop jar hadoop-examples*.jar wordcount input/test.txt output/
11/12/23 15:11:48 INFO input.FileInputFormat: Total input paths to process : 1
11/12/23 15:11:49 INFO mapred.JobClient: Running job: job_201112231511_0001
11/12/23 15:11:50 INFO mapred.JobClient:  map 0% reduce 0%


I'running hadoop 0.20.203.0, java 1.6.0 rev 25

I've done some googling, apparently the JVM crash hard (maybe Out of memory), 
does someone have any hint ?

Regards,

Anthony Garnier
/DSIN/ASTI/ETSO
IT Center
PSA Peugeot Citroen
Bessoncourt 90160



  

Re: Task process exit with nonzero status of 134

2011-12-23 Thread alo alt
Hi,

take a look into the logs for the failed attempt at your Tasktracker.
Also check the system logs with dmesg or /var/log/kern*. Could be a
syskill (segfault).

- Alex

On Fri, Dec 23, 2011 at 3:32 PM, anthony garnier sokar6...@hotmail.com wrote:

 Hi folks,

 I've just done a fresh install of Hadoop, Namenode and datanode are up, 
 Task/job Tracker also up, but when I run the Map reduce worcount exemple I 
 got this error on Task tracker:

 2011-12-23 15:11:52,679 INFO org.apache.hadoop.mapred.JvmManager: JVM : 
 jvm_201112231511_0001_m_-1653678851 exited with exit code 134. Number of 
 tasks it ran: 0
 2011-12-23 15:11:52,681 WARN org.apache.hadoop.mapred.TaskRunner: 
 attempt_201112231511_0001_m_02_0 : Child Error
 java.io.IOException: Task process exit with nonzero status of 134.
        at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:258)

 And job tracker stuck :
 # hadoop jar hadoop-examples*.jar wordcount input/test.txt output/
 11/12/23 15:11:48 INFO input.FileInputFormat: Total input paths to process : 1
 11/12/23 15:11:49 INFO mapred.JobClient: Running job: job_201112231511_0001
 11/12/23 15:11:50 INFO mapred.JobClient:  map 0% reduce 0%


 I'running hadoop 0.20.203.0, java 1.6.0 rev 25

 I've done some googling, apparently the JVM crash hard (maybe Out of memory), 
 does someone have any hint ?

 Regards,

 Anthony Garnier
 /DSIN/ASTI/ETSO
 IT Center
 PSA Peugeot Citroen
 Bessoncourt 90160







-- 
Alexander Lorenz
http://mapredit.blogspot.com

P Think of the environment: please don't print this email unless you
really need to.


Re: java.io.IOException: Task process exit with nonzero status of 134

2011-03-09 Thread Shi Yu

Here is some update of the problem.

I tried a very simple example (word count) and tried to compress the 
reducer output using default Codec, or GzipCodec. I didn't try LZO to 
avoid further troubles. I didn't use the combiner class, and set the 
number of reducer to 1.  I am trying it on a 64-bit Debian. My java 
version is

java version 1.6.0
Java(TM) SE Runtime Environment (build 1.6.0-b105)
Java HotSpot(TM) 64-Bit Server VM (build 1.6.0-b105, mixed mode)


I use SequenceFileOutputFormat, and the output value class is a Vector.  
At first, I didn't specify the compressOutput and the there was no 
problem. The sequenceFile was generated correctly.  However, when I 
compressed the output adding the following 3-line-command:


SequenceFileOutputFormat.setCompressOutput(conf,true);
SequenceFileOutputFormat.setOutputCompressorClass(conf, 
DefaultCodec.class);
SequenceFileOutputFormat.setOutputCompressionType(conf, 
SequenceFile.CompressionType.BLOCK);


The reducer kept on generating error and the task finally crashed.
11/03/09 12:28:05 INFO mapred.JobClient:  map 100% reduce 33%
11/03/09 12:28:09 INFO mapred.JobClient:  map 100% reduce 0%
11/03/09 12:28:09 INFO mapred.JobClient: Task Id : 
attempt_201103081457_0024_r_00_0, Status : FAILED

java.io.IOException: Task process exit with nonzero status of 134.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:418)

11/03/09 12:28:19 INFO mapred.JobClient:  map 100% reduce 29%
11/03/09 12:28:21 INFO mapred.JobClient:  map 100% reduce 0%
11/03/09 12:28:22 INFO mapred.JobClient: Task Id : 
attempt_201103081457_0024_r_00_1, Status : FAILED

java.io.IOException: Task process exit with nonzero status of 134.
at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:418)


I checked the logs of the reducer (only 1), there is no error in syslog. 
But in the stdout file, there is an error information:



An unexpected error has been detected by Java Runtime Environment:
#
#  SIGFPE (0x8) at pc=0x2b22eecc7b83, pid=13306, tid=1076017504
#
# Java VM: Java HotSpot(TM) 64-Bit Server VM (1.6.0-b105 mixed mode)
# Problematic frame:
# C  [ld-linux-x86-64.so.2+0x7b83]
#
# An error report file with more information is saved as hs_err_pid13306.log
#
# If you would like to submit a bug report, please visit:
#   http://java.sun.com/webapps/bugreport/crash.jsp

So I would like to know how should I solve this problem. Should I 
upgrade anything? I guess this problem is not new. Thanks for the 
information.


Shi

On 3/8/2011 4:04 PM, Shi Yu wrote:
What is the true reason of causing this? I realized there are many 
reports on web, but couldn't find the exact solution? I have this 
problem when using compressed sequence file output.


SequenceFileOutputFormat.setCompressOutput(conf, true);
SequenceFileOutputFormat.setOutputCompressorClass(conf, GzipCodec.class);
SequenceFileOutputFormat.setOutputCompressionType(conf, 
CompressionType.BLOCK);


If I remove that 3 lines, everything is fine.  I am using hadoop 
0.19.2, is there any way to avoid the problem without upgrading hadoop?


Thanks!

Shi






Re: java.io.IOException: Task process exit with nonzero status of 134

2011-03-09 Thread Greg Roelofs
Shi Yu wrote:

 An unexpected error has been detected by Java Runtime Environment:
 #
 #  SIGFPE (0x8) at pc=0x2b22eecc7b83, pid=13306, tid=1076017504
 #
 # Java VM: Java HotSpot(TM) 64-Bit Server VM (1.6.0-b105 mixed mode)
 # Problematic frame:
 # C  [ld-linux-x86-64.so.2+0x7b83]
 #
 # An error report file with more information is saved as hs_err_pid13306.log
 #
 # If you would like to submit a bug report, please visit:
 #   http://java.sun.com/webapps/bugreport/crash.jsp

So it's a 64-bit JVM, and it's invoking a 64-bit ld.so, presumably to load
a native library.  Is your zlib also 64-bit?  I vaguely recall seeing a
SIGFPE a while back, but I don't recall what it turned out to be.  (Seems
unlikely that a 32/64-bit mismatch might cause one, but that's all that
comes to mind offhand.)

Btw, the zlib codec exists in both native (JNI) and Java forms (well, with
the native part buried inside the JDK), so you could also try disabling
native support (set io.native.lib.available to false).  If it still fails,
it's not related to shared libraries.

You could also use the bzip2 codec rather than the default one; bzip2
support exists only in Java form for some inexplicable reason.  (Perhaps
users were blinded by its blazing speed.)

Greg


java.io.IOException: Task process exit with nonzero status of 134

2011-03-08 Thread Shi Yu
What is the true reason of causing this? I realized there are many 
reports on web, but couldn't find the exact solution? I have this 
problem when using compressed sequence file output.


SequenceFileOutputFormat.setCompressOutput(conf, true);
SequenceFileOutputFormat.setOutputCompressorClass(conf, GzipCodec.class);
SequenceFileOutputFormat.setOutputCompressionType(conf, 
CompressionType.BLOCK);


If I remove that 3 lines, everything is fine.  I am using hadoop 0.19.2, 
is there any way to avoid the problem without upgrading hadoop?


Thanks!

Shi