Hi I have added following line to my master node mapred-site.xml file <property> <name>mapred.child.ulimit</name> <value>3145728</value> </property>
and run the job again, and wow..., the jobs get completed in 4th attempt. I checked the at 50030. Hadoop runs job 3 times on master server and it fails but when it run on 2nd node, it succeeded and produce the desired result. Why it failed on master? Thanks Shuja On Tue, Jul 13, 2010 at 1:34 AM, Alex Kozlov <ale...@cloudera.com> wrote: > Hmm. It means your options are not propagated to the nodes. Can you put * > mapred.child.ulimit* in the mapred-siet.xml and restart the tasktrackers? > I > was under impression that the below should be enough though. Glad you got > it working in local mode. -- Alex K > > On Mon, Jul 12, 2010 at 1:24 PM, Shuja Rehman <shujamug...@gmail.com> > wrote: > > > Hi Alex, I am using putty to connect to servers. and this is almost my > > maximum screen output which i sent. putty is not allowed me to increase > the > > size of terminal. is there any other way that i get the complete output > of > > ps-aef? > > > > Now i run the following command and thnx God, it did not fails and > produce > > the desired output. > > > > hadoop jar > > /usr/lib/hadoop-0.20/contrib/streaming/hadoop-streaming-0.20.2+320.jar \ > > -D mapred.child.java.opts=-Xmx1024m \ > > -D mapred.child.ulimit=3145728 \ > > -jt local \ > > -inputformat StreamInputFormat \ > > -inputreader "StreamXmlRecordReader,begin=<mdc xmlns:HTML=\" > > http://www.w3.org/TR/REC-xml\ <http://www.w3.org/TR/REC-xml%5C> < > http://www.w3.org/TR/REC-xml%5C>">,end=</mdc>" > > \ > > -input > > > > > /user/root/RNCDATA/MDFDORKUCRAR02/A20100531.0000-0700-0015-0700_RNCCN-MDFDORKUCRAR02 > > \ > > -jobconf mapred.map.tasks=1 \ > > -jobconf mapred.reduce.tasks=0 \ > > -output RNC32 \ > > -mapper /home/ftpuser1/Nodemapper5.groovy \ > > -reducer org.apache.hadoop.mapred.lib.IdentityReducer \ > > -file /home/ftpuser1/Nodemapper5.groovy > > > > > > but when i omit the -jt local, it produces the same error. > > Thanks Alex for helping > > Regards > > Shuja > > > > On Tue, Jul 13, 2010 at 1:01 AM, Alex Kozlov <ale...@cloudera.com> > wrote: > > > > > Hi Shuja, > > > > > > Java listens to the last xmx, so if you have multiple "-Xmx ..." on the > > > command line, the last is valid. Unfortunately you have truncated > > command > > > lines. Can you show us the full command line, particularly for the > > process > > > 26162? This seems to be causing problems. > > > > > > If you are running your cluster on 2 nodes, it may be that the task was > > > scheduled on the second node. Did you run "ps -aef" on the second node > > as > > > well? You can see the task assignment in the JT web-UI ( > > > http://jt-name:50030, drill down to tasks). > > > > > > I suggest you first debug your program in the local mode first, however > > > (use > > > "*-jt local*" option). Did you try the "*-D > > mapred.child.ulimit=3145728*" > > > option? I do not see it on the command line. > > > > > > Alex K > > > > > > On Mon, Jul 12, 2010 at 12:20 PM, Shuja Rehman <shujamug...@gmail.com > > > >wrote: > > > > > > > Hi Alex > > > > > > > > I have tried with using quotes and also with -jt local but same heap > > > > error. > > > > and here is the output of ps -aef > > > > > > > > UID PID PPID C STIME TTY TIME CMD > > > > root 1 0 0 04:37 ? 00:00:00 init [3] > > > > root 2 1 0 04:37 ? 00:00:00 [migration/0] > > > > root 3 1 0 04:37 ? 00:00:00 [ksoftirqd/0] > > > > root 4 1 0 04:37 ? 00:00:00 [watchdog/0] > > > > root 5 1 0 04:37 ? 00:00:00 [events/0] > > > > root 6 1 0 04:37 ? 00:00:00 [khelper] > > > > root 7 1 0 04:37 ? 00:00:00 [kthread] > > > > root 9 7 0 04:37 ? 00:00:00 [xenwatch] > > > > root 10 7 0 04:37 ? 00:00:00 [xenbus] > > > > root 17 7 0 04:37 ? 00:00:00 [kblockd/0] > > > > root 18 7 0 04:37 ? 00:00:00 [cqueue/0] > > > > root 22 7 0 04:37 ? 00:00:00 [khubd] > > > > root 24 7 0 04:37 ? 00:00:00 [kseriod] > > > > root 84 7 0 04:37 ? 00:00:00 [khungtaskd] > > > > root 85 7 0 04:37 ? 00:00:00 [pdflush] > > > > root 86 7 0 04:37 ? 00:00:00 [pdflush] > > > > root 87 7 0 04:37 ? 00:00:00 [kswapd0] > > > > root 88 7 0 04:37 ? 00:00:00 [aio/0] > > > > root 229 7 0 04:37 ? 00:00:00 [kpsmoused] > > > > root 248 7 0 04:37 ? 00:00:00 [kstriped] > > > > root 257 7 0 04:37 ? 00:00:00 [kjournald] > > > > root 279 7 0 04:37 ? 00:00:00 [kauditd] > > > > root 307 1 0 04:37 ? 00:00:00 /sbin/udevd -d > > > > root 634 7 0 04:37 ? 00:00:00 [kmpathd/0] > > > > root 635 7 0 04:37 ? 00:00:00 [kmpath_handlerd] > > > > root 660 7 0 04:37 ? 00:00:00 [kjournald] > > > > root 662 7 0 04:37 ? 00:00:00 [kjournald] > > > > root 1032 1 0 04:38 ? 00:00:00 auditd > > > > root 1034 1032 0 04:38 ? 00:00:00 /sbin/audispd > > > > root 1049 1 0 04:38 ? 00:00:00 syslogd -m 0 > > > > root 1052 1 0 04:38 ? 00:00:00 klogd -x > > > > root 1090 7 0 04:38 ? 00:00:00 [rpciod/0] > > > > root 1158 1 0 04:38 ? 00:00:00 rpc.idmapd > > > > dbus 1171 1 0 04:38 ? 00:00:00 dbus-daemon --system > > > > root 1184 1 0 04:38 ? 00:00:00 /usr/sbin/hcid > > > > root 1190 1 0 04:38 ? 00:00:00 /usr/sbin/sdpd > > > > root 1210 1 0 04:38 ? 00:00:00 [krfcommd] > > > > root 1244 1 0 04:38 ? 00:00:00 pcscd > > > > root 1264 1 0 04:38 ? 00:00:00 /usr/bin/hidd > --server > > > > root 1295 1 0 04:38 ? 00:00:00 automount > > > > root 1314 1 0 04:38 ? 00:00:00 /usr/sbin/sshd > > > > root 1326 1 0 04:38 ? 00:00:00 xinetd -stayalive > > > -pidfile > > > > /var/run/xinetd.pid > > > > root 1337 1 0 04:38 ? 00:00:00 /usr/sbin/vsftpd > > > > /etc/vsftpd/vsftpd.conf > > > > root 1354 1 0 04:38 ? 00:00:00 sendmail: accepting > > > > connections > > > > smmsp 1362 1 0 04:38 ? 00:00:00 sendmail: Queue > > run...@01 > > > > :00:00 > > > > for /var/spool/clientmqueue > > > > root 1379 1 0 04:38 ? 00:00:00 gpm -m > /dev/input/mice > > -t > > > > exps2 > > > > root 1410 1 0 04:38 ? 00:00:00 crond > > > > xfs 1450 1 0 04:38 ? 00:00:00 xfs -droppriv -daemon > > > > root 1482 1 0 04:38 ? 00:00:00 /usr/sbin/atd > > > > 68 1508 1 0 04:38 ? 00:00:00 hald > > > > root 1509 1508 0 04:38 ? 00:00:00 hald-runner > > > > root 1533 1 0 04:38 ? 00:00:00 /usr/sbin/smartd -q > > never > > > > root 1536 1 0 04:38 xvc0 00:00:00 /sbin/agetty xvc0 > 9600 > > > > vt100-nav > > > > root 1537 1 0 04:38 ? 00:00:00 /usr/bin/python -tt > > > > /usr/sbin/yum-updatesd > > > > root 1539 1 0 04:38 ? 00:00:00 > /usr/libexec/gam_server > > > > root 21022 1314 0 11:27 ? 00:00:00 sshd: r...@pts/0 > > > > root 21024 21022 0 11:27 pts/0 00:00:00 -bash > > > > root 21103 1314 0 11:28 ? 00:00:00 sshd: r...@pts/1 > > > > root 21105 21103 0 11:28 pts/1 00:00:00 -bash > > > > root 21992 1314 0 11:47 ? 00:00:00 sshd: r...@pts/2 > > > > root 21994 21992 0 11:47 pts/2 00:00:00 -bash > > > > root 22433 1314 0 11:49 ? 00:00:00 sshd: r...@pts/3 > > > > root 22437 22433 0 11:49 pts/3 00:00:00 -bash > > > > hadoop 24808 1 0 12:01 ? 00:00:02 > > /usr/jdk1.6.0_03/bin/java > > > > -Xmx2001m -Dcom.sun.management.jmxremote > -Dcom.sun.management.jmxremote > > > > -Dhadoop.lo > > > > hadoop 24893 1 0 12:01 ? 00:00:01 > > /usr/jdk1.6.0_03/bin/java > > > > -Xmx2001m -Dcom.sun.management.jmxremote > -Dcom.sun.management.jmxremote > > > > -Dhadoop.lo > > > > hadoop 24988 1 0 12:01 ? 00:00:01 > > /usr/jdk1.6.0_03/bin/java > > > > -Xmx2001m -Dcom.sun.management.jmxremote > -Dcom.sun.management.jmxremote > > > > -Dhadoop.lo > > > > hadoop 25085 1 0 12:01 ? 00:00:00 > > /usr/jdk1.6.0_03/bin/java > > > > -Xmx2001m -Dcom.sun.management.jmxremote > -Dcom.sun.management.jmxremote > > > > -Dhadoop.lo > > > > hadoop 25175 1 0 12:01 ? 00:00:01 > > /usr/jdk1.6.0_03/bin/java > > > > -Xmx2001m -Dhadoop.log.dir=/usr/lib/hadoop-0.20/bin/../logs > > > > -Dhadoop.log.file=hadoo > > > > root 25925 21994 1 12:06 pts/2 00:00:00 > > /usr/jdk1.6.0_03/bin/java > > > > -Xmx2001m -Dhadoop.log.dir=/usr/lib/hadoop-0.20/logs > > > > -Dhadoop.log.file=hadoop.log - > > > > hadoop 26120 25175 14 12:06 ? 00:00:01 > > > > /usr/jdk1.6.0_03/jre/bin/java > > > > > > > > > > > > > > -Djava.library.path=/usr/jdk1.6.0_03/jre/lib/i386/client:/usr/jdk1.6.0_03/jre/l > > > > hadoop 26162 26120 89 12:06 ? 00:00:05 > > /usr/jdk1.6.0_03/bin/java > > > > -classpath /usr/local/groovy/lib/groovy-1.7.3.jar > > > > -Dscript.name=/usr/local/groovy/b > > > > root 26185 22437 0 12:07 pts/3 00:00:00 ps -aef > > > > > > > > > > > > *The command which i am executing is * > > > > > > > > > > > > hadoop jar > > > > > /usr/lib/hadoop-0.20/contrib/streaming/hadoop-streaming-0.20.2+320.jar > > \ > > > > -D mapred.child.java.opts=-Xmx1024m \ > > > > -inputformat StreamInputFormat \ > > > > -inputreader "StreamXmlRecordReader,begin=<mdc xmlns:HTML=\" > > > > http://www.w3.org/TR/REC-xml\ <http://www.w3.org/TR/REC-xml%5C> < > http://www.w3.org/TR/REC-xml%5C> < > > http://www.w3.org/TR/REC-xml%5C> < > > > http://www.w3.org/TR/REC-xml%5C>">,end=</mdc>" > > > > \ > > > > -input > > > > > > > > > > > > > > /user/root/RNCDATA/MDFDORKUCRAR02/A20100531.0000-0700-0015-0700_RNCCN-MDFDORKUCRAR02 > > > > \ > > > > -jobconf mapred.map.tasks=1 \ > > > > -jobconf mapred.reduce.tasks=0 \ > > > > -output RNC25 \ > > > > -mapper "/home/ftpuser1/Nodemapper5.groovy -Xmx2000m"\ > > > > -reducer org.apache.hadoop.mapred.lib.IdentityReducer \ > > > > -file /home/ftpuser1/Nodemapper5.groovy \ > > > > -jt local > > > > > > > > I have noticed that the all hadoop processes showing 2001 memory size > > > which > > > > i have set in hadoop-env.sh. and one the command, i give 2000 in > mapper > > > and > > > > 1024 in child.java.opts but i think these values(1024,2001) are not > in > > > use. > > > > secondly the following lines > > > > > > > > *hadoop 26120 25175 14 12:06 ? 00:00:01 > > > > /usr/jdk1.6.0_03/jre/bin/java > > > > > > > > > > > > > > -Djava.library.path=/usr/jdk1.6.0_03/jre/lib/i386/client:/usr/jdk1.6.0_03/jre/l > > > > hadoop 26162 26120 89 12:06 ? 00:00:05 > > /usr/jdk1.6.0_03/bin/java > > > > -classpath /usr/local/groovy/lib/groovy-1.7.3.jar > > > > -Dscript.name=/usr/local/groovy/b* > > > > > > > > did not appear for first time when job runs. they appear when job > > failed > > > > for > > > > first time and then again try to start mapping. I have one more > > question > > > > which is as all hadoop processes (namenode, datanode, tasktracker...) > > > > showing 2001 heapsize in process. will it means all the processes > > using > > > > 2001m of memory?? > > > > > > > > Regards > > > > Shuja > > > > > > > > > > > > On Mon, Jul 12, 2010 at 8:51 PM, Alex Kozlov <ale...@cloudera.com> > > > wrote: > > > > > > > > > Hi Shuja, > > > > > > > > > > I think you need to enclose the invocation string in quotes. Try: > > > > > > > > > > -mapper "/home/ftpuser1/Nodemapper5.groovy Xmx2000m" > > > > > > > > > > Also, it would be nice to see how exactly the groovy is invoked. > Is > > > > groovy > > > > > started and them gives you OOM or is OOM error during the start? > Can > > > you > > > > > see the new process with "ps -aef"? > > > > > > > > > > Can you run groovy in local mode? Try "-jt local" option. > > > > > > > > > > Thanks, > > > > > > > > > > Alex K > > > > > > > > > > On Mon, Jul 12, 2010 at 6:29 AM, Shuja Rehman < > shujamug...@gmail.com > > > > > > > > wrote: > > > > > > > > > > > Hi Patrick, > > > > > > Thanks for explanation. I have supply the heapsize in mapper in > the > > > > > > following way > > > > > > > > > > > > -mapper /home/ftpuser1/Nodemapper5.groovy Xmx2000m \ > > > > > > > > > > > > but still same error. Any other idea? > > > > > > Thanks > > > > > > > > > > > > On Mon, Jul 12, 2010 at 6:12 PM, Patrick Angeles < > > > patr...@cloudera.com > > > > > > >wrote: > > > > > > > > > > > > > Shuja, > > > > > > > > > > > > > > Those settings (mapred.child.jvm.opts and mapred.child.ulimit) > > are > > > > only > > > > > > > used > > > > > > > for child JVMs that get forked by the TaskTracker. You are > using > > > > Hadoop > > > > > > > streaming, which means the TaskTracker is forking a JVM for > > > > streaming, > > > > > > > which > > > > > > > is then forking a shell process that runs your groovy code (in > > > > another > > > > > > > JVM). > > > > > > > > > > > > > > I'm not much of a groovy expert, but if there's a way you can > > wrap > > > > your > > > > > > > code > > > > > > > around the MapReduce API that would work best. Otherwise, you > can > > > > just > > > > > > pass > > > > > > > the heapsize in '-mapper' argument. > > > > > > > > > > > > > > Regards, > > > > > > > > > > > > > > - Patrick > > > > > > > > > > > > > > On Mon, Jul 12, 2010 at 4:32 AM, Shuja Rehman < > > > shujamug...@gmail.com > > > > > > > > > > > > wrote: > > > > > > > > > > > > > > > Hi Alex, > > > > > > > > > > > > > > > > I have update the java to latest available version on all > > > machines > > > > in > > > > > > the > > > > > > > > cluster and now i run the job by adding this line > > > > > > > > > > > > > > > > -D mapred.child.ulimit=3145728 \ > > > > > > > > > > > > > > > > but still same error. Here is the output of this job. > > > > > > > > > > > > > > > > > > > > > > > > root 7845 5674 3 01:24 pts/1 00:00:00 > > > > > > /usr/jdk1.6.0_03/bin/java > > > > > > > > -Xmx10 23m -Dhadoop.log.dir=/usr/lib/hadoop-0.20/logs > > > > > > > > -Dhadoop.log.file=hadoop.log -Dha > > > > doop.home.dir=/usr/lib/hadoop-0.20 > > > > > > > > -Dhadoop.id.str= -Dhadoop.root.logger=INFO,co nsole > > > > > > > > -Dhadoop.policy.file=hadoop-policy.xml -classpath > > > > > > > /usr/lib/hadoop-0.20/con > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > f:/usr/jdk1.6.0_03/lib/tools.jar:/usr/lib/hadoop-0.20:/usr/lib/hadoop-0.20/hadoo > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > p-core-0.20.2+320.jar:/usr/lib/hadoop-0.20/lib/commons-cli-1.2.jar:/usr/lib/hado > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > op-0.20/lib/commons-codec-1.3.jar:/usr/lib/hadoop-0.20/lib/commons-el-1.0.jar:/u > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > sr/lib/hadoop-0.20/lib/commons-httpclient-3.0.1.jar:/usr/lib/hadoop-0.20/lib/com > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > mons-logging-1.0.4.jar:/usr/lib/hadoop-0.20/lib/commons-logging-api-1.0.4.jar:/u > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > sr/lib/hadoop-0.20/lib/commons-net-1.4.1.jar:/usr/lib/hadoop-0.20/lib/core-3.1.1 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > .jar:/usr/lib/hadoop-0.20/lib/hadoop-fairscheduler-0.20.2+320.jar:/usr/lib/hadoo > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > p-0.20/lib/hadoop-scribe-log4j-0.20.2+320.jar:/usr/lib/hadoop-0.20/lib/hsqldb-1. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > 8.0.10.jar:/usr/lib/hadoop-0.20/lib/hsqldb.jar:/usr/lib/hadoop-0.20/lib/jackson- > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > core-asl-1.0.1.jar:/usr/lib/hadoop-0.20/lib/jackson-mapper-asl-1.0.1.jar:/usr/li > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > b/hadoop-0.20/lib/jasper-compiler-5.5.12.jar:/usr/lib/hadoop-0.20/lib/jasper-run > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > time-5.5.12.jar:/usr/lib/hadoop-0.20/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20/l > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > ib/jetty-6.1.14.jar:/usr/lib/hadoop-0.20/lib/jetty-util-6.1.14.jar:/usr/lib/hado > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > op-0.20/lib/junit-4.5.jar:/usr/lib/hadoop-0.20/lib/kfs-0.2.2.jar:/usr/lib/hadoop > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -0.20/lib/libfb303.jar:/usr/lib/hadoop-0.20/lib/libthrift.jar:/usr/lib/hadoop-0. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > 20/lib/log4j-1.2.15.jar:/usr/lib/hadoop-0.20/lib/mockito-all-1.8.2.jar:/usr/lib/ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > hadoop-0.20/lib/mysql-connector-java-5.0.8-bin.jar:/usr/lib/hadoop-0.20/lib/oro- > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > 2.0.8.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-6.1.14.jar:/usr/lib/hadoop-0. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > 20/lib/slf4j-api-1.4.3.jar:/usr/lib/hadoop-0.20/lib/slf4j-log4j12-1.4.3.jar:/usr > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > /lib/hadoop-0.20/lib/xmlenc-0.52.jar:/usr/lib/hadoop-0.20/lib/jsp-2.1/jsp-2.1.ja > > > > > > > > r:/usr/lib/hadoop-0.20/lib/jsp-2.1/jsp-api-2.1.jar > > > > > > > > org.apache.hadoop.util.RunJar > > > > > > > > > > > > > > > /usr/lib/hadoop-0.20/contrib/streaming/hadoop-streaming-0.20.2+320.jar > > > > > > -D > > > > > > > > mapre d.child.java.opts=-Xmx2000M -D > > mapred.child.ulimit=3145728 > > > > > > > > -inputformat StreamIn putFormat -inputreader > > > > > > > > StreamXmlRecordReader,begin=<mdc xmlns:HTML="http://www.w > > > > > > > > 3.org/TR/REC-xml">,end=</mdc> > > > > > > > > -input /user/root/RNCDATA/MDFDORKUCRAR02/A20100531 > > > > > > > > .0000-0700-0015-0700_RNCCN-MDFDORKUCRAR02 -jobconf > > > > mapred.map.tasks=1 > > > > > > > > -jobconf m apred.reduce.tasks=0 -output RNC14 -mapper > > > > > > > > /home/ftpuser1/Nodemapper5.groovy -re ducer > > > > > > > > org.apache.hadoop.mapred.lib.IdentityReducer -file > > > > > > > /home/ftpuser1/Nodemapp > > > > > > > > er5.groovy > > > > > > > > root 7930 7632 0 01:24 pts/2 00:00:00 grep > > > > > Nodemapper5.groovy > > > > > > > > > > > > > > > > > > > > > > > > Any clue? > > > > > > > > Thanks > > > > > > > > > > > > > > > > On Sun, Jul 11, 2010 at 3:44 AM, Alex Kozlov < > > > ale...@cloudera.com> > > > > > > > wrote: > > > > > > > > > > > > > > > > > Hi Shuja, > > > > > > > > > > > > > > > > > > First, thank you for using CDH3. Can you also check what > m* > > > > > > > > > apred.child.ulimit* you are using? Try adding "* > > > > > > > > > -D mapred.child.ulimit=3145728*" to the command line. > > > > > > > > > > > > > > > > > > I would also recommend to upgrade java to JDK 1.6 update 8 > at > > a > > > > > > > minimum, > > > > > > > > > which you can download from the Java SE > > > > > > > > > Homepage<http://java.sun.com/javase/downloads/index.jsp> > > > > > > > > > . > > > > > > > > > > > > > > > > > > Let me know how it goes. > > > > > > > > > > > > > > > > > > Alex K > > > > > > > > > > > > > > > > > > On Sat, Jul 10, 2010 at 12:59 PM, Shuja Rehman < > > > > > > shujamug...@gmail.com > > > > > > > > > >wrote: > > > > > > > > > > > > > > > > > > > Hi Alex > > > > > > > > > > > > > > > > > > > > Yeah, I am running a job on cluster of 2 machines and > using > > > > > > Cloudera > > > > > > > > > > distribution of hadoop. and here is the output of this > > > command. > > > > > > > > > > > > > > > > > > > > root 5277 5238 3 12:51 pts/2 00:00:00 > > > > > > > > /usr/jdk1.6.0_03/bin/java > > > > > > > > > > -Xmx1023m -Dhadoop.log.dir=/usr/lib > > /hadoop-0.20/logs > > > > > > > > > > -Dhadoop.log.file=hadoop.log > > > > > -Dhadoop.home.dir=/usr/lib/hadoop-0.20 > > > > > > > > > > -Dhadoop.id.str= -Dhado > op.root.logger=INFO,console > > > > > > > > > > -Dhadoop.policy.file=hadoop-policy.xml -classpath > > > > > > > > > > /usr/lib/hadoop-0.20/conf:/usr/ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > jdk1.6.0_03/lib/tools.jar:/usr/lib/hadoop-0.20:/usr/lib/hadoop-0.20/hadoop-core-0.20.2+320.jar:/usr/lib/hadoo > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > p-0.20/lib/commons-cli-1.2.jar:/usr/lib/hadoop-0.20/lib/commons-codec-1.3.jar:/usr/lib/hadoop-0.20/lib/common > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > s-el-1.0.jar:/usr/lib/hadoop-0.20/lib/commons-httpclient-3.0.1.jar:/usr/lib/hadoop-0.20/lib/commons-logging-1 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > .0.4.jar:/usr/lib/hadoop-0.20/lib/commons-logging-api-1.0.4.jar:/usr/lib/hadoop-0.20/lib/commons-net-1.4.1.ja > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > r:/usr/lib/hadoop-0.20/lib/core-3.1.1.jar:/usr/lib/hadoop-0.20/lib/hadoop-fairscheduler-0.20.2+320.jar:/usr/l > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > ib/hadoop-0.20/lib/hadoop-scribe-log4j-0.20.2+320.jar:/usr/lib/hadoop-0.20/lib/hsqldb-1.8.0.10.jar:/usr/lib/h > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > adoop-0.20/lib/hsqldb.jar:/usr/lib/hadoop-0.20/lib/jackson-core-asl-1.0.1.jar:/usr/lib/hadoop-0.20/lib/jackso > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > n-mapper-asl-1.0.1.jar:/usr/lib/hadoop-0.20/lib/jasper-compiler-5.5.12.jar:/usr/lib/hadoop-0.20/lib/jasper-ru > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > ntime-5.5.12.jar:/usr/lib/hadoop-0.20/lib/jets3t-0.6.1.jar:/usr/lib/hadoop-0.20/lib/jetty-6.1.14.jar:/usr/lib > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > /hadoop-0.20/lib/jetty-util-6.1.14.jar:/usr/lib/hadoop-0.20/lib/junit-4.5.jar:/usr/lib/hadoop-0.20/lib/kfs-0. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > 2.2.jar:/usr/lib/hadoop-0.20/lib/libfb303.jar:/usr/lib/hadoop-0.20/lib/libthrift.jar:/usr/lib/hadoop-0.20/lib > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > /log4j-1.2.15.jar:/usr/lib/hadoop-0.20/lib/mockito-all-1.8.2.jar:/usr/lib/hadoop-0.20/lib/mysql-connector-jav > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > a-5.0.8-bin.jar:/usr/lib/hadoop-0.20/lib/oro-2.0.8.jar:/usr/lib/hadoop-0.20/lib/servlet-api-2.5-6.1.14.jar:/u > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > sr/lib/hadoop-0.20/lib/slf4j-api-1.4.3.jar:/usr/lib/hadoop-0.20/lib/slf4j-log4j12-1.4.3.jar:/usr/lib/hadoop-0 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > .20/lib/xmlenc-0.52.jar:/usr/lib/hadoop-0.20/lib/jsp-2.1/jsp-2.1.jar:/usr/lib/hadoop-0.20/lib/jsp-2.1/jsp-api > > > > > > > > > > -2.1.jar org.apache.hadoop.util.RunJar > > > > > > > > > > > > > > > > > > > > > > /usr/lib/hadoop-0.20/contrib/streaming/hadoop-streaming-0.20.2+320.jar > > > > > > > > > > -D mapred.child.java.opts=-Xmx2000M -inputformat > > > > > StreamInputFormat > > > > > > > > > > -inputreader StreamXmlRecordReader,begin= <mdc > > > > > xmlns:HTML=" > > > > > > > > > > http://www.w3.org/TR/REC-xml">,end=</mdc> -input > > > > > > > > > > /user/root/RNCDATA/MDFDORKUCRAR02/A20100531 > > > > > > > > > > .0000-0700-0015-0700_RNCCN-MDFDORKUCRAR02 -jobconf > > > > > > mapred.map.tasks=1 > > > > > > > > > > -jobconf mapred.reduce.tasks=0 -output RNC11 > > -mapper > > > > > > > > > > /home/ftpuser1/Nodemapper5.groovy -reducer > > > > > > > > > > org.apache.hadoop.mapred.lib.IdentityReducer -file / > > > > > > > > > > home/ftpuser1/Nodemapper5.groovy > > > > > > > > > > root 5360 5074 0 12:51 pts/1 00:00:00 grep > > > > > > > Nodemapper5.groovy > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > ------------------------------------------------------------------------------------------------------------------------------ > > > > > > > > > > and what is meant by OOM and thanks for helping, > > > > > > > > > > > > > > > > > > > > Best Regards > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On Sun, Jul 11, 2010 at 12:30 AM, Alex Kozlov < > > > > > ale...@cloudera.com > > > > > > > > > > > > > > > > wrote: > > > > > > > > > > > > > > > > > > > > > Hi Shuja, > > > > > > > > > > > > > > > > > > > > > > It looks like the OOM is happening in your code. Are > you > > > > > running > > > > > > > > > > MapReduce > > > > > > > > > > > in a cluster? If so, can you send the exact command > line > > > > your > > > > > > code > > > > > > > > is > > > > > > > > > > > invoked with -- you can get it with a 'ps -Af | grep > > > > > > > > > Nodemapper5.groovy' > > > > > > > > > > > command on one of the nodes which is running the task? > > > > > > > > > > > > > > > > > > > > > > Thanks, > > > > > > > > > > > > > > > > > > > > > > Alex K > > > > > > > > > > > > > > > > > > > > > > On Sat, Jul 10, 2010 at 10:40 AM, Shuja Rehman < > > > > > > > > shujamug...@gmail.com > > > > > > > > > > > >wrote: > > > > > > > > > > > > > > > > > > > > > > > Hi All > > > > > > > > > > > > > > > > > > > > > > > > I am facing a hard problem. I am running a map reduce > > job > > > > > using > > > > > > > > > > streaming > > > > > > > > > > > > but it fails and it gives the following error. > > > > > > > > > > > > > > > > > > > > > > > > Caught: java.lang.OutOfMemoryError: Java heap space > > > > > > > > > > > > at Nodemapper5.parseXML(Nodemapper5.groovy:25) > > > > > > > > > > > > > > > > > > > > > > > > java.lang.RuntimeException: > > > PipeMapRed.waitOutputThreads(): > > > > > > > > > subprocess > > > > > > > > > > > > failed with code 1 > > > > > > > > > > > > at > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > org.apache.hadoop.streaming.PipeMapRed.waitOutputThreads(PipeMapRed.java:362) > > > > > > > > > > > > at > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > org.apache.hadoop.streaming.PipeMapRed.mapRedFinished(PipeMapRed.java:572) > > > > > > > > > > > > > > > > > > > > > > > > at > > > > > > > > > > > > > > > > org.apache.hadoop.streaming.PipeMapper.close(PipeMapper.java:136) > > > > > > > > > > > > at > > > > > > > org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:57) > > > > > > > > > > > > at > > > > > > > > > > > > > > > > > > > > > > > > org.apache.hadoop.streaming.PipeMapRunner.run(PipeMapRunner.java:36) > > > > > > > > > > > > at > > > > > > > > > > > > > org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:358) > > > > > > > > > > > > > > > > > > > > > > > > at > > > > > > org.apache.hadoop.mapred.MapTask.run(MapTask.java:307) > > > > > > > > > > > > at > > > > org.apache.hadoop.mapred.Child.main(Child.java:170) > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > I have increased the heap size in hadoop-env.sh and > > make > > > it > > > > > > > 2000M. > > > > > > > > > Also > > > > > > > > > > I > > > > > > > > > > > > tell the job manually by following line. > > > > > > > > > > > > > > > > > > > > > > > > -D mapred.child.java.opts=-Xmx2000M \ > > > > > > > > > > > > > > > > > > > > > > > > but it still gives the error. The same job runs fine > if > > i > > > > run > > > > > > on > > > > > > > > > shell > > > > > > > > > > > > using > > > > > > > > > > > > 1024M heap size like > > > > > > > > > > > > > > > > > > > > > > > > cat file.xml | /root/Nodemapper5.groovy > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Any clue????????? > > > > > > > > > > > > > > > > > > > > > > > > Thanks in advance. > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > > > > > > > > Regards > > > > > > > > > > > > Shuja-ur-Rehman Baig > > > > > > > > > > > > _________________________________ > > > > > > > > > > > > MS CS - School of Science and Engineering > > > > > > > > > > > > Lahore University of Management Sciences (LUMS) > > > > > > > > > > > > Sector U, DHA, Lahore, 54792, Pakistan > > > > > > > > > > > > Cell: +92 3214207445 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > > > > > > Regards > > > > > > > > > > Shuja-ur-Rehman Baig > > > > > > > > > > _________________________________ > > > > > > > > > > MS CS - School of Science and Engineering > > > > > > > > > > Lahore University of Management Sciences (LUMS) > > > > > > > > > > Sector U, DHA, Lahore, 54792, Pakistan > > > > > > > > > > Cell: +92 3214207445 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > > > > Regards > > > > > > > > Shuja-ur-Rehman Baig > > > > > > > > _________________________________ > > > > > > > > MS CS - School of Science and Engineering > > > > > > > > Lahore University of Management Sciences (LUMS) > > > > > > > > Sector U, DHA, Lahore, 54792, Pakistan > > > > > > > > Cell: +92 3214207445 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > > Regards > > > > > > Shuja-ur-Rehman Baig > > > > > > _________________________________ > > > > > > MS CS - School of Science and Engineering > > > > > > Lahore University of Management Sciences (LUMS) > > > > > > Sector U, DHA, Lahore, 54792, Pakistan > > > > > > Cell: +92 3214207445 > > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > Regards > > > > Shuja-ur-Rehman Baig > > > > _________________________________ > > > > MS CS - School of Science and Engineering > > > > Lahore University of Management Sciences (LUMS) > > > > Sector U, DHA, Lahore, 54792, Pakistan > > > > Cell: +92 3214207445 > > > > > > > > > > > > > > > -- > > Regards > > Shuja-ur-Rehman Baig > > _________________________________ > > MS CS - School of Science and Engineering > > Lahore University of Management Sciences (LUMS) > > Sector U, DHA, Lahore, 54792, Pakistan > > Cell: +92 3214207445 > > > -- Regards Shuja-ur-Rehman Baig _________________________________ MS CS - School of Science and Engineering Lahore University of Management Sciences (LUMS) Sector U, DHA, Lahore, 54792, Pakistan Cell: +92 3214207445