Appmaster error
Hi, I am trying to start an server within the application master's container alone.But when i tried using Runtime.getRuntime.exec("command").But it throws the following execption. Application application_1325738010393_0003 failed 1 times due to AM Container for appattempt_1325738010393_0003_01 exited with exitCode: 143 due to: Container [pid=7212,containerID=container_1325738010393_0003_01_01] is running beyond virtual memory limits. Current usage: 118.4mb of 1.0gb physical memory used; 2.7gb of 2.1gb virtual memory used. Killing container. Dump of the process-tree for container_1325738010393_0003_01_0 When i tried using single node yarn cluster everything works fine.But in multi node it throws this exception.Should i increase the size of /tmp in linux... Any ideas
Appmasgter error
Hi, I am trying to start an server within the application master's container alone.But when i tried using Runtime.getRuntime.exec("command").But it throws the following execption. Application application_1325738010393_0003 failed 1 times due to AM Container for appattempt_1325738010393_0003_01 exited with exitCode: 143 due to: Container [pid=7212,containerID=container_1325738010393_0003_01_01] is running beyond virtual memory limits. Current usage: 118.4mb of 1.0gb physical memory used; 2.7gb of 2.1gb virtual memory used. Killing container. Dump of the process-tree for container_1325738010393_0003_01_0 When i tried using single node yarn cluster everything works fine.But in multi node it throws this exception.Should i increase the size of /tmp in linux... Any ideas
Launching containers in specific host
Hi, I tried to set the client node for launching the container within the application master. I have set the parameter as request.setHostName("client"); but the containers are not launched in the destined host.Instead the loop goes on continuously. 2012-01-04 15:11:48,535 INFO appmaster.ApplicationMaster (ApplicationMaster.java:run( 204)) - Current application state: loop=95, appDone=false, total=2, requested=2, completed=0, failed=0, currentAllocated=0
RE: Balancer exiting immediately despite having work to do.
James, http://pastebin.com/mYBRKDew Tomorrow I'll run the balancer again and grab a copy of the namenode logs as well. Didn't think of that today. -Landy -Original Message- From: jameswarr...@gmail.com [mailto:jameswarr...@gmail.com] On Behalf Of James Warren Sent: Wednesday, January 04, 2012 7:49 PM To: common-user@hadoop.apache.org Subject: Re: Balancer exiting immediately despite having work to do. Hi Landy - Attachments are stripped from e-mails sent to the mailing list. Could you publish your logs on pastebin and forward the url? cheers, -James On Wed, Jan 4, 2012 at 10:03 AM, Bible, Landy wrote: > Hi all, > > ** ** > > I'm running Hadoop 0.20.2. The balancer has suddenly stopped working. > I'm attempting to balance the cluster with a threshold of 1, using the > following command: > > ** ** > > ./hadoop balancer -threshold 1 > > ** ** > > This has been working fine, but suddenly it isn't. It skips though 5 > iterations without actually doing any work: > > ** ** > > Time Stamp Iteration# Bytes Already Moved Bytes Left To > Move Bytes Being Moved > > Jan 4, 2012 11:47:56 AM 0 0 KB 1.87 > GB6.68 GB > > Jan 4, 2012 11:47:56 AM 1 0 KB 1.87 > GB6.68 GB > > Jan 4, 2012 11:47:56 AM 2 0 KB 1.87 > GB6.68 GB > > Jan 4, 2012 11:47:57 AM 3 0 KB 1.87 > GB6.68 GB > > Jan 4, 2012 11:47:57 AM 4 0 KB 1.87 > GB6.68 GB > > No block has been moved for 5 iterations. Exiting... > > Balancing took 524.0 milliseconds > > ** ** > > I've attached the full log, but I can't see any errors indicating why > it is failing. Any ideas? I'd really like to get balancing working again. > My use case isn't the norm, and it is important that the cluster stay > as close to completely balanced as possible. > > ** ** > > -- > > Landy Bible > > ** ** > > Simulation and Computer Specialist > > School of Nursing - Collins College of Business > > The University of Tulsa > > ** ** >
Re: datanode failing to start.
hadoop-root-datanode-Hoot-2.local.out only shows the same Exception. d On 1/4/2012 2:37 PM, Dave Kelsey wrote: java version 1.6.0_29 hadoop: 0.20.203.0 I'm attempting to setup the pseudo-distributed config on a mac 10.6.8. I followed the steps from the QuickStart (http://wiki.apache.org./hadoop/QuickStart) and succeeded with Stage 1: Standalone Operation. I followed the steps for Stage 2: Pseudo-distributed Configuration. I set the JAVA_HOME variable in conf/hadoop-env.sh and I changed tools.jar to the location of classes.jar (a mac version of tools.jar) I've modified the three .xml files as described in the QuickStart. ssh'ing to localhost has been configured and works with passwordless authentication. I formatted the namenode with "bin/hadoop namenode -format" as the instructions say This is what I see when I run bin/start-all.sh root# bin/start-all.sh starting namenode, logging to /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-namenode-Hoot-2.local.out localhost: starting datanode, logging to /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-datanode-Hoot-2.local.out localhost: Exception in thread "main" java.lang.NoClassDefFoundError: server localhost: Caused by: java.lang.ClassNotFoundException: server localhost: at java.net.URLClassLoader$1.run(URLClassLoader.java:202) localhost: at java.security.AccessController.doPrivileged(Native Method) localhost: at java.net.URLClassLoader.findClass(URLClassLoader.java:190) localhost: at java.lang.ClassLoader.loadClass(ClassLoader.java:306) localhost: at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) localhost: at java.lang.ClassLoader.loadClass(ClassLoader.java:247) localhost: starting secondarynamenode, logging to /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-secondarynamenode-Hoot-2.local.out starting jobtracker, logging to /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-jobtracker-Hoot-2.local.out localhost: starting tasktracker, logging to /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-tasktracker-Hoot-2.local.out There are 4 processes running: ps -fax | grep hadoop | grep -v grep | wc -l 4 They are: SecondaryNameNode TaskTracker NameNode JobTracker I've searched to see if anyone else has encountered this and not found anything Dave Kelsey
Re: Balancer exiting immediately despite having work to do.
Hi Landy - Attachments are stripped from e-mails sent to the mailing list. Could you publish your logs on pastebin and forward the url? cheers, -James On Wed, Jan 4, 2012 at 10:03 AM, Bible, Landy wrote: > Hi all, > > ** ** > > I’m running Hadoop 0.20.2. The balancer has suddenly stopped working. > I’m attempting to balance the cluster with a threshold of 1, using the > following command: > > ** ** > > ./hadoop balancer –threshold 1 > > ** ** > > This has been working fine, but suddenly it isn’t. It skips though 5 > iterations without actually doing any work: > > ** ** > > Time Stamp Iteration# Bytes Already Moved Bytes Left To > Move Bytes Being Moved > > Jan 4, 2012 11:47:56 AM 0 0 KB 1.87 > GB6.68 GB > > Jan 4, 2012 11:47:56 AM 1 0 KB 1.87 > GB6.68 GB > > Jan 4, 2012 11:47:56 AM 2 0 KB 1.87 > GB6.68 GB > > Jan 4, 2012 11:47:57 AM 3 0 KB 1.87 > GB6.68 GB > > Jan 4, 2012 11:47:57 AM 4 0 KB 1.87 > GB6.68 GB > > No block has been moved for 5 iterations. Exiting... > > Balancing took 524.0 milliseconds > > ** ** > > I’ve attached the full log, but I can’t see any errors indicating why it > is failing. Any ideas? I’d really like to get balancing working again. > My use case isn’t the norm, and it is important that the cluster stay as > close to completely balanced as possible. > > ** ** > > -- > > Landy Bible > > ** ** > > Simulation and Computer Specialist > > School of Nursing – Collins College of Business > > The University of Tulsa > > ** ** >
Re: datanode failing to start.
Have you checked your logs? --Joey On Wed, Jan 4, 2012 at 4:37 PM, Dave Kelsey wrote: > > java version 1.6.0_29 > hadoop: 0.20.203.0 > > I'm attempting to setup the pseudo-distributed config on a mac 10.6.8. > I followed the steps from the QuickStart > (http://wiki.apache.org./hadoop/QuickStart) and succeeded with Stage 1: > Standalone Operation. > I followed the steps for Stage 2: Pseudo-distributed Configuration. > I set the JAVA_HOME variable in conf/hadoop-env.sh and I changed tools.jar > to the location of classes.jar (a mac version of tools.jar) > I've modified the three .xml files as described in the QuickStart. > ssh'ing to localhost has been configured and works with passwordless > authentication. > I formatted the namenode with "bin/hadoop namenode -format" as the > instructions say > > This is what I see when I run bin/start-all.sh > > root# bin/start-all.sh > starting namenode, logging to > /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-namenode-Hoot-2.local.out > localhost: starting datanode, logging to > /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-datanode-Hoot-2.local.out > localhost: Exception in thread "main" java.lang.NoClassDefFoundError: server > localhost: Caused by: java.lang.ClassNotFoundException: server > localhost: at java.net.URLClassLoader$1.run(URLClassLoader.java:202) > localhost: at java.security.AccessController.doPrivileged(Native Method) > localhost: at java.net.URLClassLoader.findClass(URLClassLoader.java:190) > localhost: at java.lang.ClassLoader.loadClass(ClassLoader.java:306) > localhost: at > sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) > localhost: at java.lang.ClassLoader.loadClass(ClassLoader.java:247) > localhost: starting secondarynamenode, logging to > /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-secondarynamenode-Hoot-2.local.out > starting jobtracker, logging to > /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-jobtracker-Hoot-2.local.out > localhost: starting tasktracker, logging to > /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-tasktracker-Hoot-2.local.out > > There are 4 processes running: > ps -fax | grep hadoop | grep -v grep | wc -l > 4 > > They are: > SecondaryNameNode > TaskTracker > NameNode > JobTracker > > > I've searched to see if anyone else has encountered this and not found > anything > > Dave Kelsey >
datanode failing to start.
java version 1.6.0_29 hadoop: 0.20.203.0 I'm attempting to setup the pseudo-distributed config on a mac 10.6.8. I followed the steps from the QuickStart (http://wiki.apache.org./hadoop/QuickStart) and succeeded with Stage 1: Standalone Operation. I followed the steps for Stage 2: Pseudo-distributed Configuration. I set the JAVA_HOME variable in conf/hadoop-env.sh and I changed tools.jar to the location of classes.jar (a mac version of tools.jar) I've modified the three .xml files as described in the QuickStart. ssh'ing to localhost has been configured and works with passwordless authentication. I formatted the namenode with "bin/hadoop namenode -format" as the instructions say This is what I see when I run bin/start-all.sh root# bin/start-all.sh starting namenode, logging to /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-namenode-Hoot-2.local.out localhost: starting datanode, logging to /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-datanode-Hoot-2.local.out localhost: Exception in thread "main" java.lang.NoClassDefFoundError: server localhost: Caused by: java.lang.ClassNotFoundException: server localhost: at java.net.URLClassLoader$1.run(URLClassLoader.java:202) localhost: at java.security.AccessController.doPrivileged(Native Method) localhost: at java.net.URLClassLoader.findClass(URLClassLoader.java:190) localhost: at java.lang.ClassLoader.loadClass(ClassLoader.java:306) localhost: at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) localhost: at java.lang.ClassLoader.loadClass(ClassLoader.java:247) localhost: starting secondarynamenode, logging to /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-secondarynamenode-Hoot-2.local.out starting jobtracker, logging to /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-jobtracker-Hoot-2.local.out localhost: starting tasktracker, logging to /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-tasktracker-Hoot-2.local.out There are 4 processes running: ps -fax | grep hadoop | grep -v grep | wc -l 4 They are: SecondaryNameNode TaskTracker NameNode JobTracker I've searched to see if anyone else has encountered this and not found anything Dave Kelsey
Setting HDFS directory time programmatically
Hi - Is it possible to set the access time of a HDFS directory programmatically? I’m using 0.20.204.0. I need to do that in unit tests, where my clean up program is going to remove files/dirs whose access time is too far in the past. I can setTimes on the test files without any problem, but not on the directories... The directories created automatically when I create the test fiels have a date (with getAccessTime) of 1969/12/31 16:00 and I can’t control that date, which makes my unit testing impossible. By the way, setTimes doesn’t allow to set the date on dirs, but getAccessTime is happy to return a date, which is inconsistent, IMHO. Finally, on our production systems, I’m seeing appropriate dates for both files and directories. Any insight appreciated, Thanks! Frank
Balancer exiting immediately despite having work to do.
Hi all, I'm running Hadoop 0.20.2. The balancer has suddenly stopped working. I'm attempting to balance the cluster with a threshold of 1, using the following command: ./hadoop balancer -threshold 1 This has been working fine, but suddenly it isn't. It skips though 5 iterations without actually doing any work: Time Stamp Iteration# Bytes Already Moved Bytes Left To Move Bytes Being Moved Jan 4, 2012 11:47:56 AM 0 0 KB 1.87 GB 6.68 GB Jan 4, 2012 11:47:56 AM 1 0 KB 1.87 GB 6.68 GB Jan 4, 2012 11:47:56 AM 2 0 KB 1.87 GB 6.68 GB Jan 4, 2012 11:47:57 AM 3 0 KB 1.87 GB 6.68 GB Jan 4, 2012 11:47:57 AM 4 0 KB 1.87 GB 6.68 GB No block has been moved for 5 iterations. Exiting... Balancing took 524.0 milliseconds I've attached the full log, but I can't see any errors indicating why it is failing. Any ideas? I'd really like to get balancing working again. My use case isn't the norm, and it is important that the cluster stay as close to completely balanced as possible. -- Landy Bible Simulation and Computer Specialist School of Nursing - Collins College of Business The University of Tulsa
Re: How do i customize the name of the job during submission?
Arun, You can't change the job IDs. Why do you seek to do this though? You can, however, specify a custom job 'name' via the regular JobConf/Job APIs. On 04-Jan-2012, at 9:31 PM, ArunKumar wrote: > Hi guys ! > > When a Job is submitted it is given an ID say job_200904211745_0018 in > Hadoop. > > But for some reason i want to submit it with ID say "job1". How can i do > that ? > > Arun > > -- > View this message in context: > http://lucene.472066.n3.nabble.com/How-do-i-customize-the-name-of-the-job-during-submission-tp3632228p3632228.html > Sent from the Hadoop lucene-users mailing list archive at Nabble.com.
Task location determination
I have a test Hadoop cluster set up using Cloudera. It consists of the Name Node and three Data Nodes. When I submit jobs, they end up piling up on one node instead of round robining through the different nodes. I understand that Hadoop tries to run the job where the data is located, but with only three data nodes and a replication factor of 3, wouldn't that mean that the same data is on every single machine? Why would it not spread the tasking out over all of the machines instead of clumping up on one, leaving the others idle? Thanks.