Appmaster error

2012-01-04 Thread raghavendhra rahul
Hi,
   I am trying to start an server within the application
master's container alone.But when i tried using
Runtime.getRuntime.exec("command").But it throws the following execption.
Application application_1325738010393_0003 failed 1 times due to AM
Container for appattempt_1325738010393_0003_01 exited with exitCode:
143 due to: Container
[pid=7212,containerID=container_1325738010393_0003_01_01] is running
beyond virtual memory limits. Current usage: 118.4mb of 1.0gb physical
memory used; 2.7gb of 2.1gb virtual memory used. Killing container. Dump of
the process-tree for container_1325738010393_0003_01_0

When i tried using single node yarn cluster everything works fine.But in
multi node it throws this exception.Should i increase the size of /tmp in
linux...
Any ideas


Appmasgter error

2012-01-04 Thread raghavendhra rahul
Hi,
   I am trying to start an server within the application
master's container alone.But when i tried using
Runtime.getRuntime.exec("command").But it throws the following execption.
Application application_1325738010393_0003 failed 1 times due to AM
Container for appattempt_1325738010393_0003_01 exited with exitCode:
143 due to: Container
[pid=7212,containerID=container_1325738010393_0003_01_01] is running
beyond virtual memory limits. Current usage: 118.4mb of 1.0gb physical
memory used; 2.7gb of 2.1gb virtual memory used. Killing container. Dump of
the process-tree for container_1325738010393_0003_01_0

When i tried using single node yarn cluster everything works fine.But in
multi node it throws this exception.Should i increase the size of /tmp in
linux...
Any ideas


Launching containers in specific host

2012-01-04 Thread raghavendhra rahul
Hi,

I tried to set the client node for launching the container within the
application master.
I have set the parameter as
request.setHostName("client");
but the containers are not launched in the destined host.Instead the loop
goes on continuously.
2012-01-04 15:11:48,535 INFO appmaster.ApplicationMaster
(ApplicationMaster.java:run(
204)) - Current application state: loop=95, appDone=false, total=2,
requested=2, completed=0, failed=0, currentAllocated=0


RE: Balancer exiting immediately despite having work to do.

2012-01-04 Thread Bible, Landy
James,

http://pastebin.com/mYBRKDew

Tomorrow I'll run the balancer again and grab a copy of the namenode logs as 
well.  Didn't think of that today.

-Landy

-Original Message-
From: jameswarr...@gmail.com [mailto:jameswarr...@gmail.com] On Behalf Of James 
Warren
Sent: Wednesday, January 04, 2012 7:49 PM
To: common-user@hadoop.apache.org
Subject: Re: Balancer exiting immediately despite having work to do.

Hi Landy -

Attachments are stripped from e-mails sent to the mailing list.  Could you 
publish your logs on pastebin and forward the url?

cheers,
-James

On Wed, Jan 4, 2012 at 10:03 AM, Bible, Landy wrote:

> Hi all,
>
> ** **
>
> I'm running Hadoop 0.20.2.  The balancer has suddenly stopped working.
> I'm attempting to balance the cluster with a threshold of 1, using the 
> following command:
>
> ** **
>
> ./hadoop balancer -threshold 1
>
> ** **
>
> This has been working fine, but suddenly it isn't.  It skips though 5 
> iterations without actually doing any work:
>
> ** **
>
> Time Stamp   Iteration#  Bytes Already Moved  Bytes Left To
> Move  Bytes Being Moved
>
> Jan 4, 2012 11:47:56 AM   0 0 KB 1.87
> GB6.68 GB
>
> Jan 4, 2012 11:47:56 AM   1 0 KB 1.87
> GB6.68 GB
>
> Jan 4, 2012 11:47:56 AM   2 0 KB 1.87
> GB6.68 GB
>
> Jan 4, 2012 11:47:57 AM   3 0 KB 1.87
> GB6.68 GB
>
> Jan 4, 2012 11:47:57 AM   4 0 KB 1.87
> GB6.68 GB
>
> No block has been moved for 5 iterations. Exiting...
>
> Balancing took 524.0 milliseconds
>
> ** **
>
> I've attached the full log, but I can't see any errors indicating why 
> it is failing.  Any ideas?  I'd really like to get balancing working again.
> My use case isn't the norm, and it is important that the cluster stay 
> as close to completely balanced as possible.
>
> ** **
>
> --
>
> Landy Bible
>
> ** **
>
> Simulation and Computer Specialist
>
> School of Nursing - Collins College of Business
>
> The University of Tulsa
>
> ** **
>


Re: datanode failing to start.

2012-01-04 Thread Dave Kelsey

hadoop-root-datanode-Hoot-2.local.out only shows the same Exception.

d

On 1/4/2012 2:37 PM, Dave Kelsey wrote:


java version 1.6.0_29
hadoop: 0.20.203.0

I'm attempting to setup the pseudo-distributed config on a mac 10.6.8.
I followed the steps from the QuickStart 
(http://wiki.apache.org./hadoop/QuickStart) and succeeded with Stage 
1: Standalone Operation.

I followed the steps for Stage 2: Pseudo-distributed Configuration.
I set the JAVA_HOME variable in conf/hadoop-env.sh and I changed 
tools.jar to the location of classes.jar (a mac version of tools.jar)

I've modified the three .xml files as described in the QuickStart.
ssh'ing to localhost has been configured and works with passwordless 
authentication.
I formatted the namenode with "bin/hadoop namenode -format" as the 
instructions say


This is what I see when I run bin/start-all.sh

root# bin/start-all.sh
starting namenode, logging to 
/Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-namenode-Hoot-2.local.out
localhost: starting datanode, logging to 
/Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-datanode-Hoot-2.local.out
localhost: Exception in thread "main" java.lang.NoClassDefFoundError: 
server

localhost: Caused by: java.lang.ClassNotFoundException: server
localhost: at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
localhost: at java.security.AccessController.doPrivileged(Native 
Method)
localhost: at 
java.net.URLClassLoader.findClass(URLClassLoader.java:190)

localhost: at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
localhost: at 
sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)

localhost: at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
localhost: starting secondarynamenode, logging to 
/Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-secondarynamenode-Hoot-2.local.out
starting jobtracker, logging to 
/Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-jobtracker-Hoot-2.local.out
localhost: starting tasktracker, logging to 
/Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-tasktracker-Hoot-2.local.out


There are 4 processes running:
ps -fax | grep hadoop | grep -v grep | wc -l
  4

They are:
SecondaryNameNode
TaskTracker
NameNode
JobTracker


I've searched to see if anyone else has encountered this and not found 
anything


Dave Kelsey



Re: Balancer exiting immediately despite having work to do.

2012-01-04 Thread James Warren
Hi Landy -

Attachments are stripped from e-mails sent to the mailing list.  Could you
publish your logs on pastebin and forward the url?

cheers,
-James

On Wed, Jan 4, 2012 at 10:03 AM, Bible, Landy wrote:

> Hi all,
>
> ** **
>
> I’m running Hadoop 0.20.2.  The balancer has suddenly stopped working.
> I’m attempting to balance the cluster with a threshold of 1, using the
> following command:
>
> ** **
>
> ./hadoop balancer –threshold 1
>
> ** **
>
> This has been working fine, but suddenly it isn’t.  It skips though 5
> iterations without actually doing any work:
>
> ** **
>
> Time Stamp   Iteration#  Bytes Already Moved  Bytes Left To
> Move  Bytes Being Moved
>
> Jan 4, 2012 11:47:56 AM   0 0 KB 1.87
> GB6.68 GB
>
> Jan 4, 2012 11:47:56 AM   1 0 KB 1.87
> GB6.68 GB
>
> Jan 4, 2012 11:47:56 AM   2 0 KB 1.87
> GB6.68 GB
>
> Jan 4, 2012 11:47:57 AM   3 0 KB 1.87
> GB6.68 GB
>
> Jan 4, 2012 11:47:57 AM   4 0 KB 1.87
> GB6.68 GB
>
> No block has been moved for 5 iterations. Exiting...
>
> Balancing took 524.0 milliseconds
>
> ** **
>
> I’ve attached the full log, but I can’t see any errors indicating why it
> is failing.  Any ideas?  I’d really like to get balancing working again.
> My use case isn’t the norm, and it is important that the cluster stay as
> close to completely balanced as possible.
>
> ** **
>
> --
>
> Landy Bible
>
> ** **
>
> Simulation and Computer Specialist
>
> School of Nursing – Collins College of Business
>
> The University of Tulsa
>
> ** **
>


Re: datanode failing to start.

2012-01-04 Thread Joey Krabacher
Have you checked your logs?

--Joey

On Wed, Jan 4, 2012 at 4:37 PM, Dave Kelsey  wrote:
>
> java version 1.6.0_29
> hadoop: 0.20.203.0
>
> I'm attempting to setup the pseudo-distributed config on a mac 10.6.8.
> I followed the steps from the QuickStart
> (http://wiki.apache.org./hadoop/QuickStart) and succeeded with Stage 1:
> Standalone Operation.
> I followed the steps for Stage 2: Pseudo-distributed Configuration.
> I set the JAVA_HOME variable in conf/hadoop-env.sh and I changed tools.jar
> to the location of classes.jar (a mac version of tools.jar)
> I've modified the three .xml files as described in the QuickStart.
> ssh'ing to localhost has been configured and works with passwordless
> authentication.
> I formatted the namenode with "bin/hadoop namenode -format" as the
> instructions say
>
> This is what I see when I run bin/start-all.sh
>
> root# bin/start-all.sh
> starting namenode, logging to
> /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-namenode-Hoot-2.local.out
> localhost: starting datanode, logging to
> /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-datanode-Hoot-2.local.out
> localhost: Exception in thread "main" java.lang.NoClassDefFoundError: server
> localhost: Caused by: java.lang.ClassNotFoundException: server
> localhost:     at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
> localhost:     at java.security.AccessController.doPrivileged(Native Method)
> localhost:     at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
> localhost:     at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
> localhost:     at
> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
> localhost:     at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
> localhost: starting secondarynamenode, logging to
> /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-secondarynamenode-Hoot-2.local.out
> starting jobtracker, logging to
> /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-jobtracker-Hoot-2.local.out
> localhost: starting tasktracker, logging to
> /Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-tasktracker-Hoot-2.local.out
>
> There are 4 processes running:
> ps -fax | grep hadoop | grep -v grep | wc -l
>      4
>
> They are:
> SecondaryNameNode
> TaskTracker
> NameNode
> JobTracker
>
>
> I've searched to see if anyone else has encountered this and not found
> anything
>
> Dave Kelsey
>


datanode failing to start.

2012-01-04 Thread Dave Kelsey


java version 1.6.0_29
hadoop: 0.20.203.0

I'm attempting to setup the pseudo-distributed config on a mac 10.6.8.
I followed the steps from the QuickStart 
(http://wiki.apache.org./hadoop/QuickStart) and succeeded with Stage 1: 
Standalone Operation.

I followed the steps for Stage 2: Pseudo-distributed Configuration.
I set the JAVA_HOME variable in conf/hadoop-env.sh and I changed 
tools.jar to the location of classes.jar (a mac version of tools.jar)

I've modified the three .xml files as described in the QuickStart.
ssh'ing to localhost has been configured and works with passwordless 
authentication.
I formatted the namenode with "bin/hadoop namenode -format" as the 
instructions say


This is what I see when I run bin/start-all.sh

root# bin/start-all.sh
starting namenode, logging to 
/Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-namenode-Hoot-2.local.out
localhost: starting datanode, logging to 
/Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-datanode-Hoot-2.local.out

localhost: Exception in thread "main" java.lang.NoClassDefFoundError: server
localhost: Caused by: java.lang.ClassNotFoundException: server
localhost: at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
localhost: at java.security.AccessController.doPrivileged(Native Method)
localhost: at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
localhost: at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
localhost: at 
sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)

localhost: at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
localhost: starting secondarynamenode, logging to 
/Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-secondarynamenode-Hoot-2.local.out
starting jobtracker, logging to 
/Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-jobtracker-Hoot-2.local.out
localhost: starting tasktracker, logging to 
/Users/admin/hadoop/hadoop-0.20.203.0/bin/../logs/hadoop-root-tasktracker-Hoot-2.local.out


There are 4 processes running:
ps -fax | grep hadoop | grep -v grep | wc -l
  4

They are:
SecondaryNameNode
TaskTracker
NameNode
JobTracker


I've searched to see if anyone else has encountered this and not found 
anything


Dave Kelsey



Setting HDFS directory time programmatically

2012-01-04 Thread Frank Astier
Hi -

Is it possible to set the access time of a HDFS directory programmatically?

I’m using 0.20.204.0.

I need to do that in unit tests, where my clean up program is going to remove 
files/dirs whose access time is too far in the past. I can setTimes on the test 
files without any problem, but not on the directories... The directories 
created automatically when I create the test fiels have a date (with 
getAccessTime) of 1969/12/31 16:00 and I can’t control that date, which makes 
my unit testing impossible.

By the way, setTimes doesn’t allow to set the date on dirs, but getAccessTime 
is happy to return a date, which is inconsistent, IMHO.

Finally, on our production systems, I’m seeing appropriate dates for both files 
and directories.

Any insight appreciated,

Thanks!

Frank


Balancer exiting immediately despite having work to do.

2012-01-04 Thread Bible, Landy
Hi all,

I'm running Hadoop 0.20.2.  The balancer has suddenly stopped working.  I'm 
attempting to balance the cluster with a threshold of 1, using the following 
command:

./hadoop balancer -threshold 1

This has been working fine, but suddenly it isn't.  It skips though 5 
iterations without actually doing any work:

Time Stamp   Iteration#  Bytes Already Moved  Bytes Left To Move  
Bytes Being Moved
Jan 4, 2012 11:47:56 AM   0 0 KB 1.87 GB
6.68 GB
Jan 4, 2012 11:47:56 AM   1 0 KB 1.87 GB
6.68 GB
Jan 4, 2012 11:47:56 AM   2 0 KB 1.87 GB
6.68 GB
Jan 4, 2012 11:47:57 AM   3 0 KB 1.87 GB
6.68 GB
Jan 4, 2012 11:47:57 AM   4 0 KB 1.87 GB
6.68 GB
No block has been moved for 5 iterations. Exiting...
Balancing took 524.0 milliseconds

I've attached the full log, but I can't see any errors indicating why it is 
failing.  Any ideas?  I'd really like to get balancing working again.  My use 
case isn't the norm, and it is important that the cluster stay as close to 
completely balanced as possible.

--
Landy Bible

Simulation and Computer Specialist
School of Nursing - Collins College of Business
The University of Tulsa



Re: How do i customize the name of the job during submission?

2012-01-04 Thread Harsh J
Arun,

You can't change the job IDs. Why do you seek to do this though?

You can, however, specify a custom job 'name' via the regular JobConf/Job APIs.

On 04-Jan-2012, at 9:31 PM, ArunKumar wrote:

> Hi guys !
> 
> When a Job is submitted it is given an ID say job_200904211745_0018 in
> Hadoop.
> 
> But for some reason i want to submit it with ID say "job1". How can i do
> that ?
> 
> Arun
> 
> --
> View this message in context: 
> http://lucene.472066.n3.nabble.com/How-do-i-customize-the-name-of-the-job-during-submission-tp3632228p3632228.html
> Sent from the Hadoop lucene-users mailing list archive at Nabble.com.



Task location determination

2012-01-04 Thread Bai Shen
I have a test Hadoop cluster set up using Cloudera.  It consists of the
Name Node and three Data Nodes.  When I submit jobs, they end up piling up
on one node instead of round robining through the different nodes.

I understand that Hadoop tries to run the job where the data is located,
but with only three data nodes and a replication factor of 3, wouldn't that
mean that the same data is on every single machine?  Why would it not
spread the tasking out over all of the machines instead of clumping up on
one, leaving the others idle?

Thanks.