Re: Scheduling in YARN according to available resources

2015-02-21 Thread R Nair
Hi Tariq,

Glad to see that your issue is resolved, thank you. This re-affirms the
compatibility issue with openJDK. Thanks

Regards,
Ravi

On Sat, Feb 21, 2015 at 1:40 PM, tesm...@gmail.com tesm...@gmail.com
wrote:

 Dear Nair,

 Your tip in your first email saved my day. Tahnks once again. I am happy
 with Oracle JDK.

 Regards,
 Tariq

 On Sat, Feb 21, 2015 at 4:05 PM, R Nair ravishankar.n...@gmail.com
 wrote:

 one of it is in the forum, if you search in google you will get more. I
 am not saying it may not work, but you will have to select and apply some
 patches. One of my friends also had the same problem and with too much
 difficulty, he got this into work. So better avoid :)

 https://github.com/elasticsearch/elasticsearch-hadoop/issues/197

 Thanks and regards,
 Nair

 On Sat, Feb 21, 2015 at 8:20 AM, tesm...@gmail.com tesm...@gmail.com
 wrote:

 Thanks Nair.

 Managed installing Oracle JDK and it is working great. Thanks for the
 tip.

 Any idea why OpenJDK is crashing and Oracle JDK works?

 Regards,
 Tariq




 On Sat, Feb 21, 2015 at 7:14 AM, tesm...@gmail.com tesm...@gmail.com
 wrote:

 Thanks for your answer Nair,
 Is installing Oracle JDK on Ubuntu is that complicated as described in
 this link

 http://askubuntu.com/questions/56104/how-can-i-install-sun-oracles-proprietary-java-jdk-6-7-8-or-jre

 Is there an alternate?

 Regards


 On Sat, Feb 21, 2015 at 6:50 AM, R Nair ravishankar.n...@gmail.com
 wrote:

 I had an issue very similar, I changed and used Oracle JDK. There is
 nothing I see wrong with your configuration in my first look, thanks

 Regards,
 Nair

 On Sat, Feb 21, 2015 at 1:42 AM, tesm...@gmail.com tesm...@gmail.com
 wrote:

 I have 7 nodes in my Hadoop cluster [8GB RAM and 4VCPUs to each
 nodes], 1 Namenode + 6 datanodes.

 I followed the link from Hortonwroks [
 http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.0.6.0/bk_installing_manually_book/content/rpm-chap1-11.html
 ] and made calculation according to the hardware configruation on my
 nodes. Added the update mapred-site and yarn-site.xml files in my 
 question.
 Still my application is crashing with the same exection

 My mapreduce application has 34 input splits with a block size of
 128MB.

 **mapred-site.xml** has the  following properties:

 mapreduce.framework.name  = yarn
 mapred.child.java.opts= -Xmx2048m
 mapreduce.map.memory.mb   = 4096
 mapreduce.map.java.opts   = -Xmx2048m

 **yarn-site.xml** has the  following properties:

 yarn.resourcemanager.hostname= hadoop-master
 yarn.nodemanager.aux-services= mapreduce_shuffle
 yarn.nodemanager.resource.memory-mb  = 6144
 yarn.scheduler.minimum-allocation-mb = 2048
 yarn.scheduler.maximum-allocation-mb = 6144


  Exception from container-launch: ExitCodeException exitCode=134:
 /bin/bash: line 1:  3876 Aborted  (core dumped)
 /usr/lib/jvm/java-7-openjdk-amd64/bin/java
 -Djava.net.preferIPv4Stack=true
 -Dhadoop.metrics.log.level=WARN -Xmx8192m
 -Djava.io.tmpdir=/tmp/hadoop-ubuntu/nm-local-dir/usercache/ubuntu/appcache/application_1424264025191_0002/container_1424264025191_0002_01_11/tmp
 -Dlog4j.configuration=container-log4j.properties
 -Dyarn.app.container.log.dir=/home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11
 -Dyarn.app.container.log.filesize=0
 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild
 192.168.0.12 50842 attempt_1424264025191_0002_m_05_0 11 

 /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stdout
 2

 /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stderr


 How can avoid this?any help is appreciated

 It looks to me that YAN is trying to launch all the container
 simultaneously and anot according to the available resources. Is
 there an option to restrict number of containers on hadoop ndoes?

 Regards,
 Tariq




 --
 Warmest Regards,

 Ravi Shankar






 --
 Warmest Regards,

 Ravi Shankar





-- 
Warmest Regards,

Ravi Shankar


Scheduling in YARN according to available resources

2015-02-20 Thread tesm...@gmail.com
I have 7 nodes in my Hadoop cluster [8GB RAM and 4VCPUs to each nodes], 1
Namenode + 6 datanodes.

I followed the link from Hortonwroks [
http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.0.6.0/bk_installing_manually_book/content/rpm-chap1-11.html
] and made calculation according to the hardware configruation on my nodes.
Added the update mapred-site and yarn-site.xml files in my question. Still
my application is crashing with the same exection

My mapreduce application has 34 input splits with a block size of 128MB.

**mapred-site.xml** has the  following properties:

mapreduce.framework.name  = yarn
mapred.child.java.opts= -Xmx2048m
mapreduce.map.memory.mb   = 4096
mapreduce.map.java.opts   = -Xmx2048m

**yarn-site.xml** has the  following properties:

yarn.resourcemanager.hostname= hadoop-master
yarn.nodemanager.aux-services= mapreduce_shuffle
yarn.nodemanager.resource.memory-mb  = 6144
yarn.scheduler.minimum-allocation-mb = 2048
yarn.scheduler.maximum-allocation-mb = 6144


 Exception from container-launch: ExitCodeException exitCode=134:
/bin/bash: line 1:  3876 Aborted  (core dumped)
/usr/lib/jvm/java-7-openjdk-amd64/bin/java
-Djava.net.preferIPv4Stack=true
-Dhadoop.metrics.log.level=WARN -Xmx8192m
-Djava.io.tmpdir=/tmp/hadoop-ubuntu/nm-local-dir/usercache/ubuntu/appcache/application_1424264025191_0002/container_1424264025191_0002_01_11/tmp
-Dlog4j.configuration=container-log4j.properties
-Dyarn.app.container.log.dir=/home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11
-Dyarn.app.container.log.filesize=0
-Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild
192.168.0.12 50842 attempt_1424264025191_0002_m_05_0 11 

/home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stdout
2

/home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stderr


How can avoid this?any help is appreciated

It looks to me that YAN is trying to launch all the container
simultaneously and anot according to the available resources. Is there an
option to restrict number of containers on hadoop ndoes?

Regards,
Tariq


Re: Scheduling in YARN according to available resources

2015-02-20 Thread R Nair
I had an issue very similar, I changed and used Oracle JDK. There is
nothing I see wrong with your configuration in my first look, thanks

Regards,
Nair

On Sat, Feb 21, 2015 at 1:42 AM, tesm...@gmail.com tesm...@gmail.com
wrote:

 I have 7 nodes in my Hadoop cluster [8GB RAM and 4VCPUs to each nodes], 1
 Namenode + 6 datanodes.

 I followed the link from Hortonwroks [
 http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.0.6.0/bk_installing_manually_book/content/rpm-chap1-11.html
 ] and made calculation according to the hardware configruation on my
 nodes. Added the update mapred-site and yarn-site.xml files in my question.
 Still my application is crashing with the same exection

 My mapreduce application has 34 input splits with a block size of 128MB.

 **mapred-site.xml** has the  following properties:

 mapreduce.framework.name  = yarn
 mapred.child.java.opts= -Xmx2048m
 mapreduce.map.memory.mb   = 4096
 mapreduce.map.java.opts   = -Xmx2048m

 **yarn-site.xml** has the  following properties:

 yarn.resourcemanager.hostname= hadoop-master
 yarn.nodemanager.aux-services= mapreduce_shuffle
 yarn.nodemanager.resource.memory-mb  = 6144
 yarn.scheduler.minimum-allocation-mb = 2048
 yarn.scheduler.maximum-allocation-mb = 6144


  Exception from container-launch: ExitCodeException exitCode=134:
 /bin/bash: line 1:  3876 Aborted  (core dumped)
 /usr/lib/jvm/java-7-openjdk-amd64/bin/java
 -Djava.net.preferIPv4Stack=true
 -Dhadoop.metrics.log.level=WARN -Xmx8192m
 -Djava.io.tmpdir=/tmp/hadoop-ubuntu/nm-local-dir/usercache/ubuntu/appcache/application_1424264025191_0002/container_1424264025191_0002_01_11/tmp
 -Dlog4j.configuration=container-log4j.properties
 -Dyarn.app.container.log.dir=/home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11
 -Dyarn.app.container.log.filesize=0
 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild
 192.168.0.12 50842 attempt_1424264025191_0002_m_05_0 11 

 /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stdout
 2

 /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stderr


 How can avoid this?any help is appreciated

 It looks to me that YAN is trying to launch all the container
 simultaneously and anot according to the available resources. Is there an
 option to restrict number of containers on hadoop ndoes?

 Regards,
 Tariq




-- 
Warmest Regards,

Ravi Shankar


Re: Scheduling in YARN according to available resources

2015-02-20 Thread tesm...@gmail.com
Thanks for your answer Nair,
Is installing Oracle JDK on Ubuntu is that complicated as described in this
link
http://askubuntu.com/questions/56104/how-can-i-install-sun-oracles-proprietary-java-jdk-6-7-8-or-jre

Is there an alternate?

Regards


On Sat, Feb 21, 2015 at 6:50 AM, R Nair ravishankar.n...@gmail.com wrote:

 I had an issue very similar, I changed and used Oracle JDK. There is
 nothing I see wrong with your configuration in my first look, thanks

 Regards,
 Nair

 On Sat, Feb 21, 2015 at 1:42 AM, tesm...@gmail.com tesm...@gmail.com
 wrote:

 I have 7 nodes in my Hadoop cluster [8GB RAM and 4VCPUs to each nodes], 1
 Namenode + 6 datanodes.

 I followed the link from Hortonwroks [
 http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.0.6.0/bk_installing_manually_book/content/rpm-chap1-11.html
 ] and made calculation according to the hardware configruation on my
 nodes. Added the update mapred-site and yarn-site.xml files in my question.
 Still my application is crashing with the same exection

 My mapreduce application has 34 input splits with a block size of 128MB.

 **mapred-site.xml** has the  following properties:

 mapreduce.framework.name  = yarn
 mapred.child.java.opts= -Xmx2048m
 mapreduce.map.memory.mb   = 4096
 mapreduce.map.java.opts   = -Xmx2048m

 **yarn-site.xml** has the  following properties:

 yarn.resourcemanager.hostname= hadoop-master
 yarn.nodemanager.aux-services= mapreduce_shuffle
 yarn.nodemanager.resource.memory-mb  = 6144
 yarn.scheduler.minimum-allocation-mb = 2048
 yarn.scheduler.maximum-allocation-mb = 6144


  Exception from container-launch: ExitCodeException exitCode=134:
 /bin/bash: line 1:  3876 Aborted  (core dumped)
 /usr/lib/jvm/java-7-openjdk-amd64/bin/java
 -Djava.net.preferIPv4Stack=true
 -Dhadoop.metrics.log.level=WARN -Xmx8192m
 -Djava.io.tmpdir=/tmp/hadoop-ubuntu/nm-local-dir/usercache/ubuntu/appcache/application_1424264025191_0002/container_1424264025191_0002_01_11/tmp
 -Dlog4j.configuration=container-log4j.properties
 -Dyarn.app.container.log.dir=/home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11
 -Dyarn.app.container.log.filesize=0
 -Dhadoop.root.logger=INFO,CLA org.apache.hadoop.mapred.YarnChild
 192.168.0.12 50842 attempt_1424264025191_0002_m_05_0 11 

 /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stdout
 2

 /home/ubuntu/hadoop/logs/userlogs/application_1424264025191_0002/container_1424264025191_0002_01_11/stderr


 How can avoid this?any help is appreciated

 It looks to me that YAN is trying to launch all the container
 simultaneously and anot according to the available resources. Is there
 an option to restrict number of containers on hadoop ndoes?

 Regards,
 Tariq




 --
 Warmest Regards,

 Ravi Shankar