I was under the understanding that with EMR 4.x, we had to create separate 
config files and place them on S3, then reference them when we create the 
cluster as described in detail in this first article:

http://docs.aws.amazon.com/ElasticMapReduce/latest/ReleaseGuide/emr-release-differences.html

This next article goes into more detail.

http://docs.aws.amazon.com/ElasticMapReduce/latest/ReleaseGuide/emr-configure-apps.html

But, this is just based on reading the documentation, I have been unable to get 
the custom configs working yet.

Thanks,

Jim


From: Aniruddha Thombare [mailto:anirud...@datatorrent.com]
Sent: Friday, March 11, 2016 9:16 AM
To: users@apex.incubator.apache.org
Subject: Re: AWS EMR: Container is running beyond virtual memory limits


Hi,

It seems that the above mentioned configurations didn't take effect.
Those changes were made in:
/etc/Hadoop/conf/yarn-site.xml
And
malted-site.xml

@Sandeep even pidemo didn't run.

On Fri, 11 Mar 2016 8:34 pm Pradeep A. Dalvi, 
<p...@apache.org<mailto:p...@apache.org>> wrote:
We are facing following error message while starting any containers on AWS EMR.


Container [pid=8107,containerID=container_1457702160744_0001_01_000007] is 
running beyond virtual memory limits. Current usage: 186.1 MB of 256 MB 
physical memory used; 2.0 GB of 1.3 GB virtual memory used. Killing container.

Dump of the process-tree for container_1457702160744_0001_01_000007 :

     |- PID PPID PGRPID SESSID CMD_NAME USER_MODE_TIME(MILLIS) 
SYSTEM_TIME(MILLIS) VMEM_USAGE(BYTES) RSSMEM_USAGE(PAGES) FULL_CMD_LINE

     |- 8222 8107 8107 8107 (java) 589 62 2041503744 46944 
/usr/lib/jvm/java-openjdk/bin/java -Xmx234881024 
-Ddt.attr.APPLICATION_PATH=hdfs://ip-172-31-9-174.ec2.internal:8020/user/hadoop/datatorrent/apps/application_1457702160744_0001
 
-Djava.io.tmpdir=/mnt1/yarn/usercache/hadoop/appcache/application_1457702160744_0001/container_1457702160744_0001_01_000007/tmp
 -Ddt.cid=container_1457702160744_0001_01_000007 -Dhadoop.root.logger=INFO,RFA 
-Dhadoop.log.dir=/var/log/hadoop-yarn/containers/application_1457702160744_0001/container_1457702160744_0001_01_000007
 com.datatorrent.stram.engine.StreamingContainer

     |- 8107 8105 8107 8107 (bash) 1 5 115806208 705 /bin/bash -c 
/usr/lib/jvm/java-openjdk/bin/java  -Xmx234881024  
-Ddt.attr.APPLICATION_PATH=hdfs://ip-172-31-9-174.ec2.internal:8020/user/hadoop/datatorrent/apps/application_1457702160744_0001
 
-Djava.io.tmpdir=/mnt1/yarn/usercache/hadoop/appcache/application_1457702160744_0001/container_1457702160744_0001_01_000007/tmp
 -Ddt.cid=container_1457702160744_0001_01_000007 -Dhadoop.root.logger=INFO,RFA 
-Dhadoop.log.dir=/var/log/hadoop-yarn/containers/application_1457702160744_0001/container_1457702160744_0001_01_000007
 com.datatorrent.stram.engine.StreamingContainer 
1>/var/log/hadoop-yarn/containers/application_1457702160744_0001/container_1457702160744_0001_01_000007/stdout
 
2>/var/log/hadoop-yarn/containers/application_1457702160744_0001/container_1457702160744_0001_01_000007/stderr



Container killed on request. Exit code is 143

Container exited with a non-zero exit code 143

We had 1 m3.xlarge MASTER & 2 m3.xlarge CORE instances provisioned. We also 
have tried m4.4xlarge instances. EMR Task configurations can be found at 
http://docs.aws.amazon.com/ElasticMapReduce/latest/DeveloperGuide/TaskConfiguration_H2.html

We tried changing following yarn configurations, however they did not seem to 
help much.

  
<property><name>yarn.nodemanager.resource.memory-mb</name><value>12288</value></property>
  
<property><name>yarn.scheduler.maximum-allocation-mb</name><value>4096</value></property>
  <property> <name>yarn.nodemanager.vmem-check-enabled</name> 
<value>false</value> </property>
  <property> <name>yarn.nodemanager.vmem-pmem-ratio</name> 
<value>50</value></property>


Thanks,
--
Pradeep A. Dalvi

Reply via email to