Changing the mapred.child.java.opts value does not change the heap size from a 
default one.
-------------------------------------------------------------------------------------------

                 Key: WHIRR-146
                 URL: https://issues.apache.org/jira/browse/WHIRR-146
             Project: Whirr
          Issue Type: Bug
         Environment: Amazon EC2, Amazon Linux images.
            Reporter: Tibor Kiss
            Assignee: Tibor Kiss


Even if I change the value for mapred.child.java.opts the task is started with 
-Xmx200m.
Since the mapred.child.java.opts and mapred.child.ulimit has been deprecated, 
we need to set the mapred.map.child.java.opts, mapred.reduce.child.java.opts 
respectively the mapred.map.child.ulimit and mapred.reduce.child.ulimit in 
order to have any effect.
Unfortunately the /scripts/cdh/install and /scripts/apache/install which 
generates the /etc/hadoop/conf.dist/hadoop-site.xml is not synchronized with 
this deprecation as a result we are not able to use mappers and reducers which 
does not fit in 200M heap size.

How to reproduce: 
1. Start a cluster on large instances where we are using 64bit jvm and run a 
simple distcp, you will experience Child jvm crash.
2. Or run a job with mappers or reducers which does not fit in 200M heap, it 
will experience OutOfMemoryError in child processes.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to