[ https://issues.apache.org/jira/browse/YARN-2041?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14000601#comment-14000601 ]
Karthik Kambatla commented on YARN-2041: ---------------------------------------- yarn.nodemanager.resource.memory-mb should ideally be fixed per node in a YARN cluster. As [~tgaves] said, we should look at how the individual tasks are scheduled (spread out) and other relevant information. > Hard to co-locate MR2 and Spark jobs on the same cluster in YARN > ---------------------------------------------------------------- > > Key: YARN-2041 > URL: https://issues.apache.org/jira/browse/YARN-2041 > Project: Hadoop YARN > Issue Type: Improvement > Components: nodemanager > Affects Versions: 2.3.0 > Reporter: Nishkam Ravi > > Performance of MR2 jobs falls drastically as YARN config parameter > yarn.nodemanager.resource.memory-mb is increased beyond a certain value. > Performance of Spark falls drastically as the value of > yarn.nodemanager.resource.memory-mb is decreased beyond a certain value for a > large data set. > This makes it hard to co-locate MR2 and Spark jobs in YARN. > The experiments are being conducted on a 6-node cluster. The following > workloads are being run: TeraGen, TeraSort, TeraValidate, WordCount, > ShuffleText and PageRank. > Will add more details to this JIRA over time. -- This message was sent by Atlassian JIRA (v6.2#6252)