Daisuke Kobayashi created SPARK-18765: -----------------------------------------
Summary: Make values for spark.yarn.{am|driver|executor}.memoryOverhead have configurable units Key: SPARK-18765 URL: https://issues.apache.org/jira/browse/SPARK-18765 Project: Spark Issue Type: Improvement Components: Spark Core Affects Versions: 1.6.3 Reporter: Daisuke Kobayashi Priority: Trivial {{spark.yarn.\{driver|executor|am\}.memoryOverhead}} values are isolated to Megabytes today. Users provide a value without a unit and Spark assumes its in MBs. Since the overhead is often a few gigabytes, we should change the memory overhead to work the same way as executor or driver memory configs. Given 2.0 has already covered this, it's worth to have 1.X code line cover this capability as well. My PR offers users being able to pass the values in multiple ways (backward compatibility is not broken) like: {code} spark.yarn.executor.memoryOverhead=300m --> converted to 300 spark.yarn.executor.memoryOverhead=500 --> converted to 500 spark.yarn.executor.memoryOverhead=1g --> converted to 1024 spark.yarn.executor.memoryOverhead=1024m --> converted to 1024 {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org