Fix Executor not getting properties in local mode
Project: http://git-wip-us.apache.org/repos/asf/incubator-spark/repo Commit: http://git-wip-us.apache.org/repos/asf/incubator-spark/commit/5bbe7386 Tree: http://git-wip-us.apache.org/repos/asf/incubator-spark/tree/5bbe7386 Diff: http://git-wip-us.apache.org/repos/asf/incubator-spark/diff/5bbe7386 Branch: refs/heads/master Commit: 5bbe73864eea78b76448ce42a7af847dad73b269 Parents: a16c52e Author: Matei Zaharia <ma...@databricks.com> Authored: Sat Dec 28 17:31:58 2013 -0500 Committer: Matei Zaharia <ma...@databricks.com> Committed: Sat Dec 28 17:31:58 2013 -0500 ---------------------------------------------------------------------- core/src/main/scala/org/apache/spark/SparkConf.scala | 2 +- .../scala/org/apache/spark/scheduler/local/LocalScheduler.scala | 3 ++- 2 files changed, 3 insertions(+), 2 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/5bbe7386/core/src/main/scala/org/apache/spark/SparkConf.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/SparkConf.scala b/core/src/main/scala/org/apache/spark/SparkConf.scala index 185ddb1..bd24cd1 100644 --- a/core/src/main/scala/org/apache/spark/SparkConf.scala +++ b/core/src/main/scala/org/apache/spark/SparkConf.scala @@ -28,7 +28,7 @@ class SparkConf(loadDefaults: Boolean) extends Serializable with Cloneable { if (loadDefaults) { val typesafeConfig = ConfigFactory.systemProperties() .withFallback(ConfigFactory.parseResources("spark.conf")) - for (e <- typesafeConfig.entrySet().asScala) { + for (e <- typesafeConfig.entrySet().asScala if e.getKey.startsWith("spark.")) { settings(e.getKey) = e.getValue.unwrapped.toString } } http://git-wip-us.apache.org/repos/asf/incubator-spark/blob/5bbe7386/core/src/main/scala/org/apache/spark/scheduler/local/LocalScheduler.scala ---------------------------------------------------------------------- diff --git a/core/src/main/scala/org/apache/spark/scheduler/local/LocalScheduler.scala b/core/src/main/scala/org/apache/spark/scheduler/local/LocalScheduler.scala index 8498cff..7c173e3 100644 --- a/core/src/main/scala/org/apache/spark/scheduler/local/LocalScheduler.scala +++ b/core/src/main/scala/org/apache/spark/scheduler/local/LocalScheduler.scala @@ -50,7 +50,8 @@ private[spark] class LocalActor(localScheduler: LocalScheduler, private var freeCores: Int) extends Actor with Logging { - val executor = new Executor("localhost", "localhost", Seq.empty, isLocal = true) + val executor = new Executor( + "localhost", "localhost", localScheduler.sc.conf.getAll, isLocal = true) def receive = { case LocalReviveOffers =>