spark git commit: Revert "[SPARK-15782][YARN] Set spark.jars system property in client mode"

2016-06-15 Thread davies
Repository: spark
Updated Branches:
  refs/heads/branch-2.0 73bf87f3c -> 9acf254ed


Revert "[SPARK-15782][YARN] Set spark.jars system property in client mode"

This reverts commit f1e9d2d92456b437803d6ebab77831e66b0eb53e.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/9acf254e
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/9acf254e
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/9acf254e

Branch: refs/heads/branch-2.0
Commit: 9acf254edadcbecb0f0bee00578dc7f228b5ec1e
Parents: 73bf87f
Author: Davies Liu 
Authored: Wed Jun 15 15:54:29 2016 -0700
Committer: Davies Liu 
Committed: Wed Jun 15 15:54:29 2016 -0700

--
 .../scala/org/apache/spark/SparkContext.scala   |  2 +-
 .../scala/org/apache/spark/util/Utils.scala | 20 
 .../apache/spark/deploy/SparkSubmitSuite.scala  | 12 
 .../org/apache/spark/repl/SparkILoop.scala  | 11 ++-
 .../main/scala/org/apache/spark/repl/Main.scala |  4 +++-
 5 files changed, 6 insertions(+), 43 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/9acf254e/core/src/main/scala/org/apache/spark/SparkContext.scala
--
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala 
b/core/src/main/scala/org/apache/spark/SparkContext.scala
index d870181..d56946e 100644
--- a/core/src/main/scala/org/apache/spark/SparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/SparkContext.scala
@@ -391,7 +391,7 @@ class SparkContext(config: SparkConf) extends Logging with 
ExecutorAllocationCli
 
 _conf.set("spark.executor.id", SparkContext.DRIVER_IDENTIFIER)
 
-_jars = Utils.getUserJars(_conf)
+_jars = 
_conf.getOption("spark.jars").map(_.split(",")).map(_.filter(_.nonEmpty)).toSeq.flatten
 _files = 
_conf.getOption("spark.files").map(_.split(",")).map(_.filter(_.nonEmpty))
   .toSeq.flatten
 

http://git-wip-us.apache.org/repos/asf/spark/blob/9acf254e/core/src/main/scala/org/apache/spark/util/Utils.scala
--
diff --git a/core/src/main/scala/org/apache/spark/util/Utils.scala 
b/core/src/main/scala/org/apache/spark/util/Utils.scala
index aebd98b..f9d0540 100644
--- a/core/src/main/scala/org/apache/spark/util/Utils.scala
+++ b/core/src/main/scala/org/apache/spark/util/Utils.scala
@@ -2352,26 +2352,6 @@ private[spark] object Utils extends Logging {
 log.info(s"Started daemon with process name: ${Utils.getProcessName()}")
 SignalUtils.registerLogger(log)
   }
-
-  /**
-   * Unions two comma-separated lists of files and filters out empty strings.
-   */
-  def unionFileLists(leftList: Option[String], rightList: Option[String]): 
Set[String] = {
-var allFiles = Set[String]()
-leftList.foreach { value => allFiles ++= value.split(",") }
-rightList.foreach { value => allFiles ++= value.split(",") }
-allFiles.filter { _.nonEmpty }
-  }
-
-  def getUserJars(conf: SparkConf): Seq[String] = {
-val sparkJars = conf.getOption("spark.jars")
-if (conf.get("spark.master") == "yarn") {
-  val yarnJars = conf.getOption("spark.yarn.dist.jars")
-  unionFileLists(sparkJars, yarnJars).toSeq
-} else {
-  sparkJars.map(_.split(",")).map(_.filter(_.nonEmpty)).toSeq.flatten
-}
-  }
 }
 
 /**

http://git-wip-us.apache.org/repos/asf/spark/blob/9acf254e/core/src/test/scala/org/apache/spark/deploy/SparkSubmitSuite.scala
--
diff --git a/core/src/test/scala/org/apache/spark/deploy/SparkSubmitSuite.scala 
b/core/src/test/scala/org/apache/spark/deploy/SparkSubmitSuite.scala
index 0b02059..2718976 100644
--- a/core/src/test/scala/org/apache/spark/deploy/SparkSubmitSuite.scala
+++ b/core/src/test/scala/org/apache/spark/deploy/SparkSubmitSuite.scala
@@ -570,18 +570,6 @@ class SparkSubmitSuite
   appArgs.executorMemory should be ("2.3g")
 }
   }
-
-  test("comma separated list of files are unioned correctly") {
-val left = Option("/tmp/a.jar,/tmp/b.jar")
-val right = Option("/tmp/c.jar,/tmp/a.jar")
-val emptyString = Option("")
-Utils.unionFileLists(left, right) should be (Set("/tmp/a.jar", 
"/tmp/b.jar", "/tmp/c.jar"))
-Utils.unionFileLists(emptyString, emptyString) should be (Set.empty)
-Utils.unionFileLists(Option("/tmp/a.jar"), emptyString) should be 
(Set("/tmp/a.jar"))
-Utils.unionFileLists(emptyString, Option("/tmp/a.jar")) should be 
(Set("/tmp/a.jar"))
-Utils.unionFileLists(None, Option("/tmp/a.jar")) should be 
(Set("/tmp/a.jar"))
-Utils.unionFileLists(Option("/tmp/a.jar"), None) should be 
(Set("/tmp/a.jar"))
-  }
   // scalastyle:on println
 
   // NOTE: This is an expensive operation in terms of time (10 seconds+).

spark git commit: Revert "[SPARK-15782][YARN] Set spark.jars system property in client mode"

2016-06-15 Thread davies
Repository: spark
Updated Branches:
  refs/heads/master 1a33f2e05 -> a153e41c0


Revert "[SPARK-15782][YARN] Set spark.jars system property in client mode"

This reverts commit 4df8df5c2e68f5a5d231c401b04d762d7a648159.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/a153e41c
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/a153e41c
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/a153e41c

Branch: refs/heads/master
Commit: a153e41c08b635b1b2a48b680d203c148deebc67
Parents: 1a33f2e
Author: Davies Liu 
Authored: Wed Jun 15 15:55:07 2016 -0700
Committer: Davies Liu 
Committed: Wed Jun 15 15:55:07 2016 -0700

--
 .../scala/org/apache/spark/SparkContext.scala   |  2 +-
 .../scala/org/apache/spark/util/Utils.scala | 20 
 .../apache/spark/deploy/SparkSubmitSuite.scala  | 12 
 .../org/apache/spark/repl/SparkILoop.scala  | 11 ++-
 .../main/scala/org/apache/spark/repl/Main.scala |  4 +++-
 5 files changed, 6 insertions(+), 43 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/a153e41c/core/src/main/scala/org/apache/spark/SparkContext.scala
--
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala 
b/core/src/main/scala/org/apache/spark/SparkContext.scala
index d870181..d56946e 100644
--- a/core/src/main/scala/org/apache/spark/SparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/SparkContext.scala
@@ -391,7 +391,7 @@ class SparkContext(config: SparkConf) extends Logging with 
ExecutorAllocationCli
 
 _conf.set("spark.executor.id", SparkContext.DRIVER_IDENTIFIER)
 
-_jars = Utils.getUserJars(_conf)
+_jars = 
_conf.getOption("spark.jars").map(_.split(",")).map(_.filter(_.nonEmpty)).toSeq.flatten
 _files = 
_conf.getOption("spark.files").map(_.split(",")).map(_.filter(_.nonEmpty))
   .toSeq.flatten
 

http://git-wip-us.apache.org/repos/asf/spark/blob/a153e41c/core/src/main/scala/org/apache/spark/util/Utils.scala
--
diff --git a/core/src/main/scala/org/apache/spark/util/Utils.scala 
b/core/src/main/scala/org/apache/spark/util/Utils.scala
index aebd98b..f9d0540 100644
--- a/core/src/main/scala/org/apache/spark/util/Utils.scala
+++ b/core/src/main/scala/org/apache/spark/util/Utils.scala
@@ -2352,26 +2352,6 @@ private[spark] object Utils extends Logging {
 log.info(s"Started daemon with process name: ${Utils.getProcessName()}")
 SignalUtils.registerLogger(log)
   }
-
-  /**
-   * Unions two comma-separated lists of files and filters out empty strings.
-   */
-  def unionFileLists(leftList: Option[String], rightList: Option[String]): 
Set[String] = {
-var allFiles = Set[String]()
-leftList.foreach { value => allFiles ++= value.split(",") }
-rightList.foreach { value => allFiles ++= value.split(",") }
-allFiles.filter { _.nonEmpty }
-  }
-
-  def getUserJars(conf: SparkConf): Seq[String] = {
-val sparkJars = conf.getOption("spark.jars")
-if (conf.get("spark.master") == "yarn") {
-  val yarnJars = conf.getOption("spark.yarn.dist.jars")
-  unionFileLists(sparkJars, yarnJars).toSeq
-} else {
-  sparkJars.map(_.split(",")).map(_.filter(_.nonEmpty)).toSeq.flatten
-}
-  }
 }
 
 /**

http://git-wip-us.apache.org/repos/asf/spark/blob/a153e41c/core/src/test/scala/org/apache/spark/deploy/SparkSubmitSuite.scala
--
diff --git a/core/src/test/scala/org/apache/spark/deploy/SparkSubmitSuite.scala 
b/core/src/test/scala/org/apache/spark/deploy/SparkSubmitSuite.scala
index 0b02059..2718976 100644
--- a/core/src/test/scala/org/apache/spark/deploy/SparkSubmitSuite.scala
+++ b/core/src/test/scala/org/apache/spark/deploy/SparkSubmitSuite.scala
@@ -570,18 +570,6 @@ class SparkSubmitSuite
   appArgs.executorMemory should be ("2.3g")
 }
   }
-
-  test("comma separated list of files are unioned correctly") {
-val left = Option("/tmp/a.jar,/tmp/b.jar")
-val right = Option("/tmp/c.jar,/tmp/a.jar")
-val emptyString = Option("")
-Utils.unionFileLists(left, right) should be (Set("/tmp/a.jar", 
"/tmp/b.jar", "/tmp/c.jar"))
-Utils.unionFileLists(emptyString, emptyString) should be (Set.empty)
-Utils.unionFileLists(Option("/tmp/a.jar"), emptyString) should be 
(Set("/tmp/a.jar"))
-Utils.unionFileLists(emptyString, Option("/tmp/a.jar")) should be 
(Set("/tmp/a.jar"))
-Utils.unionFileLists(None, Option("/tmp/a.jar")) should be 
(Set("/tmp/a.jar"))
-Utils.unionFileLists(Option("/tmp/a.jar"), None) should be 
(Set("/tmp/a.jar"))
-  }
   // scalastyle:on println
 
   // NOTE: This is an expensive operation in terms of time (10 seconds+). Use 
sp