spark git commit: [SPARK-16095][YARN] Yarn cluster mode should report correct state to SparkLauncher

2016-07-01 Thread vanzin
Repository: spark Updated Branches: refs/heads/master d17e5f2f1 -> bad0f7dbb [SPARK-16095][YARN] Yarn cluster mode should report correct state to SparkLauncher ## What changes were proposed in this pull request? Yarn cluster mode should return correct state for SparkLauncher ## How was this

spark git commit: [SPARK-16095][YARN] Yarn cluster mode should report correct state to SparkLauncher

2016-07-01 Thread vanzin
Repository: spark Updated Branches: refs/heads/branch-2.0 794d09969 -> ab4303800 [SPARK-16095][YARN] Yarn cluster mode should report correct state to SparkLauncher ## What changes were proposed in this pull request? Yarn cluster mode should return correct state for SparkLauncher ## How was

spark git commit: [SPARK-16233][R][TEST] ORC test should be enabled only when HiveContext is available.

2016-07-01 Thread shivaram
Repository: spark Updated Branches: refs/heads/branch-2.0 78387ce05 -> 794d09969 [SPARK-16233][R][TEST] ORC test should be enabled only when HiveContext is available. ## What changes were proposed in this pull request? ORC test should be enabled only when HiveContext is available. ## How

spark git commit: [SPARK-16233][R][TEST] ORC test should be enabled only when HiveContext is available.

2016-07-01 Thread shivaram
Repository: spark Updated Branches: refs/heads/master d601894c0 -> d17e5f2f1 [SPARK-16233][R][TEST] ORC test should be enabled only when HiveContext is available. ## What changes were proposed in this pull request? ORC test should be enabled only when HiveContext is available. ## How was

spark git commit: [SPARK-16335][SQL] Structured streaming should fail if source directory does not exist

2016-07-01 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-2.0 d6588115a -> 78387ce05 [SPARK-16335][SQL] Structured streaming should fail if source directory does not exist ## What changes were proposed in this pull request? In structured streaming, Spark does not report errors when the specified

spark git commit: [SPARK-16335][SQL] Structured streaming should fail if source directory does not exist

2016-07-01 Thread rxin
Repository: spark Updated Branches: refs/heads/master e4fa58c43 -> d601894c0 [SPARK-16335][SQL] Structured streaming should fail if source directory does not exist ## What changes were proposed in this pull request? In structured streaming, Spark does not report errors when the specified

spark git commit: [SPARK-16299][SPARKR] Capture errors from R workers in daemon.R to avoid deletion of R session temporary directory.

2016-07-01 Thread shivaram
Repository: spark Updated Branches: refs/heads/branch-2.0 4c96ded84 -> d6588115a [SPARK-16299][SPARKR] Capture errors from R workers in daemon.R to avoid deletion of R session temporary directory. ## What changes were proposed in this pull request? Capture errors from R workers in daemon.R

spark git commit: [SPARK-16299][SPARKR] Capture errors from R workers in daemon.R to avoid deletion of R session temporary directory.

2016-07-01 Thread shivaram
Repository: spark Updated Branches: refs/heads/master 26afb4ce4 -> e4fa58c43 [SPARK-16299][SPARKR] Capture errors from R workers in daemon.R to avoid deletion of R session temporary directory. ## What changes were proposed in this pull request? Capture errors from R workers in daemon.R to

spark git commit: [SPARK-16012][SPARKR] Implement gapplyCollect which will apply a R function on each group similar to gapply and collect the result back to R data.frame

2016-07-01 Thread shivaram
Repository: spark Updated Branches: refs/heads/branch-2.0 3665927c6 -> 4c96ded84 [SPARK-16012][SPARKR] Implement gapplyCollect which will apply a R function on each group similar to gapply and collect the result back to R data.frame ## What changes were proposed in this pull request?

spark git commit: [SPARK-16012][SPARKR] Implement gapplyCollect which will apply a R function on each group similar to gapply and collect the result back to R data.frame

2016-07-01 Thread shivaram
Repository: spark Updated Branches: refs/heads/master c55397652 -> 26afb4ce4 [SPARK-16012][SPARKR] Implement gapplyCollect which will apply a R function on each group similar to gapply and collect the result back to R data.frame ## What changes were proposed in this pull request?

spark git commit: [SPARK-16208][SQL] Add `PropagateEmptyRelation` optimizer

2016-07-01 Thread lian
Repository: spark Updated Branches: refs/heads/master 0ad6ce7e5 -> c55397652 [SPARK-16208][SQL] Add `PropagateEmptyRelation` optimizer ## What changes were proposed in this pull request? This PR adds a new logical optimizer, `PropagateEmptyRelation`, to collapse a logical plans consisting

spark git commit: [SPARK-16222][SQL] JDBC Sources - Handling illegal input values for `fetchsize` and `batchsize`

2016-07-01 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-2.0 0b64543c5 -> 3665927c6 [SPARK-16222][SQL] JDBC Sources - Handling illegal input values for `fetchsize` and `batchsize` What changes were proposed in this pull request? For JDBC data sources, users can specify `batchsize` for

spark git commit: [SPARK-16222][SQL] JDBC Sources - Handling illegal input values for `fetchsize` and `batchsize`

2016-07-01 Thread srowen
Repository: spark Updated Branches: refs/heads/master 66283ee0b -> 0ad6ce7e5 [SPARK-16222][SQL] JDBC Sources - Handling illegal input values for `fetchsize` and `batchsize` What changes were proposed in this pull request? For JDBC data sources, users can specify `batchsize` for

spark git commit: [SPARK-15761][MLLIB][PYSPARK] Load ipython when default python is Python3

2016-07-01 Thread srowen
Repository: spark Updated Branches: refs/heads/master 2075bf8ef -> 66283ee0b [SPARK-15761][MLLIB][PYSPARK] Load ipython when default python is Python3 ## What changes were proposed in this pull request? I would like to use IPython with Python 3.5. It is annoying when it fails with IPython

spark git commit: [SPARK-15761][MLLIB][PYSPARK] Load ipython when default python is Python3

2016-07-01 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-2.0 972106dd3 -> 0b64543c5 [SPARK-15761][MLLIB][PYSPARK] Load ipython when default python is Python3 ## What changes were proposed in this pull request? I would like to use IPython with Python 3.5. It is annoying when it fails with

spark git commit: [SPARK-15761][MLLIB][PYSPARK] Load ipython when default python is Python3

2016-07-01 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.6 83f860448 -> 1026aba16 [SPARK-15761][MLLIB][PYSPARK] Load ipython when default python is Python3 ## What changes were proposed in this pull request? I would like to use IPython with Python 3.5. It is annoying when it fails with

spark git commit: [SPARK-16182][CORE] Utils.scala -- terminateProcess() should call Process.destroyForcibly() if and only if Process.destroy() fails

2016-07-01 Thread srowen
Repository: spark Updated Branches: refs/heads/master fbfd0ab9d -> 2075bf8ef [SPARK-16182][CORE] Utils.scala -- terminateProcess() should call Process.destroyForcibly() if and only if Process.destroy() fails ## What changes were proposed in this pull request? Utils.terminateProcess should

spark git commit: [SPARK-12177][STREAMING][KAFKA] limit api surface area

2016-07-01 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-2.0 cc3c44b11 -> 1932bb683 [SPARK-12177][STREAMING][KAFKA] limit api surface area ## What changes were proposed in this pull request? This is an alternative to the refactoring proposed by https://github.com/apache/spark/pull/13996 ## How

spark git commit: [SPARK-12177][STREAMING][KAFKA] limit api surface area

2016-07-01 Thread tdas
Repository: spark Updated Branches: refs/heads/master 14cf61e90 -> fbfd0ab9d [SPARK-12177][STREAMING][KAFKA] limit api surface area ## What changes were proposed in this pull request? This is an alternative to the refactoring proposed by https://github.com/apache/spark/pull/13996 ## How was