spark git commit: [SPARK-17339][CORE][BRANCH-2.0] Do not use path to get a filesystem in hadoopFile and newHadoopFile APIs

2016-09-07 Thread shivaram
Repository: spark Updated Branches: refs/heads/branch-2.0 067752ce0 -> 28377da38 [SPARK-17339][CORE][BRANCH-2.0] Do not use path to get a filesystem in hadoopFile and newHadoopFile APIs ## What changes were proposed in this pull request? This PR backports

spark git commit: [SPARK-16533][CORE] - backport driver deadlock fix to 2.0

2016-09-07 Thread vanzin
Repository: spark Updated Branches: refs/heads/branch-2.0 078ac0e63 -> 067752ce0 [SPARK-16533][CORE] - backport driver deadlock fix to 2.0 ## What changes were proposed in this pull request? Backport changes from #14710 and #14925 to 2.0 Author: Marcelo Vanzin Author:

spark git commit: [SPARK-17052][SQL] Remove Duplicate Test Cases auto_join from HiveCompatibilitySuite.scala

2016-09-07 Thread joshrosen
Repository: spark Updated Branches: refs/heads/master 649fa4bf1 -> b230fb92a [SPARK-17052][SQL] Remove Duplicate Test Cases auto_join from HiveCompatibilitySuite.scala ### What changes were proposed in this pull request? The original [JIRA

spark git commit: [SPARK-17370] Shuffle service files not invalidated when a slave is lost

2016-09-07 Thread joshrosen
Repository: spark Updated Branches: refs/heads/branch-2.0 e6caceb5e -> 078ac0e63 [SPARK-17370] Shuffle service files not invalidated when a slave is lost ## What changes were proposed in this pull request? DAGScheduler invalidates shuffle files when an executor loss event occurs, but not

spark git commit: [SPARK-17370] Shuffle service files not invalidated when a slave is lost

2016-09-07 Thread joshrosen
Repository: spark Updated Branches: refs/heads/master 76ad89e92 -> 649fa4bf1 [SPARK-17370] Shuffle service files not invalidated when a slave is lost ## What changes were proposed in this pull request? DAGScheduler invalidates shuffle files when an executor loss event occurs, but not when

spark git commit: [SPARK-17245][SQL][BRANCH-1.6] Do not rely on Hive's session state to retrieve HiveConf

2016-09-07 Thread wenchen
Repository: spark Updated Branches: refs/heads/branch-1.6 69fa9458e -> 047bc3f13 [SPARK-17245][SQL][BRANCH-1.6] Do not rely on Hive's session state to retrieve HiveConf ## What changes were proposed in this pull request? Right now, we rely on Hive's `SessionState.get()` to retrieve the

spark git commit: [MINOR][SQL] Fixing the typo in unit test

2016-09-07 Thread srowen
Repository: spark Updated Branches: refs/heads/master 6f4aeccf8 -> 76ad89e92 [MINOR][SQL] Fixing the typo in unit test ## What changes were proposed in this pull request? Fixing the typo in the unit test of CodeGenerationSuite.scala ## How was this patch tested? Ran the unit test after

spark git commit: [SPARK-17378][HOTFIX] Upgrade snappy-java to 1.1.2.6 -- fix Hadoop 1 deps

2016-09-07 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.6 3f797dd87 -> 69fa9458e [SPARK-17378][HOTFIX] Upgrade snappy-java to 1.1.2.6 -- fix Hadoop 1 deps ## What changes were proposed in this pull request? Also update Hadoop 1 deps file to reflect Snappy 1.1.2.6 ## How was this patch

spark git commit: [SPARK-17427][SQL] function SIZE should return -1 when parameter is null

2016-09-07 Thread hvanhovell
Repository: spark Updated Branches: refs/heads/master 6b41195bc -> 6f4aeccf8 [SPARK-17427][SQL] function SIZE should return -1 when parameter is null ## What changes were proposed in this pull request? `select size(null)` returns -1 in Hive. In order to be compatible, we should return `-1`.

spark git commit: [SPARK-17339][SPARKR][CORE] Fix some R tests and use Path.toUri in SparkContext for Windows paths in SparkR

2016-09-07 Thread sarutak
Repository: spark Updated Branches: refs/heads/master 3ce3a282c -> 6b41195bc [SPARK-17339][SPARKR][CORE] Fix some R tests and use Path.toUri in SparkContext for Windows paths in SparkR ## What changes were proposed in this pull request? This PR fixes the Windows path issues in several APIs.

spark git commit: [SPARK-17359][SQL][MLLIB] Use ArrayBuffer.+=(A) instead of ArrayBuffer.append(A) in performance critical paths

2016-09-07 Thread srowen
Repository: spark Updated Branches: refs/heads/master 9fccde4ff -> 3ce3a282c [SPARK-17359][SQL][MLLIB] Use ArrayBuffer.+=(A) instead of ArrayBuffer.append(A) in performance critical paths ## What changes were proposed in this pull request? We should generally use `ArrayBuffer.+=(A)` rather

spark git commit: [SPARK-17296][SQL] Simplify parser join processing [BACKPORT 2.0]

2016-09-07 Thread hvanhovell
Repository: spark Updated Branches: refs/heads/branch-2.0 ee6301a88 -> c8811adaa [SPARK-17296][SQL] Simplify parser join processing [BACKPORT 2.0] ## What changes were proposed in this pull request? This PR backports https://github.com/apache/spark/pull/14867 to branch-2.0. It fixes a number

spark git commit: [SPARK-16785] R dapply doesn't return array or raw columns

2016-09-07 Thread shivaram
Repository: spark Updated Branches: refs/heads/branch-2.0 796577b43 -> ee6301a88 [SPARK-16785] R dapply doesn't return array or raw columns Fixed bug in `dapplyCollect` by changing the `compute` function of `worker.R` to explicitly handle raw (binary) vectors. cc shivaram Unit tests

spark git commit: [SPARK-16785] R dapply doesn't return array or raw columns

2016-09-07 Thread shivaram
Repository: spark Updated Branches: refs/heads/master eb1ab88a8 -> 9fccde4ff [SPARK-16785] R dapply doesn't return array or raw columns ## What changes were proposed in this pull request? Fixed bug in `dapplyCollect` by changing the `compute` function of `worker.R` to explicitly handle raw