spark git commit: [SPARK-23418][SQL] Fail DataSourceV2 reads when user schema is passed, but not supported.

2018-02-20 Thread wenchen
Repository: spark Updated Branches: refs/heads/master 95e25ed1a -> c8c4441df [SPARK-23418][SQL] Fail DataSourceV2 reads when user schema is passed, but not supported. ## What changes were proposed in this pull request? DataSourceV2 initially allowed user-supplied schemas when a source

svn commit: r25179 - in /dev/spark/2.3.1-SNAPSHOT-2018_02_20_22_01-3e7269e-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-02-20 Thread pwendell
Author: pwendell Date: Wed Feb 21 06:15:33 2018 New Revision: 25179 Log: Apache Spark 2.3.1-SNAPSHOT-2018_02_20_22_01-3e7269e docs [This commit notification would consist of 1443 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

svn commit: r25177 - in /dev/spark/2.4.0-SNAPSHOT-2018_02_20_20_01-95e25ed-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-02-20 Thread pwendell
Author: pwendell Date: Wed Feb 21 04:15:56 2018 New Revision: 25177 Log: Apache Spark 2.4.0-SNAPSHOT-2018_02_20_20_01-95e25ed docs [This commit notification would consist of 1444 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-23424][SQL] Add codegenStageId in comment

2018-02-20 Thread wenchen
Repository: spark Updated Branches: refs/heads/master 601d653bf -> 95e25ed1a [SPARK-23424][SQL] Add codegenStageId in comment ## What changes were proposed in this pull request? This PR always adds `codegenStageId` in comment of the generated class. This is a replication of #20419 for

spark git commit: [SPARK-23053][CORE][BRANCH-2.1] taskBinarySerialization and task partitions calculate in DagScheduler.submitMissingTasks should keep the same RDD checkpoint status

2018-02-20 Thread irashid
Repository: spark Updated Branches: refs/heads/branch-2.1 1df8020e1 -> 24fe6eb0f [SPARK-23053][CORE][BRANCH-2.1] taskBinarySerialization and task partitions calculate in DagScheduler.submitMissingTasks should keep the same RDD checkpoint status ## What changes were proposed in this pull

svn commit: r25174 - in /dev/spark/2.3.1-SNAPSHOT-2018_02_20_18_02-a1ee6f1-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-02-20 Thread pwendell
Author: pwendell Date: Wed Feb 21 02:17:26 2018 New Revision: 25174 Log: Apache Spark 2.3.1-SNAPSHOT-2018_02_20_18_02-a1ee6f1 docs [This commit notification would consist of 1443 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-23454][SS][DOCS] Added trigger information to the Structured Streaming programming guide

2018-02-20 Thread tdas
Repository: spark Updated Branches: refs/heads/master 6d398c05c -> 601d653bf [SPARK-23454][SS][DOCS] Added trigger information to the Structured Streaming programming guide ## What changes were proposed in this pull request? - Added clear information about triggers - Made the semantics

spark git commit: [SPARK-23454][SS][DOCS] Added trigger information to the Structured Streaming programming guide

2018-02-20 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-2.3 1d78f03ae -> 3e7269eb9 [SPARK-23454][SS][DOCS] Added trigger information to the Structured Streaming programming guide ## What changes were proposed in this pull request? - Added clear information about triggers - Made the semantics

spark git commit: [SPARK-23468][CORE] Stringify auth secret before storing it in credentials.

2018-02-20 Thread vanzin
Repository: spark Updated Branches: refs/heads/branch-2.3 a1ee6f1fc -> 1d78f03ae [SPARK-23468][CORE] Stringify auth secret before storing it in credentials. The secret is used as a string in many parts of the code, so it has to be turned into a hex string to avoid issues such as the random

spark git commit: [SPARK-23468][CORE] Stringify auth secret before storing it in credentials.

2018-02-20 Thread vanzin
Repository: spark Updated Branches: refs/heads/master 2ba77ed9e -> 6d398c05c [SPARK-23468][CORE] Stringify auth secret before storing it in credentials. The secret is used as a string in many parts of the code, so it has to be turned into a hex string to avoid issues such as the random byte

spark git commit: [SPARK-23470][UI] Use first attempt of last stage to define job description.

2018-02-20 Thread sameerag
Repository: spark Updated Branches: refs/heads/branch-2.3 c7a0dea46 -> a1ee6f1fc [SPARK-23470][UI] Use first attempt of last stage to define job description. This is much faster than finding out what the last attempt is, and the data should be the same. There's room for improvement in this

spark git commit: [SPARK-23470][UI] Use first attempt of last stage to define job description.

2018-02-20 Thread sameerag
Repository: spark Updated Branches: refs/heads/master 3e48f3b9e -> 2ba77ed9e [SPARK-23470][UI] Use first attempt of last stage to define job description. This is much faster than finding out what the last attempt is, and the data should be the same. There's room for improvement in this page

spark git commit: [SPARK-23434][SQL] Spark should not warn `metadata directory` for a HDFS file path

2018-02-20 Thread zsxwing
Repository: spark Updated Branches: refs/heads/master 83c008762 -> 3e48f3b9e [SPARK-23434][SQL] Spark should not warn `metadata directory` for a HDFS file path ## What changes were proposed in this pull request? In a kerberized cluster, when Spark reads a file path (e.g. `people.json`), it

svn commit: r25169 - in /dev/spark/2.4.0-SNAPSHOT-2018_02_20_12_01-83c0087-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-02-20 Thread pwendell
Author: pwendell Date: Tue Feb 20 20:15:42 2018 New Revision: 25169 Log: Apache Spark 2.4.0-SNAPSHOT-2018_02_20_12_01-83c0087 docs [This commit notification would consist of 1444 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-23456][SPARK-21783] Turn on `native` ORC impl and PPD by default

2018-02-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 189f56f3d -> 83c008762 [SPARK-23456][SPARK-21783] Turn on `native` ORC impl and PPD by default ## What changes were proposed in this pull request? Apache Spark 2.3 introduced `native` ORC supports with vectorization and many fixes.

svn commit: r25166 - in /dev/spark/2.4.0-SNAPSHOT-2018_02_20_08_01-189f56f-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-02-20 Thread pwendell
Author: pwendell Date: Tue Feb 20 16:18:11 2018 New Revision: 25166 Log: Apache Spark 2.4.0-SNAPSHOT-2018_02_20_08_01-189f56f docs [This commit notification would consist of 1444 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-23383][BUILD][MINOR] Make a distribution should exit with usage while detecting wrong options

2018-02-20 Thread srowen
Repository: spark Updated Branches: refs/heads/master 862fa697d -> 189f56f3d [SPARK-23383][BUILD][MINOR] Make a distribution should exit with usage while detecting wrong options ## What changes were proposed in this pull request? ```shell ./dev/make-distribution.sh --name ne-1.0.0-SNAPSHOT

svn commit: r25160 - in /dev/spark/2.4.0-SNAPSHOT-2018_02_20_04_01-862fa69-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-02-20 Thread pwendell
Author: pwendell Date: Tue Feb 20 12:18:19 2018 New Revision: 25160 Log: Apache Spark 2.4.0-SNAPSHOT-2018_02_20_04_01-862fa69 docs [This commit notification would consist of 1444 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-23240][PYTHON] Better error message when extraneous data in pyspark.daemon's stdout

2018-02-20 Thread gurwls223
Repository: spark Updated Branches: refs/heads/master aadf9535b -> 862fa697d [SPARK-23240][PYTHON] Better error message when extraneous data in pyspark.daemon's stdout ## What changes were proposed in this pull request? Print more helpful message when daemon module's stdout is empty or

svn commit: r25156 - in /dev/spark/2.4.0-SNAPSHOT-2018_02_20_00_01-651b027-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-02-20 Thread pwendell
Author: pwendell Date: Tue Feb 20 08:17:45 2018 New Revision: 25156 Log: Apache Spark 2.4.0-SNAPSHOT-2018_02_20_00_01-651b027 docs [This commit notification would consist of 1444 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-23203][SQL] DataSourceV2: Use immutable logical plans.

2018-02-20 Thread wenchen
Repository: spark Updated Branches: refs/heads/master 651b0277f -> aadf9535b [SPARK-23203][SQL] DataSourceV2: Use immutable logical plans. ## What changes were proposed in this pull request? SPARK-23203: DataSourceV2 should use immutable catalyst trees instead of wrapping a mutable