svn commit: r25786 - in /dev/spark/2.4.0-SNAPSHOT-2018_03_16_16_01-8a1efe3-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-03-16 Thread pwendell
Author: pwendell Date: Fri Mar 16 23:15:16 2018 New Revision: 25786 Log: Apache Spark 2.4.0-SNAPSHOT-2018_03_16_16_01-8a1efe3 docs [This commit notification would consist of 1449 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-23683][SQL] FileCommitProtocol.instantiate() hardening

2018-03-16 Thread wenchen
Repository: spark Updated Branches: refs/heads/master 8a72734f3 -> 8a1efe307 [SPARK-23683][SQL] FileCommitProtocol.instantiate() hardening ## What changes were proposed in this pull request? With SPARK-20236, `FileCommitProtocol.instantiate()` looks for a three argument constructor, passing

svn commit: r25780 - in /dev/spark/2.4.0-SNAPSHOT-2018_03_16_12_01-8a72734-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-03-16 Thread pwendell
Author: pwendell Date: Fri Mar 16 19:15:25 2018 New Revision: 25780 Log: Apache Spark 2.4.0-SNAPSHOT-2018_03_16_12_01-8a72734 docs [This commit notification would consist of 1448 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-15009][PYTHON][ML] Construct a CountVectorizerModel from a vocabulary list

2018-03-16 Thread holden
Repository: spark Updated Branches: refs/heads/master bd201bf61 -> 8a72734f3 [SPARK-15009][PYTHON][ML] Construct a CountVectorizerModel from a vocabulary list ## What changes were proposed in this pull request? Added a class method to construct CountVectorizerModel from a list of

spark git commit: [SPARK-23623][SS] Avoid concurrent use of cached consumers in CachedKafkaConsumer

2018-03-16 Thread zsxwing
Repository: spark Updated Branches: refs/heads/master 9945b0227 -> bd201bf61 [SPARK-23623][SS] Avoid concurrent use of cached consumers in CachedKafkaConsumer ## What changes were proposed in this pull request? CacheKafkaConsumer in the project `kafka-0-10-sql` is designed to maintain a

spark git commit: [SPARK-23680] Fix entrypoint.sh to properly support Arbitrary UIDs

2018-03-16 Thread eje
Repository: spark Updated Branches: refs/heads/master 88d8de926 -> 9945b0227 [SPARK-23680] Fix entrypoint.sh to properly support Arbitrary UIDs ## What changes were proposed in this pull request? As described in SPARK-23680, entrypoint.sh returns an error code because of a command pipeline

spark git commit: [SPARK-23581][SQL] Add interpreted unsafe projection

2018-03-16 Thread hvanhovell
Repository: spark Updated Branches: refs/heads/master dffeac369 -> 88d8de926 [SPARK-23581][SQL] Add interpreted unsafe projection ## What changes were proposed in this pull request? We currently can only create unsafe rows using code generation. This is a problem for situations in which code

[2/2] spark-website git commit: Squashed commit of the following:

2018-03-16 Thread rxin
Squashed commit of the following: commit 8e2dd71cf5613be6f019bb76b46226771422a40e Merge: 8bd24fb6d 01f0b4e0c Author: Reynold Xin Date: Fri Mar 16 10:24:54 2018 -0700 Merge pull request #104 from mateiz/history Add a project history page commit

[1/2] spark-website git commit: Squashed commit of the following:

2018-03-16 Thread rxin
Repository: spark-website Updated Branches: refs/heads/asf-site 8bd24fb6d -> a1d84bcbf http://git-wip-us.apache.org/repos/asf/spark-website/blob/a1d84bcb/site/news/spark-summit-june-2016-agenda-posted.html -- diff --git

spark git commit: [SPARK-18371][STREAMING] Spark Streaming backpressure generates batch with large number of records

2018-03-16 Thread koeninger
Repository: spark Updated Branches: refs/heads/master 5414abca4 -> dffeac369 [SPARK-18371][STREAMING] Spark Streaming backpressure generates batch with large number of records ## What changes were proposed in this pull request? Omit rounding of backpressure rate. Effects: - no batch with

svn commit: r25774 - in /dev/spark/2.3.1-SNAPSHOT-2018_03_16_10_01-21b6de4-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-03-16 Thread pwendell
Author: pwendell Date: Fri Mar 16 17:15:31 2018 New Revision: 25774 Log: Apache Spark 2.3.1-SNAPSHOT-2018_03_16_10_01-21b6de4 docs [This commit notification would consist of 1443 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-23553][TESTS] Tests should not assume the default value of `spark.sql.sources.default`

2018-03-16 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.3 d9e1f7040 -> 21b6de459 [SPARK-23553][TESTS] Tests should not assume the default value of `spark.sql.sources.default` ## What changes were proposed in this pull request? Currently, some tests have an assumption that

spark git commit: [SPARK-23553][TESTS] Tests should not assume the default value of `spark.sql.sources.default`

2018-03-16 Thread lixiao
Repository: spark Updated Branches: refs/heads/master c95200048 -> 5414abca4 [SPARK-23553][TESTS] Tests should not assume the default value of `spark.sql.sources.default` ## What changes were proposed in this pull request? Currently, some tests have an assumption that

svn commit: r25762 - in /dev/spark/2.4.0-SNAPSHOT-2018_03_16_04_01-c952000-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-03-16 Thread pwendell
Author: pwendell Date: Fri Mar 16 11:20:56 2018 New Revision: 25762 Log: Apache Spark 2.4.0-SNAPSHOT-2018_03_16_04_01-c952000 docs [This commit notification would consist of 1448 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-23635][YARN] AM env variable should not overwrite same name env variable set through spark.executorEnv.

2018-03-16 Thread jshao
Repository: spark Updated Branches: refs/heads/master ca83526de -> c95200048 [SPARK-23635][YARN] AM env variable should not overwrite same name env variable set through spark.executorEnv. ## What changes were proposed in this pull request? In the current Spark on YARN code, AM always will

spark git commit: [SPARK-23644][CORE][UI] Use absolute path for REST call in SHS

2018-03-16 Thread jshao
Repository: spark Updated Branches: refs/heads/master c2632edeb -> ca83526de [SPARK-23644][CORE][UI] Use absolute path for REST call in SHS ## What changes were proposed in this pull request? SHS is using a relative path for the REST API call to get the list of the application is a relative