svn commit: r28265 - in /dev/spark/2.3.3-SNAPSHOT-2018_07_20_22_01-bd6bfac-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-07-20 Thread pwendell
Author: pwendell Date: Sat Jul 21 05:15:52 2018 New Revision: 28265 Log: Apache Spark 2.3.3-SNAPSHOT-2018_07_20_22_01-bd6bfac docs [This commit notification would consist of 1443 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

svn commit: r28264 - in /dev/spark/2.4.0-SNAPSHOT-2018_07_20_20_01-bbd6f0c-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-07-20 Thread pwendell
Author: pwendell Date: Sat Jul 21 03:15:51 2018 New Revision: 28264 Log: Apache Spark 2.4.0-SNAPSHOT-2018_07_20_20_01-bbd6f0c docs [This commit notification would consist of 1469 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-24836][SQL] New option for Avro datasource - ignoreExtension

2018-07-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master bbd6f0c25 -> 106880edc [SPARK-24836][SQL] New option for Avro datasource - ignoreExtension ## What changes were proposed in this pull request? I propose to add new option for AVRO datasource which should control ignoring of files without

spark git commit: [SPARK-24879][SQL] Fix NPE in Hive partition pruning filter pushdown

2018-07-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.3 db1f3cc76 -> bd6bfacb2 [SPARK-24879][SQL] Fix NPE in Hive partition pruning filter pushdown ## What changes were proposed in this pull request? We get a NPE when we have a filter on a partition column of the form `col in (x, null)`.

spark git commit: [SPARK-24879][SQL] Fix NPE in Hive partition pruning filter pushdown

2018-07-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 96f312076 -> bbd6f0c25 [SPARK-24879][SQL] Fix NPE in Hive partition pruning filter pushdown ## What changes were proposed in this pull request? We get a NPE when we have a filter on a partition column of the form `col in (x, null)`. This

spark git commit: [PYSPARK][TEST][MINOR] Fix UDFInitializationTests

2018-07-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 597bdeff2 -> 96f312076 [PYSPARK][TEST][MINOR] Fix UDFInitializationTests ## What changes were proposed in this pull request? Fix a typo in pyspark sql tests Author: William Sheu Closes #21833 from PenguinToast/fix-test-typo. Project:

svn commit: r28262 - in /dev/spark/2.4.0-SNAPSHOT-2018_07_20_16_01-597bdef-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-07-20 Thread pwendell
Author: pwendell Date: Fri Jul 20 23:15:52 2018 New Revision: 28262 Log: Apache Spark 2.4.0-SNAPSHOT-2018_07_20_16_01-597bdef docs [This commit notification would consist of 1469 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-24488][SQL] Fix issue when generator is aliased multiple times

2018-07-20 Thread hvanhovell
Repository: spark Updated Branches: refs/heads/master f765bb782 -> 597bdeff2 [SPARK-24488][SQL] Fix issue when generator is aliased multiple times ## What changes were proposed in this pull request? Currently, the Analyzer throws an exception if your try to nest a generator. However, it

spark git commit: [SPARK-24880][BUILD] Fix the group id for spark-kubernetes-integration-tests

2018-07-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 00b864aa7 -> f765bb782 [SPARK-24880][BUILD] Fix the group id for spark-kubernetes-integration-tests ## What changes were proposed in this pull request? The correct group id should be `org.apache.spark`. This is causing the nightly build

spark git commit: [SPARK-24876][SQL] Avro: simplify schema serialization

2018-07-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 2333a34d3 -> 00b864aa7 [SPARK-24876][SQL] Avro: simplify schema serialization ## What changes were proposed in this pull request? Previously in the refactoring of Avro Serializer and Deserializer, a new class SerializableSchema is

spark git commit: [SPARK-22880][SQL] Add cascadeTruncate option to JDBC datasource

2018-07-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 9ad77b303 -> 2333a34d3 [SPARK-22880][SQL] Add cascadeTruncate option to JDBC datasource This commit adds the `cascadeTruncate` option to the JDBC datasource API, for databases that support this functionality (PostgreSQL and Oracle at the

spark git commit: Revert "[SPARK-24811][SQL] Avro: add new function from_avro and to_avro"

2018-07-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 3cb1b5780 -> 9ad77b303 Revert "[SPARK-24811][SQL] Avro: add new function from_avro and to_avro" This reverts commit 244bcff19463d82ec72baf15bc0a5209f21f2ef3. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit:

svn commit: r28256 - in /dev/spark/2.4.0-SNAPSHOT-2018_07_20_12_01-244bcff-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-07-20 Thread pwendell
Author: pwendell Date: Fri Jul 20 19:15:53 2018 New Revision: 28256 Log: Apache Spark 2.4.0-SNAPSHOT-2018_07_20_12_01-244bcff docs [This commit notification would consist of 1469 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-24852][ML] Update spark.ml to use Instrumentation.instrumented.

2018-07-20 Thread jkbradley
Repository: spark Updated Branches: refs/heads/master 244bcff19 -> 3cb1b5780 [SPARK-24852][ML] Update spark.ml to use Instrumentation.instrumented. ## What changes were proposed in this pull request? Followup for #21719. Update spark.ml training code to fully wrap instrumented methods and

spark git commit: [SPARK-24811][SQL] Avro: add new function from_avro and to_avro

2018-07-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master cc4d64bb1 -> 244bcff19 [SPARK-24811][SQL] Avro: add new function from_avro and to_avro ## What changes were proposed in this pull request? Add a new function from_avro for parsing a binary column of avro format and converting it into its

spark git commit: [SPARK-23451][ML] Deprecate KMeans.computeCost

2018-07-20 Thread holden
Repository: spark Updated Branches: refs/heads/master e0b638321 -> cc4d64bb1 [SPARK-23451][ML] Deprecate KMeans.computeCost ## What changes were proposed in this pull request? Deprecate `KMeans.computeCost` which was introduced as a temp fix and now it is not needed anymore, since we

svn commit: r28254 - in /dev/spark/2.4.0-SNAPSHOT-2018_07_20_08_01-e0b6383-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-07-20 Thread pwendell
Author: pwendell Date: Fri Jul 20 15:16:26 2018 New Revision: 28254 Log: Apache Spark 2.4.0-SNAPSHOT-2018_07_20_08_01-e0b6383 docs [This commit notification would consist of 1469 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-23731][SQL] Make FileSourceScanExec canonicalizable after being (de)serialized

2018-07-20 Thread wenchen
Repository: spark Updated Branches: refs/heads/branch-2.3 d0280ab81 -> db1f3cc76 [SPARK-23731][SQL] Make FileSourceScanExec canonicalizable after being (de)serialized ## What changes were proposed in this pull request? ### What's problem? In some cases, sub scalar query could throw a NPE,

spark git commit: [SPARK-23731][SQL] Make FileSourceScanExec canonicalizable after being (de)serialized

2018-07-20 Thread wenchen
Repository: spark Updated Branches: refs/heads/master 20ce1a8f8 -> e0b638321 [SPARK-23731][SQL] Make FileSourceScanExec canonicalizable after being (de)serialized ## What changes were proposed in this pull request? ### What's problem? In some cases, sub scalar query could throw a NPE,

spark git commit: [SPARK-24551][K8S] Add integration tests for secrets

2018-07-20 Thread srowen
Repository: spark Updated Branches: refs/heads/master 7b6d36bc9 -> 20ce1a8f8 [SPARK-24551][K8S] Add integration tests for secrets ## What changes were proposed in this pull request? - Adds integration tests for env and mount secrets. ## How was this patch tested? Manually by checking that

spark git commit: [SPARK-24871][SQL] Refactor Concat and MapConcat to avoid creating concatenator object for each row.

2018-07-20 Thread wenchen
Repository: spark Updated Branches: refs/heads/master 0ab07b357 -> 7b6d36bc9 [SPARK-24871][SQL] Refactor Concat and MapConcat to avoid creating concatenator object for each row. ## What changes were proposed in this pull request? Refactor `Concat` and `MapConcat` to: - avoid creating

svn commit: r28251 - in /dev/spark/2.4.0-SNAPSHOT-2018_07_20_04_01-0ab07b3-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-07-20 Thread pwendell
Author: pwendell Date: Fri Jul 20 11:19:36 2018 New Revision: 28251 Log: Apache Spark 2.4.0-SNAPSHOT-2018_07_20_04_01-0ab07b3 docs [This commit notification would consist of 1469 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-24868][PYTHON] add sequence function in Python

2018-07-20 Thread gurwls223
Repository: spark Updated Branches: refs/heads/master 2b91d9918 -> 0ab07b357 [SPARK-24868][PYTHON] add sequence function in Python ## What changes were proposed in this pull request? Add ```sequence``` in functions.py ## How was this patch tested? Add doctest. Author: Huaxin Gao Closes

svn commit: r28244 - in /dev/spark/2.4.0-SNAPSHOT-2018_07_20_00_01-2b91d99-docs: ./ _site/ _site/api/ _site/api/R/ _site/api/java/ _site/api/java/lib/ _site/api/java/org/ _site/api/java/org/apache/ _s

2018-07-20 Thread pwendell
Author: pwendell Date: Fri Jul 20 07:17:44 2018 New Revision: 28244 Log: Apache Spark 2.4.0-SNAPSHOT-2018_07_20_00_01-2b91d99 docs [This commit notification would consist of 1469 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.]

spark git commit: [SPARK-24424][SQL] Support ANSI-SQL compliant syntax for GROUPING SET

2018-07-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master a5925c163 -> 2b91d9918 [SPARK-24424][SQL] Support ANSI-SQL compliant syntax for GROUPING SET ## What changes were proposed in this pull request? Enhances the parser and analyzer to support ANSI compliant syntax for GROUPING SET. As part

[2/2] spark git commit: [SPARK-24268][SQL] Use datatype.catalogString in error messages

2018-07-20 Thread lixiao
[SPARK-24268][SQL] Use datatype.catalogString in error messages ## What changes were proposed in this pull request? As stated in https://github.com/apache/spark/pull/21321, in the error messages we should use `catalogString`. This is not the case, as SPARK-22893 used `simpleString` in order to

[1/2] spark git commit: [SPARK-24268][SQL] Use datatype.catalogString in error messages

2018-07-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 1462b1766 -> a5925c163 http://git-wip-us.apache.org/repos/asf/spark/blob/a5925c16/sql/core/src/main/java/org/apache/spark/sql/execution/datasources/parquet/VectorizedColumnReader.java