spark git commit: [SPARK-19149][SQL] Unify two sets of statistics in LogicalPlan

2017-01-10 Thread rxin
Repository: spark Updated Branches: refs/heads/master 3b19c74e7 -> a61551356 [SPARK-19149][SQL] Unify two sets of statistics in LogicalPlan ## What changes were proposed in this pull request? Currently we have two sets of statistics in LogicalPlan: a simple stats and a stats estimated by cbo

spark git commit: [SPARK-19157][SQL] should be able to change spark.sql.runSQLOnFiles at runtime

2017-01-10 Thread lixiao
Repository: spark Updated Branches: refs/heads/master bc6c56e94 -> 3b19c74e7 [SPARK-19157][SQL] should be able to change spark.sql.runSQLOnFiles at runtime ## What changes were proposed in this pull request? The analyzer rule that supports to query files directly will be added to `Analyzer.e

spark git commit: [SPARK-19133][SPARKR][ML][BACKPORT-2.1] fix glm for Gamma, clarify glm family supported

2017-01-10 Thread felixcheung
Repository: spark Updated Branches: refs/heads/branch-2.1 230607d62 -> 1022049c7 [SPARK-19133][SPARKR][ML][BACKPORT-2.1] fix glm for Gamma, clarify glm family supported ## What changes were proposed in this pull request? backporting to 2.1, 2.0 and 1.6 ## How was this patch tested? unit te

spark git commit: [SPARK-19140][SS] Allow update mode for non-aggregation streaming queries

2017-01-10 Thread zsxwing
Repository: spark Updated Branches: refs/heads/branch-2.1 81c943090 -> 230607d62 [SPARK-19140][SS] Allow update mode for non-aggregation streaming queries ## What changes were proposed in this pull request? This PR allow update mode for non-aggregation streaming queries. It will be same as t

spark git commit: [SPARK-19140][SS] Allow update mode for non-aggregation streaming queries

2017-01-10 Thread zsxwing
Repository: spark Updated Branches: refs/heads/master 856bae6af -> bc6c56e94 [SPARK-19140][SS] Allow update mode for non-aggregation streaming queries ## What changes were proposed in this pull request? This PR allow update mode for non-aggregation streaming queries. It will be same as the a

spark git commit: [SPARK-18997][CORE] Recommended upgrade libthrift to 0.9.3

2017-01-10 Thread vanzin
Repository: spark Updated Branches: refs/heads/branch-2.0 e70c4198d -> 6fe676c09 [SPARK-18997][CORE] Recommended upgrade libthrift to 0.9.3 ## What changes were proposed in this pull request? Updates to libthrift 0.9.3 to address a CVE. ## How was this patch tested? Existing tests. Author:

spark git commit: [SPARK-18997][CORE] Recommended upgrade libthrift to 0.9.3

2017-01-10 Thread vanzin
Repository: spark Updated Branches: refs/heads/branch-2.1 e0af4b726 -> 81c943090 [SPARK-18997][CORE] Recommended upgrade libthrift to 0.9.3 ## What changes were proposed in this pull request? Updates to libthrift 0.9.3 to address a CVE. ## How was this patch tested? Existing tests. Author:

spark git commit: [SPARK-18997][CORE] Recommended upgrade libthrift to 0.9.3

2017-01-10 Thread vanzin
Repository: spark Updated Branches: refs/heads/master 9bc3507e4 -> 856bae6af [SPARK-18997][CORE] Recommended upgrade libthrift to 0.9.3 ## What changes were proposed in this pull request? Updates to libthrift 0.9.3 to address a CVE. ## How was this patch tested? Existing tests. Author: Sea

spark git commit: [SPARK-19133][SPARKR][ML] fix glm for Gamma, clarify glm family supported

2017-01-10 Thread felixcheung
Repository: spark Updated Branches: refs/heads/master d5b1dc934 -> 9bc3507e4 [SPARK-19133][SPARKR][ML] fix glm for Gamma, clarify glm family supported ## What changes were proposed in this pull request? R family is a longer list than what Spark supports. ## How was this patch tested? manual

spark git commit: [SPARK-19113][SS][TESTS] Set UncaughtExceptionHandler in onQueryStarted to ensure catching fatal errors during query initialization

2017-01-10 Thread zsxwing
Repository: spark Updated Branches: refs/heads/branch-2.1 69d1c4c5c -> e0af4b726 [SPARK-19113][SS][TESTS] Set UncaughtExceptionHandler in onQueryStarted to ensure catching fatal errors during query initialization ## What changes were proposed in this pull request? StreamTest sets `UncaughtEx

spark git commit: [SPARK-19137][SQL] Fix `withSQLConf` to reset `OptionalConfigEntry` correctly

2017-01-10 Thread zsxwing
Repository: spark Updated Branches: refs/heads/branch-2.1 65c866ef9 -> 69d1c4c5c [SPARK-19137][SQL] Fix `withSQLConf` to reset `OptionalConfigEntry` correctly ## What changes were proposed in this pull request? `DataStreamReaderWriterSuite` makes test files in source folder like the followin

spark git commit: [SPARK-19137][SQL] Fix `withSQLConf` to reset `OptionalConfigEntry` correctly

2017-01-10 Thread zsxwing
Repository: spark Updated Branches: refs/heads/master 3ef183a94 -> d5b1dc934 [SPARK-19137][SQL] Fix `withSQLConf` to reset `OptionalConfigEntry` correctly ## What changes were proposed in this pull request? `DataStreamReaderWriterSuite` makes test files in source folder like the followings.

spark git commit: [SPARK-19113][SS][TESTS] Set UncaughtExceptionHandler in onQueryStarted to ensure catching fatal errors during query initialization

2017-01-10 Thread srowen
Repository: spark Updated Branches: refs/heads/master a2c6adcc5 -> 3ef183a94 [SPARK-19113][SS][TESTS] Set UncaughtExceptionHandler in onQueryStarted to ensure catching fatal errors during query initialization ## What changes were proposed in this pull request? StreamTest sets `UncaughtExcept

spark git commit: [SPARK-18857][SQL] Don't use `Iterator.duplicate` for `incrementalCollect` in Thrift Server

2017-01-10 Thread srowen
Repository: spark Updated Branches: refs/heads/master 2cfd41ac0 -> a2c6adcc5 [SPARK-18857][SQL] Don't use `Iterator.duplicate` for `incrementalCollect` in Thrift Server ## What changes were proposed in this pull request? To support `FETCH_FIRST`, SPARK-16563 used Scala `Iterator.duplicate`.

spark git commit: [SPARK-19117][TESTS] Skip the tests using script transformation on Windows

2017-01-10 Thread srowen
Repository: spark Updated Branches: refs/heads/master 4e27578fa -> 2cfd41ac0 [SPARK-19117][TESTS] Skip the tests using script transformation on Windows ## What changes were proposed in this pull request? This PR proposes to skip the tests for script transformation failed on Windows due to fi

spark git commit: [SPARK-18922][SQL][CORE][STREAMING][TESTS] Fix all identified tests failed due to path and resource-not-closed problems on Windows

2017-01-10 Thread srowen
Repository: spark Updated Branches: refs/heads/master 32286ba68 -> 4e27578fa [SPARK-18922][SQL][CORE][STREAMING][TESTS] Fix all identified tests failed due to path and resource-not-closed problems on Windows ## What changes were proposed in this pull request? This PR proposes to fix all the

spark git commit: [SPARK-17645][MLLIB][ML][FOLLOW-UP] document minor change

2017-01-10 Thread srowen
Repository: spark Updated Branches: refs/heads/master acfc5f354 -> 32286ba68 [SPARK-17645][MLLIB][ML][FOLLOW-UP] document minor change ## What changes were proposed in this pull request? Add FDR test case in ml/feature/ChiSqSelectorSuite. Improve some comments in the code. This is a follow-up

spark git commit: [SPARK-16845][SQL] `GeneratedClass$SpecificOrdering` grows beyond 64 KB

2017-01-10 Thread wenchen
Repository: spark Updated Branches: refs/heads/branch-2.1 3b6ac323b -> 65c866ef9 [SPARK-16845][SQL] `GeneratedClass$SpecificOrdering` grows beyond 64 KB ## What changes were proposed in this pull request? Prior to this patch, we'll generate `compare(...)` for `GeneratedClass$SpecificOrdering

spark git commit: [SPARK-16845][SQL] `GeneratedClass$SpecificOrdering` grows beyond 64 KB

2017-01-10 Thread wenchen
Repository: spark Updated Branches: refs/heads/master b0319c2ec -> acfc5f354 [SPARK-16845][SQL] `GeneratedClass$SpecificOrdering` grows beyond 64 KB ## What changes were proposed in this pull request? Prior to this patch, we'll generate `compare(...)` for `GeneratedClass$SpecificOrdering` li

spark git commit: [SPARK-19107][SQL] support creating hive table with DataFrameWriter and Catalog

2017-01-10 Thread wenchen
Repository: spark Updated Branches: refs/heads/master b0e5840d4 -> b0319c2ec [SPARK-19107][SQL] support creating hive table with DataFrameWriter and Catalog ## What changes were proposed in this pull request? After unifying the CREATE TABLE syntax in https://github.com/apache/spark/pull/1629

spark git commit: [SPARK-19134][EXAMPLE] Fix several sql, mllib and status api examples not working

2017-01-10 Thread yliang
Repository: spark Updated Branches: refs/heads/master 3ef6d98a8 -> b0e5840d4 [SPARK-19134][EXAMPLE] Fix several sql, mllib and status api examples not working ## What changes were proposed in this pull request? **binary_classification_metrics_example.py** LibSVM datasource loads `ml.linalg.