spark git commit: [SPARK-16101][HOTFIX] Fix the build with Scala 2.10 by explicit typed argument

2017-01-23 Thread tdas
Repository: spark Updated Branches: refs/heads/master 60bd91a34 -> ec9493b44 [SPARK-16101][HOTFIX] Fix the build with Scala 2.10 by explicit typed argument ## What changes were proposed in this pull request? I goofed in https://github.com/apache/spark/pull/16669 which introduces the break in

spark git commit: [SPARK-19268][SS] Disallow adaptive query execution for streaming queries

2017-01-23 Thread zsxwing
Repository: spark Updated Branches: refs/heads/master e576c1ed7 -> 60bd91a34 [SPARK-19268][SS] Disallow adaptive query execution for streaming queries ## What changes were proposed in this pull request? As adaptive query execution may change the number of partitions in different batches, it

spark git commit: [SPARK-19268][SS] Disallow adaptive query execution for streaming queries

2017-01-23 Thread zsxwing
Repository: spark Updated Branches: refs/heads/branch-2.1 4a2be0902 -> 570e5e11d [SPARK-19268][SS] Disallow adaptive query execution for streaming queries ## What changes were proposed in this pull request? As adaptive query execution may change the number of partitions in different batches,

spark git commit: [SPARK-9435][SQL] Reuse function in Java UDF to correctly support expressions that require equality comparison between ScalaUDF

2017-01-23 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 ed5d1e725 -> 4a2be0902 [SPARK-9435][SQL] Reuse function in Java UDF to correctly support expressions that require equality comparison between ScalaUDF ## What changes were proposed in this pull request? Currently, running the codes in

spark git commit: [SPARK-9435][SQL] Reuse function in Java UDF to correctly support expressions that require equality comparison between ScalaUDF

2017-01-23 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 3bdf3ee86 -> e576c1ed7 [SPARK-9435][SQL] Reuse function in Java UDF to correctly support expressions that require equality comparison between ScalaUDF ## What changes were proposed in this pull request? Currently, running the codes in Jav

spark git commit: [SPARK-19272][SQL] Remove the param `viewOriginalText` from `CatalogTable`

2017-01-23 Thread wenchen
Repository: spark Updated Branches: refs/heads/master fcfd5d0bb -> 3bdf3ee86 [SPARK-19272][SQL] Remove the param `viewOriginalText` from `CatalogTable` ## What changes were proposed in this pull request? Hive will expand the view text, so it needs 2 fields: originalText and viewText. Since w

spark git commit: [SPARK-19290][SQL] add a new extending interface in Analyzer for post-hoc resolution

2017-01-23 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 49f5b0ae4 -> fcfd5d0bb [SPARK-19290][SQL] add a new extending interface in Analyzer for post-hoc resolution ## What changes were proposed in this pull request? To implement DDL commands, we added several analyzer rules in sql/hive module

spark git commit: [SPARK-17747][ML] WeightCol support non-double numeric datatypes

2017-01-23 Thread jkbradley
Repository: spark Updated Branches: refs/heads/master e4974721f -> 49f5b0ae4 [SPARK-17747][ML] WeightCol support non-double numeric datatypes ## What changes were proposed in this pull request? 1, add test for `WeightCol` in `MLTestingUtils.checkNumericTypes` 2, move datatype cast to `Predict

spark git commit: [SPARK-19306][CORE] Fix inconsistent state in DiskBlockObject when expection occurred

2017-01-23 Thread vanzin
Repository: spark Updated Branches: refs/heads/branch-2.1 1e07a7192 -> ed5d1e725 [SPARK-19306][CORE] Fix inconsistent state in DiskBlockObject when expection occurred ## What changes were proposed in this pull request? In `DiskBlockObjectWriter`, when some errors happened during writing, it

spark git commit: [SPARK-19306][CORE] Fix inconsistent state in DiskBlockObject when expection occurred

2017-01-23 Thread vanzin
Repository: spark Updated Branches: refs/heads/master 5b258b8b0 -> e4974721f [SPARK-19306][CORE] Fix inconsistent state in DiskBlockObject when expection occurred ## What changes were proposed in this pull request? In `DiskBlockObjectWriter`, when some errors happened during writing, it will

spark git commit: [SPARK-16473][MLLIB] Fix BisectingKMeans Algorithm failing in edge case

2017-01-23 Thread jkbradley
Repository: spark Updated Branches: refs/heads/master c8aea7445 -> 5b258b8b0 [SPARK-16473][MLLIB] Fix BisectingKMeans Algorithm failing in edge case [SPARK-16473][MLLIB] Fix BisectingKMeans Algorithm failing in edge case where no children exist in updateAssignments ## What changes were propo

spark git commit: [SPARK-17455][MLLIB] Improve PAVA implementation in IsotonicRegression

2017-01-23 Thread jkbradley
Repository: spark Updated Branches: refs/heads/master 4a11d029d -> c8aea7445 [SPARK-17455][MLLIB] Improve PAVA implementation in IsotonicRegression ## What changes were proposed in this pull request? New implementation of the Pool Adjacent Violators Algorithm (PAVA) in mllib.IsotonicRegressi

spark git commit: [SPARK-14709][ML] spark.ml API for linear SVM

2017-01-23 Thread jkbradley
Repository: spark Updated Branches: refs/heads/master 0ef1421a6 -> 4a11d029d [SPARK-14709][ML] spark.ml API for linear SVM ## What changes were proposed in this pull request? jira: https://issues.apache.org/jira/browse/SPARK-14709 Provide API for SVM algorithm for DataFrames. As discussed in

spark git commit: [SPARK-19284][SQL] append to partitioned datasource table should without custom partition location

2017-01-23 Thread wenchen
Repository: spark Updated Branches: refs/heads/master c99492141 -> 0ef1421a6 [SPARK-19284][SQL] append to partitioned datasource table should without custom partition location ## What changes were proposed in this pull request? when we append data to a existed partitioned datasource table, t

spark git commit: [SPARK-19146][CORE] Drop more elements when stageData.taskData.size > retainedTasks

2017-01-23 Thread srowen
Repository: spark Updated Branches: refs/heads/master c4a6519c4 -> c99492141 [SPARK-19146][CORE] Drop more elements when stageData.taskData.size > retainedTasks ## What changes were proposed in this pull request? Drop more elements when `stageData.taskData.size > retainedTasks` to reduce the

spark git commit: [SPARK-19218][SQL] Fix SET command to show a result correctly and in a sorted order

2017-01-23 Thread lixiao
Repository: spark Updated Branches: refs/heads/master f067acefa -> c4a6519c4 [SPARK-19218][SQL] Fix SET command to show a result correctly and in a sorted order ## What changes were proposed in this pull request? This PR aims to fix the following two things. 1. `sql("SET -v").collect()` or

spark git commit: [SPARK-19155][ML] Make family case insensitive in GLM

2017-01-23 Thread yliang
Repository: spark Updated Branches: refs/heads/branch-2.1 8daf10e3f -> 1e07a7192 [SPARK-19155][ML] Make family case insensitive in GLM ## What changes were proposed in this pull request? This is a supplement to PR #16516 which did not make the value from `getFamily` case insensitive. Current

spark git commit: [SPARK-19155][ML] Make family case insensitive in GLM

2017-01-23 Thread yliang
Repository: spark Updated Branches: refs/heads/master de6ad3dfa -> f067acefa [SPARK-19155][ML] Make family case insensitive in GLM ## What changes were proposed in this pull request? This is a supplement to PR #16516 which did not make the value from `getFamily` case insensitive. Current test