spark git commit: [SPARK-10169] [SQL] [BRANCH-1.4] Partial aggregation's plan is wrong when a grouping expression is used as an argument of the aggregate fucntion

2015-08-24 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.4 c73498773 - 2671551a9 [SPARK-10169] [SQL] [BRANCH-1.4] Partial aggregation's plan is wrong when a grouping expression is used as an argument of the aggregate fucntion https://issues.apache.org/jira/browse/SPARK-10169 Author: Yin Huai

spark git commit: [SPARK-9791] [PACKAGE] Change private class to private class to prevent unnecessary classes from showing up in the docs

2015-08-24 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-1.5 36bc50c8d - d003373bd [SPARK-9791] [PACKAGE] Change private class to private class to prevent unnecessary classes from showing up in the docs In addition, some random cleanup of import ordering Author: Tathagata Das

spark git commit: [SPARK-10169] [SQL] [BRANCH-1.3] Partial aggregation's plan is wrong when a grouping expression is used as an argument of the aggregate fucntion

2015-08-24 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.3 a98603f8c - 3d2eaf0a7 [SPARK-10169] [SQL] [BRANCH-1.3] Partial aggregation's plan is wrong when a grouping expression is used as an argument of the aggregate fucntion https://issues.apache.org/jira/browse/SPARK-10169 Author: Wenchen

spark git commit: [SPARK-9758] [TEST] [SQL] Compilation issue for hive test / wrong package?

2015-08-24 Thread srowen
Repository: spark Updated Branches: refs/heads/master a2f4cdceb - cb2d2e158 [SPARK-9758] [TEST] [SQL] Compilation issue for hive test / wrong package? Move `test.org.apache.spark.sql.hive` package tests to apparent intended `org.apache.spark.sql.hive` as they don't intend to test behavior

spark git commit: [SPARK-9758] [TEST] [SQL] Compilation issue for hive test / wrong package?

2015-08-24 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.5 d36f3517c - 92234439d [SPARK-9758] [TEST] [SQL] Compilation issue for hive test / wrong package? Move `test.org.apache.spark.sql.hive` package tests to apparent intended `org.apache.spark.sql.hive` as they don't intend to test

spark git commit: [SPARK-10168] [STREAMING] Fix the issue that maven publishes wrong artifact jars

2015-08-24 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-1.5 b40059dbd - 36bc50c8d [SPARK-10168] [STREAMING] Fix the issue that maven publishes wrong artifact jars This PR removed the `outputFile` configuration from pom.xml and updated `tests.py` to search jars for both sbt build and maven

spark git commit: [SPARK-7710] [SPARK-7998] [DOCS] Docs for DataFrameStatFunctions

2015-08-24 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.5 d003373bd - 43dcf95e4 [SPARK-7710] [SPARK-7998] [DOCS] Docs for DataFrameStatFunctions This PR contains examples on how to use some of the Stat Functions available for DataFrames under `df.stat`. rxin Author: Burak Yavuz

spark git commit: [SPARK-10144] [UI] Actually show peak execution memory by default

2015-08-24 Thread yhuai
Repository: spark Updated Branches: refs/heads/master 9ce0c7ad3 - 662bb9667 [SPARK-10144] [UI] Actually show peak execution memory by default The peak execution memory metric was introduced in SPARK-8735. That was before Tungsten was enabled by default, so it assumed that

spark git commit: [SPARK-8580] [SQL] Refactors ParquetHiveCompatibilitySuite and adds more test cases

2015-08-24 Thread davies
Repository: spark Updated Branches: refs/heads/master 662bb9667 - a2f4cdceb [SPARK-8580] [SQL] Refactors ParquetHiveCompatibilitySuite and adds more test cases This PR refactors `ParquetHiveCompatibilitySuite` so that it's easier to add new test cases. Hit two bugs, SPARK-10177 and

spark git commit: [SPARK-10144] [UI] Actually show peak execution memory by default

2015-08-24 Thread yhuai
Repository: spark Updated Branches: refs/heads/branch-1.5 43dcf95e4 - 831f78ee5 [SPARK-10144] [UI] Actually show peak execution memory by default The peak execution memory metric was introduced in SPARK-8735. That was before Tungsten was enabled by default, so it assumed that

spark git commit: [SPARK-9791] [PACKAGE] Change private class to private class to prevent unnecessary classes from showing up in the docs

2015-08-24 Thread tdas
Repository: spark Updated Branches: refs/heads/master 4e0395ddb - 7478c8b66 [SPARK-9791] [PACKAGE] Change private class to private class to prevent unnecessary classes from showing up in the docs In addition, some random cleanup of import ordering Author: Tathagata Das

spark git commit: [SPARK-7710] [SPARK-7998] [DOCS] Docs for DataFrameStatFunctions

2015-08-24 Thread rxin
Repository: spark Updated Branches: refs/heads/master 7478c8b66 - 9ce0c7ad3 [SPARK-7710] [SPARK-7998] [DOCS] Docs for DataFrameStatFunctions This PR contains examples on how to use some of the Stat Functions available for DataFrames under `df.stat`. rxin Author: Burak Yavuz

Git Push Summary

2015-08-24 Thread rxin
Repository: spark Updated Tags: refs/tags/v1.5.0-rc2 [deleted] e2569282a - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org

spark git commit: [SPARK-10168] [STREAMING] Fix the issue that maven publishes wrong artifact jars

2015-08-24 Thread tdas
Repository: spark Updated Branches: refs/heads/master 053d94fcf - 4e0395ddb [SPARK-10168] [STREAMING] Fix the issue that maven publishes wrong artifact jars This PR removed the `outputFile` configuration from pom.xml and updated `tests.py` to search jars for both sbt build and maven build.

spark git commit: [SPARK-8580] [SQL] Refactors ParquetHiveCompatibilitySuite and adds more test cases

2015-08-24 Thread davies
Repository: spark Updated Branches: refs/heads/branch-1.5 831f78ee5 - d36f3517c [SPARK-8580] [SQL] Refactors ParquetHiveCompatibilitySuite and adds more test cases This PR refactors `ParquetHiveCompatibilitySuite` so that it's easier to add new test cases. Hit two bugs, SPARK-10177 and

spark git commit: [SPARK-10190] Fix NPE in CatalystTypeConverters Decimal toScala converter

2015-08-24 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.5 aadb9de4c - a4bad5f25 [SPARK-10190] Fix NPE in CatalystTypeConverters Decimal toScala converter This adds a missing null check to the Decimal `toScala` converter in `CatalystTypeConverters`, fixing an NPE. Author: Josh Rosen

spark git commit: [SPARK-10190] Fix NPE in CatalystTypeConverters Decimal toScala converter

2015-08-24 Thread rxin
Repository: spark Updated Branches: refs/heads/master 13db11cb0 - d7b4c0952 [SPARK-10190] Fix NPE in CatalystTypeConverters Decimal toScala converter This adds a missing null check to the Decimal `toScala` converter in `CatalystTypeConverters`, fixing an NPE. Author: Josh Rosen

spark git commit: [SPARK-10118] [SPARKR] [DOCS] Improve SparkR API docs for 1.5 release

2015-08-24 Thread shivaram
Repository: spark Updated Branches: refs/heads/master 2bf338c62 - 6511bf559 [SPARK-10118] [SPARKR] [DOCS] Improve SparkR API docs for 1.5 release cc: shivaram ## Summary - Modify `tdname` of expression functions. i.e. `ascii`: `rdname functions` = `rdname ascii` - Replace the dynamical

spark git commit: HOTFIX: Adding missing 1.4.1 ec2 version

2015-08-24 Thread pwendell
Repository: spark Updated Branches: refs/heads/branch-1.5 a4bad5f25 - 8ca8bdd01 HOTFIX: Adding missing 1.4.1 ec2 version Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8ca8bdd0 Tree:

spark git commit: [SPARK-10165] [SQL] Await child resolution in ResolveFunctions

2015-08-24 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.5 8ca8bdd01 - 228e429eb [SPARK-10165] [SQL] Await child resolution in ResolveFunctions Currently, we eagerly attempt to resolve functions, even before their children are resolved. However, this is not valid in cases where we need to

spark git commit: [SPARK-10165] [SQL] Await child resolution in ResolveFunctions

2015-08-24 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master d7b4c0952 - 2bf338c62 [SPARK-10165] [SQL] Await child resolution in ResolveFunctions Currently, we eagerly attempt to resolve functions, even before their children are resolved. However, this is not valid in cases where we need to know

spark git commit: [SPARK-10061] [DOC] ML ensemble docs

2015-08-24 Thread meng
Repository: spark Updated Branches: refs/heads/branch-1.5 92234439d - aadb9de4c [SPARK-10061] [DOC] ML ensemble docs User guide for spark.ml GBTs and Random Forests. The examples are copied from the decision tree guide and modified to run. I caught some issues I had somehow missed in the

spark git commit: [SPARK-10061] [DOC] ML ensemble docs

2015-08-24 Thread meng
Repository: spark Updated Branches: refs/heads/master cb2d2e158 - 13db11cb0 [SPARK-10061] [DOC] ML ensemble docs User guide for spark.ml GBTs and Random Forests. The examples are copied from the decision tree guide and modified to run. I caught some issues I had somehow missed in the tree

spark git commit: [SPARK-10118] [SPARKR] [DOCS] Improve SparkR API docs for 1.5 release

2015-08-24 Thread shivaram
Repository: spark Updated Branches: refs/heads/branch-1.5 228e429eb - ec5d09c0f [SPARK-10118] [SPARKR] [DOCS] Improve SparkR API docs for 1.5 release cc: shivaram ## Summary - Modify `tdname` of expression functions. i.e. `ascii`: `rdname functions` = `rdname ascii` - Replace the dynamical

spark git commit: [SQL] [MINOR] [DOC] Clarify docs for inferring DataFrame from RDD of Products

2015-08-24 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.5 ec5d09c0f - 2f7e4b416 [SQL] [MINOR] [DOC] Clarify docs for inferring DataFrame from RDD of Products * Makes `SQLImplicits.rddToDataFrameHolder` scaladoc consistent with `SQLContext.createDataFrame[A : Product](rdd: RDD[A])` since the

spark git commit: [SQL] [MINOR] [DOC] Clarify docs for inferring DataFrame from RDD of Products

2015-08-24 Thread rxin
Repository: spark Updated Branches: refs/heads/master 6511bf559 - 642c43c81 [SQL] [MINOR] [DOC] Clarify docs for inferring DataFrame from RDD of Products * Makes `SQLImplicits.rddToDataFrameHolder` scaladoc consistent with `SQLContext.createDataFrame[A : Product](rdd: RDD[A])` since the

spark git commit: [SPARK-10121] [SQL] Thrift server always use the latest class loader provided by the conf of executionHive's state

2015-08-24 Thread lian
Repository: spark Updated Branches: refs/heads/branch-1.5 2f7e4b416 - c99f4160b [SPARK-10121] [SQL] Thrift server always use the latest class loader provided by the conf of executionHive's state https://issues.apache.org/jira/browse/SPARK-10121 Looks like the problem is that if we add a jar

spark git commit: [SPARK-10121] [SQL] Thrift server always use the latest class loader provided by the conf of executionHive's state

2015-08-24 Thread lian
Repository: spark Updated Branches: refs/heads/master 642c43c81 - a0c0aae1d [SPARK-10121] [SQL] Thrift server always use the latest class loader provided by the conf of executionHive's state https://issues.apache.org/jira/browse/SPARK-10121 Looks like the problem is that if we add a jar