spark git commit: [TRIVIAL][DOCS][STREAMING][SQL] The return type mentioned in the Javadoc is incorrect for toJavaRDD, …

2016-06-29 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-2.0 6650c0533 -> 904122335 [TRIVIAL][DOCS][STREAMING][SQL] The return type mentioned in the Javadoc is incorrect for toJavaRDD, … ## What changes were proposed in this pull request? Change the return type mentioned in the JavaDoc for `t

spark git commit: [TRIVIAL][DOCS][STREAMING][SQL] The return type mentioned in the Javadoc is incorrect for toJavaRDD, …

2016-06-29 Thread tdas
Repository: spark Updated Branches: refs/heads/master f454a7f9f -> 757dc2c09 [TRIVIAL][DOCS][STREAMING][SQL] The return type mentioned in the Javadoc is incorrect for toJavaRDD, … ## What changes were proposed in this pull request? Change the return type mentioned in the JavaDoc for `toJav

spark git commit: [SPARK-16291][SQL] CheckAnalysis should capture nested aggregate functions that reference no input attributes

2016-06-29 Thread wenchen
Repository: spark Updated Branches: refs/heads/master 757dc2c09 -> d1e810885 [SPARK-16291][SQL] CheckAnalysis should capture nested aggregate functions that reference no input attributes ## What changes were proposed in this pull request? `MAX(COUNT(*))` is invalid since aggregate expression

spark git commit: [SPARK-16291][SQL] CheckAnalysis should capture nested aggregate functions that reference no input attributes

2016-06-29 Thread wenchen
Repository: spark Updated Branches: refs/heads/branch-2.0 904122335 -> 1b4d63f6f [SPARK-16291][SQL] CheckAnalysis should capture nested aggregate functions that reference no input attributes ## What changes were proposed in this pull request? `MAX(COUNT(*))` is invalid since aggregate expres

spark git commit: [SPARK-16157][SQL] Add New Methods for comments in StructField and StructType

2016-06-29 Thread wenchen
Repository: spark Updated Branches: refs/heads/master d1e810885 -> 7ee9e39cb [SPARK-16157][SQL] Add New Methods for comments in StructField and StructType What changes were proposed in this pull request? Based on the previous discussion with cloud-fan hvanhovell in another related PR htt

spark git commit: [SPARK-16261][EXAMPLES][ML] Fixed incorrect appNames in ML Examples

2016-06-29 Thread mlnick
Repository: spark Updated Branches: refs/heads/master 7ee9e39cb -> 21385d02a [SPARK-16261][EXAMPLES][ML] Fixed incorrect appNames in ML Examples ## What changes were proposed in this pull request? Some appNames in ML examples are incorrect, mostly in PySpark but one in Scala. This corrects

spark git commit: [SPARK-16261][EXAMPLES][ML] Fixed incorrect appNames in ML Examples

2016-06-29 Thread mlnick
Repository: spark Updated Branches: refs/heads/branch-2.0 1b4d63f6f -> ba71cf451 [SPARK-16261][EXAMPLES][ML] Fixed incorrect appNames in ML Examples ## What changes were proposed in this pull request? Some appNames in ML examples are incorrect, mostly in PySpark but one in Scala. This corre

spark git commit: [SPARK-15858][ML] Fix calculating error by tree stack over flow prob…

2016-06-29 Thread srowen
Repository: spark Updated Branches: refs/heads/master 21385d02a -> 393db655c [SPARK-15858][ML] Fix calculating error by tree stack over flow prob… ## What changes were proposed in this pull request? What changes were proposed in this pull request? Improving evaluateEachIteration function i

spark git commit: [SPARK-15990][YARN] Add rolling log aggregation support for Spark on yarn

2016-06-29 Thread tgraves
Repository: spark Updated Branches: refs/heads/master 393db655c -> 272a2f78f [SPARK-15990][YARN] Add rolling log aggregation support for Spark on yarn ## What changes were proposed in this pull request? Yarn supports rolling log aggregation since 2.6, previously log will only be aggregated t

spark git commit: [MINOR][SPARKR] Fix arguments of survreg in SparkR

2016-06-29 Thread meng
Repository: spark Updated Branches: refs/heads/master 272a2f78f -> c6a220d75 [MINOR][SPARKR] Fix arguments of survreg in SparkR ## What changes were proposed in this pull request? Fix wrong arguments description of ```survreg``` in SparkR. ## How was this patch tested? ```Arguments``` section

spark git commit: [MINOR][SPARKR] Fix arguments of survreg in SparkR

2016-06-29 Thread meng
Repository: spark Updated Branches: refs/heads/branch-2.0 ba71cf451 -> d96e8c2dd [MINOR][SPARKR] Fix arguments of survreg in SparkR ## What changes were proposed in this pull request? Fix wrong arguments description of ```survreg``` in SparkR. ## How was this patch tested? ```Arguments``` sec

spark git commit: [SPARK-16140][MLLIB][SPARKR][DOCS] Group k-means method in generated R doc

2016-06-29 Thread meng
Repository: spark Updated Branches: refs/heads/master c6a220d75 -> 8c9cd0a7a [SPARK-16140][MLLIB][SPARKR][DOCS] Group k-means method in generated R doc https://issues.apache.org/jira/browse/SPARK-16140 ## What changes were proposed in this pull request? Group the R doc of spark.kmeans, predi

spark git commit: [SPARK-16140][MLLIB][SPARKR][DOCS] Group k-means method in generated R doc

2016-06-29 Thread meng
Repository: spark Updated Branches: refs/heads/branch-2.0 d96e8c2dd -> 1cde325e2 [SPARK-16140][MLLIB][SPARKR][DOCS] Group k-means method in generated R doc https://issues.apache.org/jira/browse/SPARK-16140 ## What changes were proposed in this pull request? Group the R doc of spark.kmeans, p

spark git commit: [SPARK-16236][SQL][FOLLOWUP] Add Path Option back to Load API in DataFrameReader

2016-06-29 Thread zsxwing
Repository: spark Updated Branches: refs/heads/branch-2.0 1cde325e2 -> edd1905c0 [SPARK-16236][SQL][FOLLOWUP] Add Path Option back to Load API in DataFrameReader What changes were proposed in this pull request? In Python API, we have the same issue. Thanks for identifying this issue, zsx

spark git commit: [SPARK-16236][SQL][FOLLOWUP] Add Path Option back to Load API in DataFrameReader

2016-06-29 Thread zsxwing
Repository: spark Updated Branches: refs/heads/master 8c9cd0a7a -> 39f2eb1da [SPARK-16236][SQL][FOLLOWUP] Add Path Option back to Load API in DataFrameReader What changes were proposed in this pull request? In Python API, we have the same issue. Thanks for identifying this issue, zsxwing

spark git commit: [SPARK-14480][SQL] Remove meaningless StringIteratorReader for CSV data source.

2016-06-29 Thread rxin
Repository: spark Updated Branches: refs/heads/master 39f2eb1da -> cb1b9d34f [SPARK-14480][SQL] Remove meaningless StringIteratorReader for CSV data source. ## What changes were proposed in this pull request? This PR removes meaningless `StringIteratorReader` for CSV data source. In `CSVPars

spark git commit: [SPARK-16256][SQL][STREAMING] Added Structured Streaming Programming Guide

2016-06-29 Thread tdas
Repository: spark Updated Branches: refs/heads/master cb1b9d34f -> 64132a14f [SPARK-16256][SQL][STREAMING] Added Structured Streaming Programming Guide Title defines all. Author: Tathagata Das Closes #13945 from tdas/SPARK-16256. Project: http://git-wip-us.apache.org/repos/asf/spark/repo

spark git commit: [SPARK-16256][SQL][STREAMING] Added Structured Streaming Programming Guide

2016-06-29 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-2.0 edd1905c0 -> 3cc258efb [SPARK-16256][SQL][STREAMING] Added Structured Streaming Programming Guide Title defines all. Author: Tathagata Das Closes #13945 from tdas/SPARK-16256. (cherry picked from commit 64132a14fb7a7255feeb5847a54f5

spark git commit: [SPARK-16044][SQL] Backport input_file_name() for data source based on NewHadoopRDD to branch 1.6

2016-06-29 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.6 0cb06c993 -> 1ac830aca [SPARK-16044][SQL] Backport input_file_name() for data source based on NewHadoopRDD to branch 1.6 ## What changes were proposed in this pull request? This PR backports https://github.com/apache/spark/pull/13759.

spark git commit: [TRIVIAL] [PYSPARK] Clean up orc compression option as well

2016-06-29 Thread davies
Repository: spark Updated Branches: refs/heads/master 64132a14f -> d8a87a3ed [TRIVIAL] [PYSPARK] Clean up orc compression option as well ## What changes were proposed in this pull request? This PR corrects ORC compression option for PySpark as well. I think this was missed mistakenly in http

spark git commit: [TRIVIAL] [PYSPARK] Clean up orc compression option as well

2016-06-29 Thread davies
Repository: spark Updated Branches: refs/heads/branch-2.0 3cc258efb -> 809af6d9d [TRIVIAL] [PYSPARK] Clean up orc compression option as well ## What changes were proposed in this pull request? This PR corrects ORC compression option for PySpark as well. I think this was missed mistakenly in

spark git commit: [SPARK-16301] [SQL] The analyzer rule for resolving using joins should respect the case sensitivity setting.

2016-06-29 Thread davies
Repository: spark Updated Branches: refs/heads/master d8a87a3ed -> 8b5a8b25b [SPARK-16301] [SQL] The analyzer rule for resolving using joins should respect the case sensitivity setting. ## What changes were proposed in this pull request? The analyzer rule for resolving using joins should resp

spark git commit: [SPARK-16301] [SQL] The analyzer rule for resolving using joins should respect the case sensitivity setting.

2016-06-29 Thread davies
Repository: spark Updated Branches: refs/heads/branch-2.0 809af6d9d -> a7f66ef62 [SPARK-16301] [SQL] The analyzer rule for resolving using joins should respect the case sensitivity setting. ## What changes were proposed in this pull request? The analyzer rule for resolving using joins should

spark git commit: [SPARK-16006][SQL] Attemping to write empty DataFrame with no fields throws non-intuitive exception

2016-06-29 Thread rxin
Repository: spark Updated Branches: refs/heads/master 8b5a8b25b -> 9b1b3ae77 [SPARK-16006][SQL] Attemping to write empty DataFrame with no fields throws non-intuitive exception ## What changes were proposed in this pull request? This PR allows `emptyDataFrame.write` since the user didn't spe

spark git commit: [SPARK-16006][SQL] Attemping to write empty DataFrame with no fields throws non-intuitive exception

2016-06-29 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-2.0 a7f66ef62 -> ef0253ff6 [SPARK-16006][SQL] Attemping to write empty DataFrame with no fields throws non-intuitive exception ## What changes were proposed in this pull request? This PR allows `emptyDataFrame.write` since the user didn't

spark git commit: [SPARK-16238] Metrics for generated method and class bytecode size

2016-06-29 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-2.0 ef0253ff6 -> c4cebd572 [SPARK-16238] Metrics for generated method and class bytecode size ## What changes were proposed in this pull request? This extends SPARK-15860 to include metrics for the actual bytecode size of janino-generated

spark git commit: [SPARK-16238] Metrics for generated method and class bytecode size

2016-06-29 Thread rxin
Repository: spark Updated Branches: refs/heads/master 9b1b3ae77 -> 23c58653f [SPARK-16238] Metrics for generated method and class bytecode size ## What changes were proposed in this pull request? This extends SPARK-15860 to include metrics for the actual bytecode size of janino-generated met

spark git commit: [SPARK-16228][SQL] HiveSessionCatalog should return `double`-param functions for decimal param lookups

2016-06-29 Thread rxin
Repository: spark Updated Branches: refs/heads/master 23c58653f -> 2eaabfa41 [SPARK-16228][SQL] HiveSessionCatalog should return `double`-param functions for decimal param lookups ## What changes were proposed in this pull request? This PR supports a fallback lookup by casting `DecimalType`

spark git commit: [SPARK-16228][SQL] HiveSessionCatalog should return `double`-param functions for decimal param lookups

2016-06-29 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-2.0 c4cebd572 -> 011befd20 [SPARK-16228][SQL] HiveSessionCatalog should return `double`-param functions for decimal param lookups ## What changes were proposed in this pull request? This PR supports a fallback lookup by casting `DecimalTy

spark git commit: [SPARK-16134][SQL] optimizer rules for typed filter

2016-06-29 Thread lian
Repository: spark Updated Branches: refs/heads/master 2eaabfa41 -> d063898be [SPARK-16134][SQL] optimizer rules for typed filter ## What changes were proposed in this pull request? This PR adds 3 optimizer rules for typed filter: 1. push typed filter down through `SerializeFromObject` and el

spark git commit: [SPARK-16134][SQL] optimizer rules for typed filter

2016-06-29 Thread lian
Repository: spark Updated Branches: refs/heads/branch-2.0 011befd20 -> 8da431473 [SPARK-16134][SQL] optimizer rules for typed filter ## What changes were proposed in this pull request? This PR adds 3 optimizer rules for typed filter: 1. push typed filter down through `SerializeFromObject` an

spark git commit: Revert "[SPARK-16134][SQL] optimizer rules for typed filter"

2016-06-29 Thread lian
Repository: spark Updated Branches: refs/heads/branch-2.0 8da431473 -> e1bdf1e02 Revert "[SPARK-16134][SQL] optimizer rules for typed filter" This reverts commit 8da4314735ed55f259642e2977d8d7bf2212474f. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apa

spark git commit: [SPARK-16267][TEST] Replace deprecated `CREATE TEMPORARY TABLE ... USING` from testsuites.

2016-06-29 Thread rxin
Repository: spark Updated Branches: refs/heads/master d063898be -> 831a04f5d [SPARK-16267][TEST] Replace deprecated `CREATE TEMPORARY TABLE ... USING` from testsuites. ## What changes were proposed in this pull request? After SPARK-15674, `DDLStrategy` prints out the following deprecation me

spark git commit: [SPARK-16267][TEST] Replace deprecated `CREATE TEMPORARY TABLE ... USING` from testsuites.

2016-06-29 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-2.0 e1bdf1e02 -> b52bd8070 [SPARK-16267][TEST] Replace deprecated `CREATE TEMPORARY TABLE ... USING` from testsuites. ## What changes were proposed in this pull request? After SPARK-15674, `DDLStrategy` prints out the following deprecatio

spark git commit: [SPARK-16274][SQL] Implement xpath_boolean

2016-06-29 Thread wenchen
Repository: spark Updated Branches: refs/heads/master 831a04f5d -> d3af6731f [SPARK-16274][SQL] Implement xpath_boolean ## What changes were proposed in this pull request? This patch implements xpath_boolean expression for Spark SQL, a xpath function that returns true or false. The implementa

spark git commit: [SPARK-16294][SQL] Labelling support for the include_example Jekyll plugin

2016-06-29 Thread meng
Repository: spark Updated Branches: refs/heads/master d3af6731f -> bde1d6a61 [SPARK-16294][SQL] Labelling support for the include_example Jekyll plugin ## What changes were proposed in this pull request? This PR adds labelling support for the `include_example` Jekyll plugin, so that we may s

spark git commit: [SPARK-16294][SQL] Labelling support for the include_example Jekyll plugin

2016-06-29 Thread meng
Repository: spark Updated Branches: refs/heads/branch-2.0 b52bd8070 -> a54852350 [SPARK-16294][SQL] Labelling support for the include_example Jekyll plugin ## What changes were proposed in this pull request? This PR adds labelling support for the `include_example` Jekyll plugin, so that we m

[2/2] spark git commit: [SPARK-12177][STREAMING][KAFKA] Update KafkaDStreams to new Kafka 0.10 Consumer API

2016-06-29 Thread tdas
[SPARK-12177][STREAMING][KAFKA] Update KafkaDStreams to new Kafka 0.10 Consumer API ## What changes were proposed in this pull request? New Kafka consumer api for the released 0.10 version of Kafka ## How was this patch tested? Unit tests, manual tests Author: cody koeninger Closes #11863 f

[1/2] spark git commit: [SPARK-12177][STREAMING][KAFKA] Update KafkaDStreams to new Kafka 0.10 Consumer API

2016-06-29 Thread tdas
Repository: spark Updated Branches: refs/heads/master bde1d6a61 -> dedbceec1 http://git-wip-us.apache.org/repos/asf/spark/blob/dedbceec/external/kafka-0-10/src/test/java/org/apache/spark/streaming/kafka010/JavaConsumerStrategySuite.java --

[2/2] spark git commit: [SPARK-12177][STREAMING][KAFKA] Update KafkaDStreams to new Kafka 0.10 Consumer API

2016-06-29 Thread tdas
[SPARK-12177][STREAMING][KAFKA] Update KafkaDStreams to new Kafka 0.10 Consumer API ## What changes were proposed in this pull request? New Kafka consumer api for the released 0.10 version of Kafka ## How was this patch tested? Unit tests, manual tests Author: cody koeninger Closes #11863 f

[1/2] spark git commit: [SPARK-12177][STREAMING][KAFKA] Update KafkaDStreams to new Kafka 0.10 Consumer API

2016-06-29 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-2.0 a54852350 -> 3134f116a http://git-wip-us.apache.org/repos/asf/spark/blob/3134f116/external/kafka-0-10/src/test/java/org/apache/spark/streaming/kafka010/JavaConsumerStrategySuite.java --

spark git commit: [SPARK-16256][DOCS] Minor fixes on the Structured Streaming Programming Guide

2016-06-29 Thread tdas
Repository: spark Updated Branches: refs/heads/master dedbceec1 -> 2c3d96134 [SPARK-16256][DOCS] Minor fixes on the Structured Streaming Programming Guide Author: Tathagata Das Closes #13978 from tdas/SPARK-16256-1. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://

spark git commit: [SPARK-16256][DOCS] Minor fixes on the Structured Streaming Programming Guide

2016-06-29 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-2.0 3134f116a -> c8a7c2305 [SPARK-16256][DOCS] Minor fixes on the Structured Streaming Programming Guide Author: Tathagata Das Closes #13978 from tdas/SPARK-16256-1. (cherry picked from commit 2c3d96134dcc0428983eea087db7e91072215aea) Si

spark git commit: [SPARK-16257][BUILD] Update spark_ec2.py to support Spark 1.6.2 and 1.6.3.

2016-06-29 Thread srowen
Repository: spark Updated Branches: refs/heads/branch-1.6 1ac830aca -> ccc7fa357 [SPARK-16257][BUILD] Update spark_ec2.py to support Spark 1.6.2 and 1.6.3. ## What changes were proposed in this pull request? - Adds 1.6.2 and 1.6.3 as supported Spark versions within the bundled spark-ec2 scri