spark git commit: [SPARK-11522][SQL] input_file_name() returns "" for external tables

2015-11-16 Thread yhuai
Repository: spark Updated Branches: refs/heads/branch-1.6 a0f9cd77a -> c37ed52ec [SPARK-11522][SQL] input_file_name() returns "" for external tables When computing partition for non-parquet relation, `HadoopRDD.compute` is used. but it does not set the thread local variable `inputFileName`

spark git commit: [SPARK-11044][SQL] Parquet writer version fixed as version1

2015-11-16 Thread lian
Repository: spark Updated Branches: refs/heads/master 42de5253f -> 7f8eb3bf6 [SPARK-11044][SQL] Parquet writer version fixed as version1 https://issues.apache.org/jira/browse/SPARK-11044 Spark writes a parquet file only with writer version1 ignoring the writer version given by user. So, in

spark git commit: [SPARK-11752] [SQL] fix timezone problem for DateTimeUtils.getSeconds

2015-11-16 Thread davies
Repository: spark Updated Branches: refs/heads/branch-1.6 c37ed52ec -> 949c9b7c6 [SPARK-11752] [SQL] fix timezone problem for DateTimeUtils.getSeconds code snippet to reproduce it: ``` TimeZone.setDefault(TimeZone.getTimeZone("Asia/Shanghai")) val t = Timestamp.valueOf("1900-06-11

spark git commit: [SPARK-11752] [SQL] fix timezone problem for DateTimeUtils.getSeconds

2015-11-16 Thread davies
Repository: spark Updated Branches: refs/heads/master 0e79604ae -> 06f1fdba6 [SPARK-11752] [SQL] fix timezone problem for DateTimeUtils.getSeconds code snippet to reproduce it: ``` TimeZone.setDefault(TimeZone.getTimeZone("Asia/Shanghai")) val t = Timestamp.valueOf("1900-06-11 12:14:50.789")

spark git commit: [SPARK-11692][SQL] Support for Parquet logical types, JSON and BSON (embedded types)

2015-11-16 Thread lian
Repository: spark Updated Branches: refs/heads/master 7f8eb3bf6 -> e388b39d1 [SPARK-11692][SQL] Support for Parquet logical types, JSON and BSON (embedded types) Parquet supports some JSON and BSON datatypes. They are represented as binary for BSON and string (UTF-8) for JSON internally.

spark git commit: [SPARK-10181][SQL] Do kerberos login for credentials during hive client initialization

2015-11-16 Thread yhuai
Repository: spark Updated Branches: refs/heads/branch-1.5 bf79a171e -> 51fc152b7 [SPARK-10181][SQL] Do kerberos login for credentials during hive client initialization On driver process start up, UserGroupInformation.loginUserFromKeytab is called with the principal and keytab passed in, and

spark git commit: [SPARK-11044][SQL] Parquet writer version fixed as version1

2015-11-16 Thread lian
Repository: spark Updated Branches: refs/heads/branch-1.6 1887fa228 -> f14fb291d [SPARK-11044][SQL] Parquet writer version fixed as version1 https://issues.apache.org/jira/browse/SPARK-11044 Spark writes a parquet file only with writer version1 ignoring the writer version given by user.

spark git commit: [SPARK-11754][SQL] consolidate `ExpressionEncoder.tuple` and `Encoders.tuple`

2015-11-16 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.6 38fe092ff -> fbe65c592 [SPARK-11754][SQL] consolidate `ExpressionEncoder.tuple` and `Encoders.tuple` These 2 are very similar, we can consolidate them into one. Also add tests for it and fix a bug. Author: Wenchen Fan

spark git commit: [SPARK-11390][SQL] Query plan with/without filterPushdown indistinguishable

2015-11-16 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.6 fbe65c592 -> 90d71bff0 [SPARK-11390][SQL] Query plan with/without filterPushdown indistinguishable …ishable Propagate pushed filters to PhyicalRDD in DataSourceStrategy.apply Author: Zee Chen Closes #9679 from

spark git commit: [SPARK-11390][SQL] Query plan with/without filterPushdown indistinguishable

2015-11-16 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master b1a966262 -> 985b38dd2 [SPARK-11390][SQL] Query plan with/without filterPushdown indistinguishable …ishable Propagate pushed filters to PhyicalRDD in DataSourceStrategy.apply Author: Zee Chen Closes #9679 from

spark git commit: Revert "[SPARK-11271][SPARK-11016][CORE] Use Spark BitSet instead of RoaringBitmap to reduce memory usage"

2015-11-16 Thread davies
Repository: spark Updated Branches: refs/heads/branch-1.6 90d71bff0 -> 64439f7d6 Revert "[SPARK-11271][SPARK-11016][CORE] Use Spark BitSet instead of RoaringBitmap to reduce memory usage" This reverts commit e209fa271ae57dc8849f8b1241bf1ea7d6d3d62c. Project:

spark git commit: Revert "[SPARK-11271][SPARK-11016][CORE] Use Spark BitSet instead of RoaringBitmap to reduce memory usage"

2015-11-16 Thread davies
Repository: spark Updated Branches: refs/heads/master 985b38dd2 -> 3c025087b Revert "[SPARK-11271][SPARK-11016][CORE] Use Spark BitSet instead of RoaringBitmap to reduce memory usage" This reverts commit e209fa271ae57dc8849f8b1241bf1ea7d6d3d62c. Project:

spark git commit: [SPARK-11731][STREAMING] Enable batching on Driver WriteAheadLog by default

2015-11-16 Thread tdas
Repository: spark Updated Branches: refs/heads/master b0c3fd34e -> de5e531d3 [SPARK-11731][STREAMING] Enable batching on Driver WriteAheadLog by default Using batching on the driver for the WriteAheadLog should be an improvement for all environments and use cases. Users will be able to scale

spark git commit: [SPARK-11731][STREAMING] Enable batching on Driver WriteAheadLog by default

2015-11-16 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-1.6 f14fb291d -> 38673d7e6 [SPARK-11731][STREAMING] Enable batching on Driver WriteAheadLog by default Using batching on the driver for the WriteAheadLog should be an improvement for all environments and use cases. Users will be able to

spark git commit: [SPARK-11718][YARN][CORE] Fix explicitly killed executor dies silently issue

2015-11-16 Thread vanzin
Repository: spark Updated Branches: refs/heads/master ace0db471 -> 24477d270 [SPARK-11718][YARN][CORE] Fix explicitly killed executor dies silently issue Currently if dynamic allocation is enabled, explicitly killing executor will not get response, so the executor metadata is wrong in driver

spark git commit: [SPARK-11718][YARN][CORE] Fix explicitly killed executor dies silently issue

2015-11-16 Thread vanzin
Repository: spark Updated Branches: refs/heads/branch-1.6 c83177d30 -> 38fe092ff [SPARK-11718][YARN][CORE] Fix explicitly killed executor dies silently issue Currently if dynamic allocation is enabled, explicitly killing executor will not get response, so the executor metadata is wrong in

spark git commit: [SPARK-11754][SQL] consolidate `ExpressionEncoder.tuple` and `Encoders.tuple`

2015-11-16 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master 24477d270 -> b1a966262 [SPARK-11754][SQL] consolidate `ExpressionEncoder.tuple` and `Encoders.tuple` These 2 are very similar, we can consolidate them into one. Also add tests for it and fix a bug. Author: Wenchen Fan

spark git commit: [SPARK-6328][PYTHON] Python API for StreamingListener

2015-11-16 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-1.6 38673d7e6 -> c83177d30 [SPARK-6328][PYTHON] Python API for StreamingListener Author: Daniel Jalova Closes #9186 from djalova/SPARK-6328. (cherry picked from commit ace0db47141ffd457c2091751038fc291f6d5a8b)

spark git commit: [SPARK-6328][PYTHON] Python API for StreamingListener

2015-11-16 Thread tdas
Repository: spark Updated Branches: refs/heads/master de5e531d3 -> ace0db471 [SPARK-6328][PYTHON] Python API for StreamingListener Author: Daniel Jalova Closes #9186 from djalova/SPARK-6328. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit:

spark git commit: [SPARK-11625][SQL] add java test for typed aggregate

2015-11-16 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master 75ee12f09 -> fd14936be [SPARK-11625][SQL] add java test for typed aggregate Author: Wenchen Fan Closes #9591 from cloud-fan/agg-test. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit:

spark git commit: [SPARK-11742][STREAMING] Add the failure info to the batch lists

2015-11-16 Thread tdas
Repository: spark Updated Branches: refs/heads/master 3c025087b -> bcea0bfda [SPARK-11742][STREAMING] Add the failure info to the batch lists https://cloud.githubusercontent.com/assets/1000778/11162322/9b88e204-8a51-11e5-8c57-a44889cab713.png;> Author: Shixiong Zhu

spark git commit: [SPARK-11742][STREAMING] Add the failure info to the batch lists

2015-11-16 Thread tdas
Repository: spark Updated Branches: refs/heads/branch-1.6 64439f7d6 -> 3bd72eafc [SPARK-11742][STREAMING] Add the failure info to the batch lists https://cloud.githubusercontent.com/assets/1000778/11162322/9b88e204-8a51-11e5-8c57-a44889cab713.png;> Author: Shixiong Zhu

spark git commit: [SPARK-11553][SQL] Primitive Row accessors should not convert null to default value

2015-11-16 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.6 3bd72eafc -> 6c8e0c0ff [SPARK-11553][SQL] Primitive Row accessors should not convert null to default value Invocation of getters for type extending AnyVal returns default value (if field value is null) instead of throwing NPE. Please

spark git commit: [SPARKR][HOTFIX] Disable flaky SparkR package build test

2015-11-16 Thread andrewor14
Repository: spark Updated Branches: refs/heads/branch-1.6 4f8c7e18f -> bb044ec22 [SPARKR][HOTFIX] Disable flaky SparkR package build test See https://github.com/apache/spark/pull/9390#issuecomment-157160063 and https://gist.github.com/shivaram/3a2fecce60768a603dac for more information

spark git commit: [SPARKR][HOTFIX] Disable flaky SparkR package build test

2015-11-16 Thread andrewor14
Repository: spark Updated Branches: refs/heads/master fd14936be -> ea6f53e48 [SPARKR][HOTFIX] Disable flaky SparkR package build test See https://github.com/apache/spark/pull/9390#issuecomment-157160063 and https://gist.github.com/shivaram/3a2fecce60768a603dac for more information Author:

spark git commit: [SPARK-8658][SQL] AttributeReference's equals method compares all the members

2015-11-16 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.6 6c8e0c0ff -> e042780cd [SPARK-8658][SQL] AttributeReference's equals method compares all the members This fix is to change the equals method to check all of the specified fields for equality of AttributeReference. Author: gatorsmile

spark git commit: [SPARK-8658][SQL] AttributeReference's equals method compares all the members

2015-11-16 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master 31296628a -> 75ee12f09 [SPARK-8658][SQL] AttributeReference's equals method compares all the members This fix is to change the equals method to check all of the specified fields for equality of AttributeReference. Author: gatorsmile

spark git commit: [SPARK-11625][SQL] add java test for typed aggregate

2015-11-16 Thread marmbrus
Repository: spark Updated Branches: refs/heads/branch-1.6 e042780cd -> 4f8c7e18f [SPARK-11625][SQL] add java test for typed aggregate Author: Wenchen Fan Closes #9591 from cloud-fan/agg-test. (cherry picked from commit fd14936be7beff543dbbcf270f2f9749f7a803c4)

spark git commit: [SPARK-11480][CORE][WEBUI] Wrong callsite is displayed when using AsyncRDDActions#takeAsync

2015-11-16 Thread andrewor14
Repository: spark Updated Branches: refs/heads/branch-1.6 bb044ec22 -> e4abfe932 [SPARK-11480][CORE][WEBUI] Wrong callsite is displayed when using AsyncRDDActions#takeAsync When we call AsyncRDDActions#takeAsync, actually another DAGScheduler#runJob is called from another thread so we

spark git commit: [SPARK-11480][CORE][WEBUI] Wrong callsite is displayed when using AsyncRDDActions#takeAsync

2015-11-16 Thread andrewor14
Repository: spark Updated Branches: refs/heads/master ea6f53e48 -> 30f3cfda1 [SPARK-11480][CORE][WEBUI] Wrong callsite is displayed when using AsyncRDDActions#takeAsync When we call AsyncRDDActions#takeAsync, actually another DAGScheduler#runJob is called from another thread so we cannot

spark git commit: [SPARK-11710] Document new memory management model

2015-11-16 Thread andrewor14
Repository: spark Updated Branches: refs/heads/master 30f3cfda1 -> 33a0ec937 [SPARK-11710] Document new memory management model Author: Andrew Or Closes #9676 from andrewor14/memory-management-docs. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit:

spark git commit: [EXAMPLE][MINOR] Add missing awaitTermination in click stream example

2015-11-16 Thread andrewor14
Repository: spark Updated Branches: refs/heads/master 33a0ec937 -> bd10eb81c [EXAMPLE][MINOR] Add missing awaitTermination in click stream example Author: jerryshao Closes #9730 from jerryshao/clickstream-fix. Project:

spark git commit: [SPARK-11768][SPARK-9196][SQL] Support now function in SQL (alias for current_timestamp).

2015-11-16 Thread rxin
Repository: spark Updated Branches: refs/heads/master 540bf58f1 -> fbad920db [SPARK-11768][SPARK-9196][SQL] Support now function in SQL (alias for current_timestamp). This patch adds an alias for current_timestamp (now function). Also fixes SPARK-9196 to re-enable the test case for

spark git commit: [SPARK-11768][SPARK-9196][SQL] Support now function in SQL (alias for current_timestamp).

2015-11-16 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.6 5a6f40459 -> 07ac8e950 [SPARK-11768][SPARK-9196][SQL] Support now function in SQL (alias for current_timestamp). This patch adds an alias for current_timestamp (now function). Also fixes SPARK-9196 to re-enable the test case for

spark git commit: [SPARK-11553][SQL] Primitive Row accessors should not convert null to default value

2015-11-16 Thread marmbrus
Repository: spark Updated Branches: refs/heads/master bcea0bfda -> 31296628a [SPARK-11553][SQL] Primitive Row accessors should not convert null to default value Invocation of getters for type extending AnyVal returns default value (if field value is null) instead of throwing NPE. Please

spark git commit: [SPARK-11612][ML] Pipeline and PipelineModel persistence

2015-11-16 Thread jkbradley
Repository: spark Updated Branches: refs/heads/branch-1.6 32a69e4c1 -> 505eceef3 [SPARK-11612][ML] Pipeline and PipelineModel persistence Pipeline and PipelineModel extend Readable and Writable. Persistence succeeds only when all stages are Writable. Note: This PR reinstates tests for

spark git commit: [SPARK-11612][ML] Pipeline and PipelineModel persistence

2015-11-16 Thread jkbradley
Repository: spark Updated Branches: refs/heads/master bd10eb81c -> 1c5475f14 [SPARK-11612][ML] Pipeline and PipelineModel persistence Pipeline and PipelineModel extend Readable and Writable. Persistence succeeds only when all stages are Writable. Note: This PR reinstates tests for other

spark git commit: [SPARK-11617][NETWORK] Fix leak in TransportFrameDecoder.

2015-11-16 Thread vanzin
Repository: spark Updated Branches: refs/heads/branch-1.6 505eceef3 -> e12ecfa36 [SPARK-11617][NETWORK] Fix leak in TransportFrameDecoder. The code was using the wrong API to add data to the internal composite buffer, causing buffers to leak in certain situations. Use the right API and

[1/2] spark git commit: Preparing Spark release v1.6.0-preview

2015-11-16 Thread pwendell
Repository: spark Updated Branches: refs/heads/branch-1.6 e12ecfa36 -> 5a6f40459 Preparing Spark release v1.6.0-preview Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/31db3610 Tree:

[2/2] spark git commit: Preparing development version 1.6.0-SNAPSHOT

2015-11-16 Thread pwendell
Preparing development version 1.6.0-SNAPSHOT Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/5a6f4045 Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/5a6f4045 Diff:

Git Push Summary

2015-11-16 Thread pwendell
Repository: spark Updated Tags: refs/tags/v1.6.0-preview [created] 31db36100 - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org

spark git commit: [SPARK-11617][NETWORK] Fix leak in TransportFrameDecoder.

2015-11-16 Thread vanzin
Repository: spark Updated Branches: refs/heads/master 1c5475f14 -> 540bf58f1 [SPARK-11617][NETWORK] Fix leak in TransportFrameDecoder. The code was using the wrong API to add data to the internal composite buffer, causing buffers to leak in certain situations. Use the right API and enhance

spark git commit: [SPARK-11694][FOLLOW-UP] Clean up imports, use a common function for metadata and add a test for FIXED_LEN_BYTE_ARRAY

2015-11-16 Thread lian
Repository: spark Updated Branches: refs/heads/master fbad920db -> 75d202073 [SPARK-11694][FOLLOW-UP] Clean up imports, use a common function for metadata and add a test for FIXED_LEN_BYTE_ARRAY As discussed https://github.com/apache/spark/pull/9660

spark git commit: [SPARK-11745][SQL] Enable more JSON parsing options

2015-11-16 Thread rxin
Repository: spark Updated Branches: refs/heads/master fd50fa4c3 -> 42de5253f [SPARK-11745][SQL] Enable more JSON parsing options This patch adds the following options to the JSON data source, for dealing with non-standard JSON files: * `allowComments` (default `false`): ignores Java/C++

spark git commit: [SPARK-11745][SQL] Enable more JSON parsing options

2015-11-16 Thread rxin
Repository: spark Updated Branches: refs/heads/branch-1.6 053c63ecf -> a0f9cd77a [SPARK-11745][SQL] Enable more JSON parsing options This patch adds the following options to the JSON data source, for dealing with non-standard JSON files: * `allowComments` (default `false`): ignores Java/C++

spark git commit: [SPARK-11447][SQL] change NullType to StringType during binaryComparison between NullType and StringType

2015-11-16 Thread yhuai
Repository: spark Updated Branches: refs/heads/branch-1.6 07ac8e950 -> 113410c12 [SPARK-11447][SQL] change NullType to StringType during binaryComparison between NullType and StringType During executing PromoteStrings rule, if one side of binaryComparison is StringType and the other side is

spark git commit: [MINOR] [SQL] Fix randomly generated ArrayData in RowEncoderSuite

2015-11-16 Thread davies
Repository: spark Updated Branches: refs/heads/master e01865af0 -> d79d8b08f [MINOR] [SQL] Fix randomly generated ArrayData in RowEncoderSuite The randomly generated ArrayData used for the UDT `ExamplePoint` in `RowEncoderSuite` sometimes doesn't have enough elements. In this case, this

spark git commit: [MINOR] [SQL] Fix randomly generated ArrayData in RowEncoderSuite

2015-11-16 Thread davies
Repository: spark Updated Branches: refs/heads/branch-1.6 113410c12 -> 2ae1fa074 [MINOR] [SQL] Fix randomly generated ArrayData in RowEncoderSuite The randomly generated ArrayData used for the UDT `ExamplePoint` in `RowEncoderSuite` sometimes doesn't have enough elements. In this case, this

spark git commit: [SPARK-11447][SQL] change NullType to StringType during binaryComparison between NullType and StringType

2015-11-16 Thread yhuai
Repository: spark Updated Branches: refs/heads/master 75d202073 -> e01865af0 [SPARK-11447][SQL] change NullType to StringType during binaryComparison between NullType and StringType During executing PromoteStrings rule, if one side of binaryComparison is StringType and the other side is not