spark git commit: [SPARK-19386][SPARKR][FOLLOWUP] fix error in vignettes

2017-02-03 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 48aafeda7 -> 050c20cc9 [SPARK-19386][SPARKR][FOLLOWUP] fix error in vignettes ## What changes were proposed in this pull request? Current version has error in vignettes: ``` model <- spark.bisectingKmeans(df, Sepal_Length ~ Sepal_Width, k

spark git commit: [SPARK-10063] Follow-up: remove dead code related to an old output committer.

2017-02-03 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 050c20cc9 -> 22d4aae8b [SPARK-10063] Follow-up: remove dead code related to an old output committer. ## What changes were proposed in this pull request? DirectParquetOutputCommitter was removed from Spark as it was deemed unsafe to use.

spark git commit: [SPARK-19296][SQL] Deduplicate url and table in JdbcUtils

2017-02-01 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 04ee8cf63 -> 5ed397baa [SPARK-19296][SQL] Deduplicate url and table in JdbcUtils ## What changes were proposed in this pull request? This PR deduplicates arguments, `url` and `table` in `JdbcUtils` with `JDBCOptions`. It avoids to use

spark git commit: [SPARK-19279][SQL][FOLLOW-UP] Infer Schema for Hive Serde Tables

2017-02-08 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 0077bfcb9 -> 4d4d0de7f [SPARK-19279][SQL][FOLLOW-UP] Infer Schema for Hive Serde Tables ### What changes were proposed in this pull request? `table.schema` is always not empty for partitioned tables, because `table.schema` also contains

spark git commit: [SPARK-18872][SQL][TESTS] New test cases for EXISTS subquery (Aggregate, Having, Orderby, Limit)

2017-02-08 Thread lixiao
Repository: spark Updated Branches: refs/heads/master c618ccdbe -> 64cae22f7 [SPARK-18872][SQL][TESTS] New test cases for EXISTS subquery (Aggregate, Having, Orderby, Limit) ## What changes were proposed in this pull request? This PR adds the second set of tests for EXISTS subquery. File

spark git commit: [SPARK-19472][SQL] Parser should not mistake CASE WHEN(...) for a function call

2017-02-06 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 7c23bd49e -> f55bd4c73 [SPARK-19472][SQL] Parser should not mistake CASE WHEN(...) for a function call ## What changes were proposed in this pull request? The SQL parser can mistake a `WHEN (...)` used in `CASE` for a function call.

spark git commit: [SPARK-19472][SQL] Parser should not mistake CASE WHEN(...) for a function call

2017-02-06 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.0 b41294bb8 -> 8bf642226 [SPARK-19472][SQL] Parser should not mistake CASE WHEN(...) for a function call ## What changes were proposed in this pull request? The SQL parser can mistake a `WHEN (...)` used in `CASE` for a function call.

spark git commit: [SPARK-19229][SQL] Disallow Creating Hive Source Tables when Hive Support is Not Enabled

2017-01-22 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 74e65cb74 -> 772035e77 [SPARK-19229][SQL] Disallow Creating Hive Source Tables when Hive Support is Not Enabled ### What changes were proposed in this pull request? It is weird to create Hive source tables when using InMemoryCatalog. We

spark git commit: [SPARK-19218][SQL] Fix SET command to show a result correctly and in a sorted order

2017-01-23 Thread lixiao
Repository: spark Updated Branches: refs/heads/master f067acefa -> c4a6519c4 [SPARK-19218][SQL] Fix SET command to show a result correctly and in a sorted order ## What changes were proposed in this pull request? This PR aims to fix the following two things. 1. `sql("SET -v").collect()` or

spark git commit: [SPARK-9435][SQL] Reuse function in Java UDF to correctly support expressions that require equality comparison between ScalaUDF

2017-01-23 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 ed5d1e725 -> 4a2be0902 [SPARK-9435][SQL] Reuse function in Java UDF to correctly support expressions that require equality comparison between ScalaUDF ## What changes were proposed in this pull request? Currently, running the codes

spark git commit: [SPARK-9435][SQL] Reuse function in Java UDF to correctly support expressions that require equality comparison between ScalaUDF

2017-01-23 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 3bdf3ee86 -> e576c1ed7 [SPARK-9435][SQL] Reuse function in Java UDF to correctly support expressions that require equality comparison between ScalaUDF ## What changes were proposed in this pull request? Currently, running the codes in

spark git commit: [SPARK-19290][SQL] add a new extending interface in Analyzer for post-hoc resolution

2017-01-23 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 49f5b0ae4 -> fcfd5d0bb [SPARK-19290][SQL] add a new extending interface in Analyzer for post-hoc resolution ## What changes were proposed in this pull request? To implement DDL commands, we added several analyzer rules in sql/hive module

spark git commit: [SPARK-16046][DOCS] Aggregations in the Spark SQL programming guide

2017-01-24 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 c13378796 -> e2f773923 [SPARK-16046][DOCS] Aggregations in the Spark SQL programming guide ## What changes were proposed in this pull request? - A separate subsection for Aggregations under “Getting Started” in the Spark SQL

spark git commit: [SPARK-16046][DOCS] Aggregations in the Spark SQL programming guide

2017-01-24 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 40a4cfc7c -> 3fdce8143 [SPARK-16046][DOCS] Aggregations in the Spark SQL programming guide ## What changes were proposed in this pull request? - A separate subsection for Aggregations under “Getting Started” in the Spark SQL

spark git commit: [SPARK-19338][SQL] Add UDF names in explain

2017-01-26 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 2969fb437 -> 9f523d319 [SPARK-19338][SQL] Add UDF names in explain ## What changes were proposed in this pull request? This pr added a variable for a UDF name in `ScalaUDF`. Then, if the variable filled, `DataFrame#explain` prints the

spark git commit: [SPARK-19338][SQL] Add UDF names in explain

2017-01-26 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 0d7e38524 -> b12a76a41 [SPARK-19338][SQL] Add UDF names in explain ## What changes were proposed in this pull request? This pr added a variable for a UDF name in `ScalaUDF`. Then, if the variable filled, `DataFrame#explain` prints the

spark git commit: [SPARK-19396][DOC] JDBC Options are Case In-sensitive

2017-01-30 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 9a49f9afa -> 445438c9f [SPARK-19396][DOC] JDBC Options are Case In-sensitive ### What changes were proposed in this pull request? The case are not sensitive in JDBC options, after the PR https://github.com/apache/spark/pull/15884 is

spark git commit: [SPARK-19396][DOC] JDBC Options are Case In-sensitive

2017-01-30 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 06fbc3554 -> c0eda7e87 [SPARK-19396][DOC] JDBC Options are Case In-sensitive ### What changes were proposed in this pull request? The case are not sensitive in JDBC options, after the PR https://github.com/apache/spark/pull/15884 is

spark git commit: [SPARK-19406][SQL] Fix function to_json to respect user-provided options

2017-01-30 Thread lixiao
Repository: spark Updated Branches: refs/heads/master c0eda7e87 -> f9156d295 [SPARK-19406][SQL] Fix function to_json to respect user-provided options ### What changes were proposed in this pull request? Currently, the function `to_json` allows users to provide options for generating JSON.

spark git commit: [SPARK-19406][SQL] Fix function to_json to respect user-provided options

2017-01-30 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 445438c9f -> 07a1788ee [SPARK-19406][SQL] Fix function to_json to respect user-provided options ### What changes were proposed in this pull request? Currently, the function `to_json` allows users to provide options for generating

spark git commit: [SPARK-19359][SQL] clear useless path after rename a partition with upper-case by HiveExternalCatalog

2017-01-27 Thread lixiao
Repository: spark Updated Branches: refs/heads/master bb1a1fe05 -> 1b5ee2003 [SPARK-19359][SQL] clear useless path after rename a partition with upper-case by HiveExternalCatalog ## What changes were proposed in this pull request? Hive metastore is not case preserving and keep partition

spark git commit: [SPARK-19359][SQL] Revert Clear useless path after rename a partition with upper-case by HiveExternalCatalog

2017-01-28 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 42ad93b2c -> cfcfc92f7 [SPARK-19359][SQL] Revert Clear useless path after rename a partition with upper-case by HiveExternalCatalog ### What changes were proposed in this pull request? This PR is to revert the changes made in

spark git commit: [SPARK-19385][SQL] During canonicalization, `NOT(...(l, r))` should not expect such cases that l.hashcode > r.hashcode

2017-01-29 Thread lixiao
Repository: spark Updated Branches: refs/heads/master e2e7b12ce -> ade075aed [SPARK-19385][SQL] During canonicalization, `NOT(...(l, r))` should not expect such cases that l.hashcode > r.hashcode ## What changes were proposed in this pull request? During canonicalization, `NOT(...(l, r))`

spark git commit: [SPARK-18872][SQL][TESTS] New test cases for EXISTS subquery

2017-01-29 Thread lixiao
Repository: spark Updated Branches: refs/heads/master f7c07db85 -> e2e7b12ce [SPARK-18872][SQL][TESTS] New test cases for EXISTS subquery ## What changes were proposed in this pull request? This PR adds the first set of tests for EXISTS subquery. File name| Brief

spark git commit: [SPARK-19311][SQL] fix UDT hierarchy issue

2017-01-25 Thread lixiao
Repository: spark Updated Branches: refs/heads/master f1ddca5fc -> f6480b146 [SPARK-19311][SQL] fix UDT hierarchy issue ## What changes were proposed in this pull request? acceptType() in UDT will no only accept the same type but also all base types ## How was this patch tested? Manual test

spark git commit: [TESTS][SQL] Setup testdata at the beginning for tests to run independently

2017-01-25 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 256a3a801 -> 9effc2cdc [TESTS][SQL] Setup testdata at the beginning for tests to run independently ## What changes were proposed in this pull request? In CachedTableSuite, we are not setting up the test data at the beginning. Some tests

spark git commit: [SPARK-14536][SQL] fix to handle null value in array type column for postgres.

2017-01-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 54268b42d -> f174cdc74 [SPARK-14536][SQL] fix to handle null value in array type column for postgres. ## What changes were proposed in this pull request? JDBC read is failing with NPE due to missing null value check for array data

spark git commit: [SPARK-19152][SQL][FOLLOWUP] simplify CreateHiveTableAsSelectCommand

2017-01-28 Thread lixiao
Repository: spark Updated Branches: refs/heads/master cfcfc92f7 -> f7c07db85 [SPARK-19152][SQL][FOLLOWUP] simplify CreateHiveTableAsSelectCommand ## What changes were proposed in this pull request? After https://github.com/apache/spark/pull/16552 , `CreateHiveTableAsSelectCommand` becomes

spark git commit: [SPARK-18871][SQL][TESTS] New test cases for IN/NOT IN subquery 3rd batch

2017-02-16 Thread lixiao
Repository: spark Updated Branches: refs/heads/master f041e55ee -> 3871d94a6 [SPARK-18871][SQL][TESTS] New test cases for IN/NOT IN subquery 3rd batch ## What changes were proposed in this pull request? This is 3ird batch of test case for IN/NOT IN subquery. In this PR, it has these test

spark git commit: [SPARK-18871][SQL][TESTS] New test cases for IN/NOT IN subquery 4th batch

2017-02-15 Thread lixiao
Repository: spark Updated Branches: refs/heads/master fc02ef95c -> 8487902a9 [SPARK-18871][SQL][TESTS] New test cases for IN/NOT IN subquery 4th batch ## What changes were proposed in this pull request? This is 4th batch of test case for IN/NOT IN subquery. In this PR, it has these test

spark git commit: [SPARK-15453][SQL][FOLLOW-UP] FileSourceScanExec to extract `outputOrdering` information

2017-02-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master d0ecca607 -> ead4ba0eb [SPARK-15453][SQL][FOLLOW-UP] FileSourceScanExec to extract `outputOrdering` information ### What changes were proposed in this pull request? `outputOrdering` is also dependent on whether the bucket has more than

spark git commit: [SPARK-19669][SQL] Open up visibility for sharedState, sessionState, and a few other functions

2017-02-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master ead4ba0eb -> 0733a54a4 [SPARK-19669][SQL] Open up visibility for sharedState, sessionState, and a few other functions ## What changes were proposed in this pull request? To ease debugging, most of Spark SQL internals have public level

spark git commit: [SPARK-19669][HOTFIX][SQL] sessionState access privileges compiled failed in TestSQLContext

2017-02-20 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 0733a54a4 -> 73f065569 [SPARK-19669][HOTFIX][SQL] sessionState access privileges compiled failed in TestSQLContext ## What changes were proposed in this pull request? In

spark git commit: [SPARK-19459][SQL][BRANCH-2.1] Support for nested char/varchar fields in ORC

2017-02-23 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 d30238f1b -> 43084b3cc [SPARK-19459][SQL][BRANCH-2.1] Support for nested char/varchar fields in ORC ## What changes were proposed in this pull request? This is a backport of the two following commits:

spark git commit: [MINOR][DOCS] Fixes two problems in the SQL programing guide page

2017-02-25 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 ed9aaa314 -> 97866e198 [MINOR][DOCS] Fixes two problems in the SQL programing guide page ## What changes were proposed in this pull request? Removed duplicated lines in sql python example and found a typo. ## How was this patch

spark git commit: [MINOR][DOCS] Fixes two problems in the SQL programing guide page

2017-02-25 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 8f0511ed4 -> 061bcfb86 [MINOR][DOCS] Fixes two problems in the SQL programing guide page ## What changes were proposed in this pull request? Removed duplicated lines in sql python example and found a typo. ## How was this patch tested?

spark git commit: [SPARK-19598][SQL] Remove the alias parameter in UnresolvedRelation

2017-02-19 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 1487c9af2 -> 65fe902e1 [SPARK-19598][SQL] Remove the alias parameter in UnresolvedRelation ## What changes were proposed in this pull request? Remove the alias parameter in `UnresolvedRelation`, and use `SubqueryAlias` to replace it.

spark git commit: [SPARK-19435][SQL] Type coercion between ArrayTypes

2017-02-13 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 905fdf0c2 -> 9af8f743b [SPARK-19435][SQL] Type coercion between ArrayTypes ## What changes were proposed in this pull request? This PR proposes to support type coercion between `ArrayType`s where the element types are compatible.

spark git commit: [SPARK-19448][SQL] optimize some duplication functions between HiveClientImpl and HiveUtils

2017-02-11 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 0fbecc736 -> 3881f342b [SPARK-19448][SQL] optimize some duplication functions between HiveClientImpl and HiveUtils ## What changes were proposed in this pull request? There are some duplicate functions between `HiveClientImpl` and

spark git commit: [SPARK-19115][SQL] Supporting Create Table Like Location

2017-02-13 Thread lixiao
Repository: spark Updated Branches: refs/heads/master e02ac303c -> 6e45b547c [SPARK-19115][SQL] Supporting Create Table Like Location What changes were proposed in this pull request? Support CREATE [EXTERNAL] TABLE LIKE LOCATION... syntax for Hive serde and datasource tables. In this PR,we

spark git commit: [SPARK-19539][SQL] Block duplicate temp table during creation

2017-02-13 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 6e45b547c -> 1ab97310e [SPARK-19539][SQL] Block duplicate temp table during creation ## What changes were proposed in this pull request? Current `CREATE TEMPORARY TABLE ... ` is deprecated and recommend users to use `CREATE TEMPORARY VIEW

spark git commit: [SPARK-19585][DOC][SQL] Fix the cacheTable and uncacheTable api call in the doc

2017-02-13 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 1ab97310e -> 9b5e460a9 [SPARK-19585][DOC][SQL] Fix the cacheTable and uncacheTable api call in the doc ## What changes were proposed in this pull request?

spark git commit: [SPARK-19585][DOC][SQL] Fix the cacheTable and uncacheTable api call in the doc

2017-02-13 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 7fe3543fd -> c8113b0ee [SPARK-19585][DOC][SQL] Fix the cacheTable and uncacheTable api call in the doc ## What changes were proposed in this pull request?

spark git commit: [SPARK-19436][SQL] Add missing tests for approxQuantile

2017-02-16 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 3b4376876 -> 54a30c8a7 [SPARK-19436][SQL] Add missing tests for approxQuantile ## What changes were proposed in this pull request? 1, check the behavior with illegal `quantiles` and `relativeError` 2, add tests for `relativeError` > 1 3,

spark git commit: [SPARK-19329][SQL] Reading from or writing to a datasource table with a non pre-existing location should succeed

2017-02-15 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 59dc26e37 -> 6a9a85b84 [SPARK-19329][SQL] Reading from or writing to a datasource table with a non pre-existing location should succeed ## What changes were proposed in this pull request? when we insert data into a datasource table use

[1/2] spark git commit: [SPARK-18243][SQL] Port Hive writing to use FileFormat interface

2017-01-17 Thread lixiao
Repository: spark Updated Branches: refs/heads/master e7f982b20 -> 4494cd971 http://git-wip-us.apache.org/repos/asf/spark/blob/4494cd97/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveComparisonTest.scala --

[2/2] spark git commit: [SPARK-18243][SQL] Port Hive writing to use FileFormat interface

2017-01-17 Thread lixiao
[SPARK-18243][SQL] Port Hive writing to use FileFormat interface ## What changes were proposed in this pull request? Inserting data into Hive tables has its own implementation that is distinct from data sources: `InsertIntoHiveTable`, `SparkHiveWriterContainer` and

spark git commit: [SPARK-19265][SQL] make table relation cache general and does not depend on hive

2017-01-19 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 0c9231858 -> 2e6256002 [SPARK-19265][SQL] make table relation cache general and does not depend on hive ## What changes were proposed in this pull request? We have a table relation plan cache in `HiveMetastoreCatalog`, which caches a lot

spark git commit: [SPARK-16968][SQL][BACKPORT-2.0] Add additional options in jdbc when creating a new table

2017-01-18 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.0 ee4e8faff -> 9fc053c30 [SPARK-16968][SQL][BACKPORT-2.0] Add additional options in jdbc when creating a new table ### What changes were proposed in this pull request? This PR is to backport the PRs

spark git commit: [SPARK-19239][PYSPARK] Check parameters whether equals None when specify the column in jdbc API

2017-01-17 Thread lixiao
Repository: spark Updated Branches: refs/heads/master a23debd7b -> 843ec8ec4 [SPARK-19239][PYSPARK] Check parameters whether equals None when specify the column in jdbc API ## What changes were proposed in this pull request? The `jdbc` API do not check the `lowerBound` and `upperBound` when

spark git commit: [SPARK-19292][SQL] filter with partition columns should be case-insensitive on Hive tables

2017-01-19 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 148a84b37 -> 0bf605c2c [SPARK-19292][SQL] filter with partition columns should be case-insensitive on Hive tables ## What changes were proposed in this pull request? When we query a table with a filter on partitioned columns, we will

spark git commit: [SPARK-19271][SQL] Change non-cbo estimation of aggregate

2017-01-19 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 0bf605c2c -> 039ed9fe8 [SPARK-19271][SQL] Change non-cbo estimation of aggregate ## What changes were proposed in this pull request? Change non-cbo estimation behavior of aggregate: - If groupExpression is empty, we can know row count

spark git commit: [SQL] Duplicate test exception in SQLQueryTestSuite due to meta files(.DS_Store) on Mac

2017-02-25 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 89608cf26 -> 68f2142cf [SQL] Duplicate test exception in SQLQueryTestSuite due to meta files(.DS_Store) on Mac ## What changes were proposed in this pull request? After adding the tests for subquery, we now have multiple level of

[2/2] spark git commit: [SPARK-19678][SQL] remove MetastoreRelation

2017-02-28 Thread lixiao
[SPARK-19678][SQL] remove MetastoreRelation ## What changes were proposed in this pull request? `MetastoreRelation` is used to represent table relation for hive tables, and provides some hive related information. We will resolve `SimpleCatalogRelation` to `MetastoreRelation` for hive tables,

[1/2] spark git commit: [SPARK-19678][SQL] remove MetastoreRelation

2017-02-28 Thread lixiao
Repository: spark Updated Branches: refs/heads/master b40546651 -> 7c7fc30b4 http://git-wip-us.apache.org/repos/asf/spark/blob/7c7fc30b/sql/hive/src/test/scala/org/apache/spark/sql/hive/StatisticsSuite.scala -- diff --git

spark git commit: [SPARK-19761][SQL] create InMemoryFileIndex with an empty rootPaths when set PARALLEL_PARTITION_DISCOVERY_THRESHOLD to zero failed

2017-03-01 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 5502a9cf8 -> 8aa560b75 [SPARK-19761][SQL] create InMemoryFileIndex with an empty rootPaths when set PARALLEL_PARTITION_DISCOVERY_THRESHOLD to zero failed ## What changes were proposed in this pull request? If we create a

spark git commit: [SPARK-19766][SQL] Constant alias columns in INNER JOIN should not be folded by FoldablePropagation rule

2017-03-01 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 38e783534 -> 5502a9cf8 [SPARK-19766][SQL] Constant alias columns in INNER JOIN should not be folded by FoldablePropagation rule ## What changes were proposed in this pull request? This PR fixes the code in Optimizer phase where the

spark git commit: [SPARK-19766][SQL] Constant alias columns in INNER JOIN should not be folded by FoldablePropagation rule

2017-03-01 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 f719cccdc -> bbe0d8caa [SPARK-19766][SQL] Constant alias columns in INNER JOIN should not be folded by FoldablePropagation rule ## What changes were proposed in this pull request? This PR fixes the code in Optimizer phase where the

spark git commit: [SPARK-14525][SQL][FOLLOWUP] Clean up JdbcRelationProvider

2016-10-07 Thread lixiao
Repository: spark Updated Branches: refs/heads/master cff560755 -> aa3a6841e [SPARK-14525][SQL][FOLLOWUP] Clean up JdbcRelationProvider ## What changes were proposed in this pull request? This PR proposes cleaning up the confusing part in `createRelation` as discussed in

spark git commit: [SPARK-17750][SQL][BACKPORT-2.0] Fix CREATE VIEW with INTERVAL arithmetic

2016-10-06 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.0 e355ca8e8 -> b1a9c41e8 [SPARK-17750][SQL][BACKPORT-2.0] Fix CREATE VIEW with INTERVAL arithmetic ## What changes were proposed in this pull request? Currently, Spark raises `RuntimeException` when creating a view with timestamp with

spark git commit: [SPARK-17659][SQL] Partitioned View is Not Supported By SHOW CREATE TABLE

2016-11-09 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 f67208369 -> b89c38b2e [SPARK-17659][SQL] Partitioned View is Not Supported By SHOW CREATE TABLE ### What changes were proposed in this pull request? `Partitioned View` is not supported by SPARK SQL. For Hive partitioned view, SHOW

spark git commit: [SPARK-17659][SQL] Partitioned View is Not Supported By SHOW CREATE TABLE

2016-11-09 Thread lixiao
Repository: spark Updated Branches: refs/heads/master b9192bb3f -> e256392a1 [SPARK-17659][SQL] Partitioned View is Not Supported By SHOW CREATE TABLE ### What changes were proposed in this pull request? `Partitioned View` is not supported by SPARK SQL. For Hive partitioned view, SHOW

spark git commit: [SPARK-17620][SQL] Determine Serde by hive.default.fileformat when Creating Hive Serde Tables

2016-10-14 Thread lixiao
Repository: spark Updated Branches: refs/heads/master de1c1ca5c -> 7ab86244e [SPARK-17620][SQL] Determine Serde by hive.default.fileformat when Creating Hive Serde Tables ## What changes were proposed in this pull request? Make sure the hive.default.fileformat is used to when creating the

spark git commit: [SPARK-18035][SQL] Introduce performant and memory efficient APIs to create ArrayBasedMapData

2016-10-22 Thread lixiao
Repository: spark Updated Branches: refs/heads/master bc167a2a5 -> eff4aed1a [SPARK-18035][SQL] Introduce performant and memory efficient APIs to create ArrayBasedMapData ## What changes were proposed in this pull request? Jira: https://issues.apache.org/jira/browse/SPARK-18035 In

spark git commit: [SPARK-17620][SQL] Determine Serde by hive.default.fileformat when Creating Hive Serde Tables

2016-10-17 Thread lixiao
Repository: spark Updated Branches: refs/heads/master d88a1bae6 -> 813ab5e02 [SPARK-17620][SQL] Determine Serde by hive.default.fileformat when Creating Hive Serde Tables ## What changes were proposed in this pull request? Reopens the closed PR https://github.com/apache/spark/pull/15190

spark git commit: [SPARK-18533] Raise correct error upon specification of schema for datasource tables created using CTAS

2016-11-22 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 64b9de9c0 -> 4b96ffb13 [SPARK-18533] Raise correct error upon specification of schema for datasource tables created using CTAS ## What changes were proposed in this pull request? Fixes the inconsistency of error raised between data

spark git commit: [SPARK-16803][SQL] SaveAsTable does not work when target table is a Hive serde table

2016-11-22 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 2fd101b2f -> 9c42d4a76 [SPARK-16803][SQL] SaveAsTable does not work when target table is a Hive serde table ### What changes were proposed in this pull request? In Spark 2.0, `SaveAsTable` does not work when the target table is a Hive

spark git commit: [SPARK-18533] Raise correct error upon specification of schema for datasource tables created using CTAS

2016-11-22 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 9c42d4a76 -> 39a1d3063 [SPARK-18533] Raise correct error upon specification of schema for datasource tables created using CTAS ## What changes were proposed in this pull request? Fixes the inconsistency of error raised between data source

spark git commit: [SPARK-18594][SQL] Name Validation of Databases/Tables

2016-11-27 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 9c5495728 -> 1e8fbefa3 [SPARK-18594][SQL] Name Validation of Databases/Tables ### What changes were proposed in this pull request? Currently, the name validation checks are limited to table creation. It is enfored by Analyzer rule:

spark git commit: [SPARK-18594][SQL] Name Validation of Databases/Tables

2016-11-27 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 9c03c5646 -> 07f32c228 [SPARK-18594][SQL] Name Validation of Databases/Tables ### What changes were proposed in this pull request? Currently, the name validation checks are limited to table creation. It is enfored by Analyzer rule:

spark git commit: [SPARK-18300][SQL] Do not apply foldable propagation with expand as a child.

2016-11-15 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 33be4da53 -> f14ae4900 [SPARK-18300][SQL] Do not apply foldable propagation with expand as a child. ## What changes were proposed in this pull request? The `FoldablePropagation` optimizer rule, pulls foldable values out from under an

spark git commit: [SPARK-18300][SQL] Do not apply foldable propagation with expand as a child.

2016-11-15 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 0762c0ceb -> 0af94e772 [SPARK-18300][SQL] Do not apply foldable propagation with expand as a child. ## What changes were proposed in this pull request? The `FoldablePropagation` optimizer rule, pulls foldable values out from under an

spark git commit: [SPARK-17982][SQL][BACKPORT-2.0] SQLBuilder should wrap the generated SQL with parenthesis for LIMIT

2016-11-11 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.0 99575e88f -> 80c1a1f30 [SPARK-17982][SQL][BACKPORT-2.0] SQLBuilder should wrap the generated SQL with parenthesis for LIMIT ## What changes were proposed in this pull request? Currently, `SQLBuilder` handles `LIMIT` by always adding

spark git commit: [SPARK-17982][SQL] SQLBuilder should wrap the generated SQL with parenthesis for LIMIT

2016-11-11 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 00c9c7d96 -> 465e4b40b [SPARK-17982][SQL] SQLBuilder should wrap the generated SQL with parenthesis for LIMIT ## What changes were proposed in this pull request? Currently, `SQLBuilder` handles `LIMIT` by always adding `LIMIT` at the

spark git commit: [SPARK-17982][SQL] SQLBuilder should wrap the generated SQL with parenthesis for LIMIT

2016-11-11 Thread lixiao
Repository: spark Updated Branches: refs/heads/master a531fe1a8 -> d42bb7cc4 [SPARK-17982][SQL] SQLBuilder should wrap the generated SQL with parenthesis for LIMIT ## What changes were proposed in this pull request? Currently, `SQLBuilder` handles `LIMIT` by always adding `LIMIT` at the end

spark git commit: [SPARK-17693][SQL] Fixed Insert Failure To Data Source Tables when the Schema has the Comment Field

2016-10-26 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 12b3e8d2e -> 93b8ad184 [SPARK-17693][SQL] Fixed Insert Failure To Data Source Tables when the Schema has the Comment Field ### What changes were proposed in this pull request? ```SQL CREATE TABLE tab1(col1 int COMMENT 'a', col2 int) USING

spark git commit: [SPARK-17122][SQL] support drop current database

2016-11-03 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 965c964c2 -> c4c5328f2 [SPARK-17122][SQL] support drop current database ## What changes were proposed in this pull request? In Spark 1.6 and earlier, we can drop the database we are using. In Spark 2.0, native implementation prevent

spark git commit: [SPARK-17122][SQL] support drop current database

2016-11-03 Thread lixiao
Repository: spark Updated Branches: refs/heads/master d24e73647 -> 96cc1b567 [SPARK-17122][SQL] support drop current database ## What changes were proposed in this pull request? In Spark 1.6 and earlier, we can drop the database we are using. In Spark 2.0, native implementation prevent us

[2/3] spark git commit: [SPARK-17963][SQL][DOCUMENTATION] Add examples (extend) in each expression and improve documentation

2016-11-02 Thread lixiao
http://git-wip-us.apache.org/repos/asf/spark/blob/1e29f0a0/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/mathExpressions.scala -- diff --git

[3/3] spark git commit: [SPARK-17963][SQL][DOCUMENTATION] Add examples (extend) in each expression and improve documentation

2016-11-02 Thread lixiao
[SPARK-17963][SQL][DOCUMENTATION] Add examples (extend) in each expression and improve documentation ## What changes were proposed in this pull request? This PR proposes to change the documentation for functions. Please refer the discussion from https://github.com/apache/spark/pull/15513 The

[2/3] spark git commit: [SPARK-17963][SQL][DOCUMENTATION] Add examples (extend) in each expression and improve documentation

2016-11-02 Thread lixiao
http://git-wip-us.apache.org/repos/asf/spark/blob/7eb2ca8e/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/mathExpressions.scala -- diff --git

[1/3] spark git commit: [SPARK-17963][SQL][DOCUMENTATION] Add examples (extend) in each expression and improve documentation

2016-11-02 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 3a1bc6f47 -> 7eb2ca8e3 http://git-wip-us.apache.org/repos/asf/spark/blob/7eb2ca8e/sql/core/src/main/scala/org/apache/spark/sql/execution/command/functions.scala -- diff

[3/3] spark git commit: [SPARK-17963][SQL][DOCUMENTATION] Add examples (extend) in each expression and improve documentation

2016-11-02 Thread lixiao
[SPARK-17963][SQL][DOCUMENTATION] Add examples (extend) in each expression and improve documentation ## What changes were proposed in this pull request? This PR proposes to change the documentation for functions. Please refer the discussion from https://github.com/apache/spark/pull/15513 The

[1/3] spark git commit: [SPARK-17963][SQL][DOCUMENTATION] Add examples (extend) in each expression and improve documentation

2016-11-02 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 5ea2f9e5e -> 1e29f0a0d http://git-wip-us.apache.org/repos/asf/spark/blob/1e29f0a0/sql/core/src/main/scala/org/apache/spark/sql/execution/command/functions.scala -- diff

spark git commit: [SPARK-18175][SQL] Improve the test case coverage of implicit type casting

2016-11-02 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 7eb2ca8e3 -> 9ddec8636 [SPARK-18175][SQL] Improve the test case coverage of implicit type casting ### What changes were proposed in this pull request? So far, we have limited test case coverage about implicit type casting. We need to

spark git commit: [SPARK-18175][SQL] Improve the test case coverage of implicit type casting

2016-11-02 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 1e29f0a0d -> 2cf39d638 [SPARK-18175][SQL] Improve the test case coverage of implicit type casting ### What changes were proposed in this pull request? So far, we have limited test case coverage about implicit type casting. We need to

spark git commit: [SPARK-17720][SQL] introduce static SQL conf

2016-10-11 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 5b77e66dd -> b9a147181 [SPARK-17720][SQL] introduce static SQL conf ## What changes were proposed in this pull request? SQLConf is session-scoped and mutable. However, we do have the requirement for a static SQL conf, which is global and

spark git commit: [SPARK-17719][SPARK-17776][SQL] Unify and tie up options in a single place in JDBC datasource package

2016-10-10 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 19a5bae47 -> 0c0ad436a [SPARK-17719][SPARK-17776][SQL] Unify and tie up options in a single place in JDBC datasource package ## What changes were proposed in this pull request? This PR proposes to fix arbitrary usages among `Map[String,

spark git commit: [SPARK-18766][SQL] Push Down Filter Through BatchEvalPython (Python UDF)

2016-12-10 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 3a3e65ada -> 422a45cf0 [SPARK-18766][SQL] Push Down Filter Through BatchEvalPython (Python UDF) ### What changes were proposed in this pull request? Currently, when users use Python UDF in Filter, BatchEvalPython is always generated below

spark git commit: [SPARK-18752][HIVE] isSrcLocal" value should be set from user query.

2016-12-12 Thread lixiao
Repository: spark Updated Branches: refs/heads/master bf42c2db5 -> 476b34c23 [SPARK-18752][HIVE] isSrcLocal" value should be set from user query. The value of the "isSrcLocal" parameter passed to Hive's loadTable and loadPartition methods needs to be set according to the user query (e.g.

spark git commit: [SPARK-16848][SQL] Check schema validation for user-specified schema in jdbc and table APIs

2017-01-11 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 43fa21b3e -> 24100f162 [SPARK-16848][SQL] Check schema validation for user-specified schema in jdbc and table APIs ## What changes were proposed in this pull request? This PR proposes to throw an exception for both jdbc APIs when user

spark git commit: [SPARK-19157][SQL] should be able to change spark.sql.runSQLOnFiles at runtime

2017-01-10 Thread lixiao
Repository: spark Updated Branches: refs/heads/master bc6c56e94 -> 3b19c74e7 [SPARK-19157][SQL] should be able to change spark.sql.runSQLOnFiles at runtime ## What changes were proposed in this pull request? The analyzer rule that supports to query files directly will be added to

spark git commit: [SPARK-19178][SQL] convert string of large numbers to int should return null

2017-01-12 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 7f24a0b6c -> 6b34e745b [SPARK-19178][SQL] convert string of large numbers to int should return null ## What changes were proposed in this pull request? When we convert a string to integral, we will convert that string to `decimal(20, 0)`

spark git commit: [SPARK-19151][SQL] DataFrameWriter.saveAsTable support hive overwrite

2017-01-14 Thread lixiao
Repository: spark Updated Branches: refs/heads/master b6a7aa4f7 -> 894235390 [SPARK-19151][SQL] DataFrameWriter.saveAsTable support hive overwrite ## What changes were proposed in this pull request? After [SPARK-19107](https://issues.apache.org/jira/browse/SPARK-19107), we now can treat

spark git commit: [SPARK-17237][SQL] Remove backticks in a pivot result schema

2017-01-12 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 042e32d18 -> 23944d0d6 [SPARK-17237][SQL] Remove backticks in a pivot result schema ## What changes were proposed in this pull request? Pivoting adds backticks (e.g. 3_count(\`c\`)) in column names and, in some cases, thes causes

spark git commit: [SPARK-17237][SQL] Remove backticks in a pivot result schema

2017-01-12 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 2bc4d4e28 -> 5585ed93b [SPARK-17237][SQL] Remove backticks in a pivot result schema ## What changes were proposed in this pull request? Pivoting adds backticks (e.g. 3_count(\`c\`)) in column names and, in some cases, thes causes analysis

spark git commit: [SPARK-18141][SQL] Fix to quote column names in the predicate clause of the JDBC RDD generated sql statement

2016-12-01 Thread lixiao
Repository: spark Updated Branches: refs/heads/master 37e52f879 -> 70c5549ee [SPARK-18141][SQL] Fix to quote column names in the predicate clause of the JDBC RDD generated sql statement ## What changes were proposed in this pull request? SQL query generated for the JDBC data source is not

spark git commit: [SPARK-18141][SQL] Fix to quote column names in the predicate clause of the JDBC RDD generated sql statement

2016-12-01 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-2.1 2d2e80180 -> 2f91b0154 [SPARK-18141][SQL] Fix to quote column names in the predicate clause of the JDBC RDD generated sql statement ## What changes were proposed in this pull request? SQL query generated for the JDBC data source is

spark git commit: [SPARK-12446][SQL][BACKPORT-1.6] Add unit tests for JDBCRDD internal functions

2016-12-02 Thread lixiao
Repository: spark Updated Branches: refs/heads/branch-1.6 8f25cb26f -> 70f271b56 [SPARK-12446][SQL][BACKPORT-1.6] Add unit tests for JDBCRDD internal functions No tests done for JDBCRDD#compileFilter. Author: Takeshi YAMAMURO Closes #10409 from maropu/AddTestsInJdbcRdd. (cherry picked

spark git commit: [SPARK-18871][SQL] New test cases for IN/NOT IN subquery

2017-01-05 Thread lixiao
Repository: spark Updated Branches: refs/heads/master dfc4c935b -> bcc510b02 [SPARK-18871][SQL] New test cases for IN/NOT IN subquery ## What changes were proposed in this pull request? This PR extends the existing IN/NOT IN subquery test cases coverage, adds more test cases to the IN

  1   2   3   4   5   6   7   8   9   10   >