[ https://issues.apache.org/jira/browse/SPARK-12975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Michael Armbrust resolved SPARK-12975. -------------------------------------- Resolution: Fixed Fix Version/s: 2.0.0 Issue resolved by pull request 10891 [https://github.com/apache/spark/pull/10891] > Throwing Exception when Bucketing Columns are part of Partitioning Columns > -------------------------------------------------------------------------- > > Key: SPARK-12975 > URL: https://issues.apache.org/jira/browse/SPARK-12975 > Project: Spark > Issue Type: Improvement > Components: SQL > Affects Versions: 2.0.0 > Reporter: Xiao Li > Fix For: 2.0.0 > > > When users are using partitionBy and bucketBy at the same time, some > bucketing columns might be part of partitioning columns. For example, > {code} > df.write > .format(source) > .partitionBy("i") > .bucketBy(8, "i", "k") > .sortBy("k") > .saveAsTable("bucketed_table") > {code} > However, in the above case, adding column `i` into `bucketBy` is useless. It > is just wasting extra CPU when reading or writing bucket tables. Thus, like > Hive, we can issue an exception and let users do the change. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org