[ https://issues.apache.org/jira/browse/SPARK-32767?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Apache Spark reassigned SPARK-32767: ------------------------------------ Assignee: (was: Apache Spark) > Bucket join should work if spark.sql.shuffle.partitions larger than bucket > number > --------------------------------------------------------------------------------- > > Key: SPARK-32767 > URL: https://issues.apache.org/jira/browse/SPARK-32767 > Project: Spark > Issue Type: Bug > Components: SQL > Affects Versions: 3.1.0 > Reporter: Yuming Wang > Priority: Major > > How to reproduce this issue: > {code:scala} > spark.range(1000).write.bucketBy(500, "id").saveAsTable("t1") > spark.range(1000).write.bucketBy(50, "id").saveAsTable("t2") > sql("set spark.sql.shuffle.partitions=600") > sql("set spark.sql.autoBroadcastJoinThreshold=-1") > sql("select * from t1 join t2 on t1.id = t2.id").explain() > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org