Yuming Wang created SPARK-32767: ----------------------------------- Summary: Bucket join should work if SHUFFLE_PARTITIONS larger than bucket number Key: SPARK-32767 URL: https://issues.apache.org/jira/browse/SPARK-32767 Project: Spark Issue Type: Bug Components: SQL Affects Versions: 3.1.0 Reporter: Yuming Wang
How to reproduce this issue: {code:scala} spark.range(1000).write.bucketBy(500, "id").saveAsTable("t1") spark.range(1000).write.bucketBy(50, "id").saveAsTable("t2") sql("set spark.sql.shuffle.partitions=600") sql("set spark.sql.autoBroadcastJoinThreshold=-1") sql("select * from t1 join t2 on t1.id = t2.id").explain() {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org