This is an automated email from the ASF dual-hosted git repository.

agrove pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/datafusion-comet.git


The following commit(s) were added to refs/heads/main by this push:
     new 8112e1aca feat: [iceberg] Enable Comet shuffle in Iceberg diff (#2205)
8112e1aca is described below

commit 8112e1acab497ca3a915d4ab3fdce4ce9e64c88a
Author: Andy Grove <agr...@apache.org>
AuthorDate: Thu Aug 21 18:52:54 2025 -0600

    feat: [iceberg] Enable Comet shuffle in Iceberg diff (#2205)
---
 dev/diffs/iceberg/1.8.1.diff | 279 +++++++++++++++++++++++++++----------------
 1 file changed, 177 insertions(+), 102 deletions(-)

diff --git a/dev/diffs/iceberg/1.8.1.diff b/dev/diffs/iceberg/1.8.1.diff
index 15c66af20..2f80453de 100644
--- a/dev/diffs/iceberg/1.8.1.diff
+++ b/dev/diffs/iceberg/1.8.1.diff
@@ -699,15 +699,17 @@ index e2d2c7a..f64232d 100644
      relocate 'io.airlift', 'org.apache.iceberg.shaded.io.airlift'
      relocate 'org.apache.hc.client5', 
'org.apache.iceberg.shaded.org.apache.hc.client5'
 diff --git 
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/ExtensionsTestBase.java
 
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/ExtensionsTestBase.java
-index 578845e..2b16f8e 100644
+index 578845e..0118b30 100644
 --- 
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/ExtensionsTestBase.java
 +++ 
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/ExtensionsTestBase.java
-@@ -57,6 +57,14 @@ public abstract class ExtensionsTestBase extends 
CatalogTestBase {
+@@ -57,6 +57,16 @@ public abstract class ExtensionsTestBase extends 
CatalogTestBase {
              
.config("spark.sql.legacy.respectNullabilityInTextDatasetConversion", "true")
              .config(
                  SQLConf.ADAPTIVE_EXECUTION_ENABLED().key(), 
String.valueOf(RANDOM.nextBoolean()))
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -718,15 +720,17 @@ index 578845e..2b16f8e 100644
              .getOrCreate();
  
 diff --git 
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestCallStatementParser.java
 
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestCallStatementParser.java
-index ade19de..255c416 100644
+index ade19de..150a2cd 100644
 --- 
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestCallStatementParser.java
 +++ 
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestCallStatementParser.java
-@@ -56,6 +56,14 @@ public class TestCallStatementParser {
+@@ -56,6 +56,16 @@ public class TestCallStatementParser {
              .master("local[2]")
              .config("spark.sql.extensions", 
IcebergSparkSessionExtensions.class.getName())
              .config("spark.extra.prop", "value")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -737,15 +741,17 @@ index ade19de..255c416 100644
      TestCallStatementParser.parser = spark.sessionState().sqlParser();
    }
 diff --git 
a/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/action/DeleteOrphanFilesBenchmark.java
 
b/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/action/DeleteOrphanFilesBenchmark.java
-index 64edb10..ea4d76f 100644
+index 64edb10..5bb449f 100644
 --- 
a/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/action/DeleteOrphanFilesBenchmark.java
 +++ 
b/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/action/DeleteOrphanFilesBenchmark.java
-@@ -179,6 +179,14 @@ public class DeleteOrphanFilesBenchmark {
+@@ -179,6 +179,16 @@ public class DeleteOrphanFilesBenchmark {
              .config("spark.sql.catalog.spark_catalog", 
SparkSessionCatalog.class.getName())
              .config("spark.sql.catalog.spark_catalog.type", "hadoop")
              .config("spark.sql.catalog.spark_catalog.warehouse", 
catalogWarehouse())
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -756,15 +762,17 @@ index 64edb10..ea4d76f 100644
      spark = builder.getOrCreate();
    }
 diff --git 
a/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/action/IcebergSortCompactionBenchmark.java
 
b/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/action/IcebergSortCompactionBenchmark.java
-index a5d0456..5aec87e 100644
+index a5d0456..4af408f 100644
 --- 
a/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/action/IcebergSortCompactionBenchmark.java
 +++ 
b/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/action/IcebergSortCompactionBenchmark.java
-@@ -392,6 +392,14 @@ public class IcebergSortCompactionBenchmark {
+@@ -392,6 +392,16 @@ public class IcebergSortCompactionBenchmark {
                  "spark.sql.catalog.spark_catalog", 
"org.apache.iceberg.spark.SparkSessionCatalog")
              .config("spark.sql.catalog.spark_catalog.type", "hadoop")
              .config("spark.sql.catalog.spark_catalog.warehouse", 
getCatalogWarehouse())
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -775,15 +783,17 @@ index a5d0456..5aec87e 100644
      spark = builder.getOrCreate();
      Configuration sparkHadoopConf = spark.sessionState().newHadoopConf();
 diff --git 
a/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/DVReaderBenchmark.java
 
b/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/DVReaderBenchmark.java
-index c6794e4..60bf66c 100644
+index c6794e4..f735919 100644
 --- 
a/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/DVReaderBenchmark.java
 +++ 
b/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/DVReaderBenchmark.java
-@@ -239,6 +239,14 @@ public class DVReaderBenchmark {
+@@ -239,6 +239,16 @@ public class DVReaderBenchmark {
              .config("spark.sql.catalog.spark_catalog", 
SparkSessionCatalog.class.getName())
              .config("spark.sql.catalog.spark_catalog.type", "hadoop")
              .config("spark.sql.catalog.spark_catalog.warehouse", 
newWarehouseDir())
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -794,15 +804,17 @@ index c6794e4..60bf66c 100644
              .getOrCreate();
    }
 diff --git 
a/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/DVWriterBenchmark.java
 
b/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/DVWriterBenchmark.java
-index ac74fb5..3fca3bd 100644
+index ac74fb5..e011b8b 100644
 --- 
a/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/DVWriterBenchmark.java
 +++ 
b/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/DVWriterBenchmark.java
-@@ -223,6 +223,14 @@ public class DVWriterBenchmark {
+@@ -223,6 +223,16 @@ public class DVWriterBenchmark {
              .config("spark.sql.catalog.spark_catalog", 
SparkSessionCatalog.class.getName())
              .config("spark.sql.catalog.spark_catalog.type", "hadoop")
              .config("spark.sql.catalog.spark_catalog.warehouse", 
newWarehouseDir())
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -813,10 +825,10 @@ index ac74fb5..3fca3bd 100644
              .getOrCreate();
    }
 diff --git 
a/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/IcebergSourceBenchmark.java
 
b/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/IcebergSourceBenchmark.java
-index 68c537e..2b821f8 100644
+index 68c537e..f66be2f 100644
 --- 
a/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/IcebergSourceBenchmark.java
 +++ 
b/spark/v3.5/spark/src/jmh/java/org/apache/iceberg/spark/source/IcebergSourceBenchmark.java
-@@ -94,7 +94,17 @@ public abstract class IcebergSourceBenchmark {
+@@ -94,7 +94,19 @@ public abstract class IcebergSourceBenchmark {
    }
  
    protected void setupSpark(boolean enableDictionaryEncoding) {
@@ -825,7 +837,9 @@ index 68c537e..2b821f8 100644
 +        SparkSession.builder()
 +            .config("spark.ui.enabled", false)
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1158,15 +1172,17 @@ index 019f391..656e060 100644
 +  }
  }
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/SparkDistributedDataScanTestBase.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/SparkDistributedDataScanTestBase.java
-index 404ba72..64c5e51 100644
+index 404ba72..19d0cd5 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/SparkDistributedDataScanTestBase.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/SparkDistributedDataScanTestBase.java
-@@ -90,6 +90,14 @@ public abstract class SparkDistributedDataScanTestBase
+@@ -90,6 +90,16 @@ public abstract class SparkDistributedDataScanTestBase
          .master("local[2]")
          .config("spark.serializer", serializer)
          .config(SQLConf.SHUFFLE_PARTITIONS().key(), "4")
 +        .config("spark.plugins", "org.apache.spark.CometPlugin")
-+        .config("spark.comet.exec.shuffle.enabled", "false")
++        .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +        .config("spark.comet.explainFallback.enabled", "true")
 +        .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +        .config("spark.memory.offHeap.enabled", "true")
@@ -1177,15 +1193,17 @@ index 404ba72..64c5e51 100644
    }
  }
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanDeletes.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanDeletes.java
-index 659507e..ead2de6 100644
+index 659507e..eb9cedc 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanDeletes.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanDeletes.java
-@@ -73,6 +73,14 @@ public class TestSparkDistributedDataScanDeletes
+@@ -73,6 +73,16 @@ public class TestSparkDistributedDataScanDeletes
              .master("local[2]")
              .config("spark.serializer", 
"org.apache.spark.serializer.KryoSerializer")
              .config(SQLConf.SHUFFLE_PARTITIONS().key(), "4")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1196,15 +1214,17 @@ index 659507e..ead2de6 100644
    }
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanFilterFiles.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanFilterFiles.java
-index a218f96..a11f574 100644
+index a218f96..395c024 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanFilterFiles.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanFilterFiles.java
-@@ -62,6 +62,14 @@ public class TestSparkDistributedDataScanFilterFiles
+@@ -62,6 +62,16 @@ public class TestSparkDistributedDataScanFilterFiles
              .master("local[2]")
              .config("spark.serializer", 
"org.apache.spark.serializer.KryoSerializer")
              .config(SQLConf.SHUFFLE_PARTITIONS().key(), "4")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1215,15 +1235,17 @@ index a218f96..a11f574 100644
    }
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanReporting.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanReporting.java
-index 2665d7b..95647d7 100644
+index 2665d7b..306e859 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanReporting.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/TestSparkDistributedDataScanReporting.java
-@@ -63,6 +63,14 @@ public class TestSparkDistributedDataScanReporting
+@@ -63,6 +63,16 @@ public class TestSparkDistributedDataScanReporting
              .master("local[2]")
              .config("spark.serializer", 
"org.apache.spark.serializer.KryoSerializer")
              .config(SQLConf.SHUFFLE_PARTITIONS().key(), "4")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1234,15 +1256,17 @@ index 2665d7b..95647d7 100644
    }
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/TestBase.java 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/TestBase.java
-index de68351..248f927 100644
+index de68351..3ab0d3d 100644
 --- a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/TestBase.java
 +++ b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/TestBase.java
-@@ -77,6 +77,15 @@ public abstract class TestBase extends SparkTestHelperBase {
+@@ -77,6 +77,17 @@ public abstract class TestBase extends SparkTestHelperBase {
              .config(SQLConf.PARTITION_OVERWRITE_MODE().key(), "dynamic")
              .config("spark.hadoop." + METASTOREURIS.varname, 
hiveConf.get(METASTOREURIS.varname))
              
.config("spark.sql.legacy.respectNullabilityInTextDatasetConversion", "true")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1254,10 +1278,10 @@ index de68351..248f927 100644
              .getOrCreate();
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/data/parquet/vectorized/TestParquetDictionaryEncodedVectorizedReads.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/data/parquet/vectorized/TestParquetDictionaryEncodedVectorizedReads.java
-index bc4e722..a2d8346 100644
+index bc4e722..1a40d8e 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/data/parquet/vectorized/TestParquetDictionaryEncodedVectorizedReads.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/data/parquet/vectorized/TestParquetDictionaryEncodedVectorizedReads.java
-@@ -59,7 +59,18 @@ public class TestParquetDictionaryEncodedVectorizedReads 
extends TestParquetVect
+@@ -59,7 +59,20 @@ public class TestParquetDictionaryEncodedVectorizedReads 
extends TestParquetVect
  
    @BeforeAll
    public static void startSpark() {
@@ -1266,7 +1290,9 @@ index bc4e722..a2d8346 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1278,10 +1304,10 @@ index bc4e722..a2d8346 100644
  
    @AfterAll
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/ScanTestBase.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/ScanTestBase.java
-index 3a26974..9be02b9 100644
+index 3a26974..59592eb 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/ScanTestBase.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/ScanTestBase.java
-@@ -54,7 +54,18 @@ public abstract class ScanTestBase extends AvroDataTest {
+@@ -54,7 +54,20 @@ public abstract class ScanTestBase extends AvroDataTest {
  
    @BeforeAll
    public static void startSpark() {
@@ -1290,7 +1316,9 @@ index 3a26974..9be02b9 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1302,10 +1330,10 @@ index 3a26974..9be02b9 100644
    }
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestCompressionSettings.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestCompressionSettings.java
-index f411920..f2a9593 100644
+index f411920..367c2e2 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestCompressionSettings.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestCompressionSettings.java
-@@ -144,7 +144,18 @@ public class TestCompressionSettings extends 
CatalogTestBase {
+@@ -144,7 +144,20 @@ public class TestCompressionSettings extends 
CatalogTestBase {
  
    @BeforeAll
    public static void startSpark() {
@@ -1314,7 +1342,9 @@ index f411920..f2a9593 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1326,10 +1356,10 @@ index f411920..f2a9593 100644
  
    @BeforeEach
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestDataSourceOptions.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestDataSourceOptions.java
-index c4ba96e..5a1073d 100644
+index c4ba96e..a0e77db 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestDataSourceOptions.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestDataSourceOptions.java
-@@ -75,7 +75,18 @@ public class TestDataSourceOptions extends 
TestBaseWithCatalog {
+@@ -75,7 +75,20 @@ public class TestDataSourceOptions extends 
TestBaseWithCatalog {
  
    @BeforeAll
    public static void startSpark() {
@@ -1338,7 +1368,9 @@ index c4ba96e..5a1073d 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1350,10 +1382,10 @@ index c4ba96e..5a1073d 100644
  
    @AfterAll
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestFilteredScan.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestFilteredScan.java
-index 3481735..688daba 100644
+index 3481735..bdd3152 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestFilteredScan.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestFilteredScan.java
-@@ -110,7 +110,18 @@ public class TestFilteredScan {
+@@ -110,7 +110,20 @@ public class TestFilteredScan {
  
    @BeforeAll
    public static void startSpark() {
@@ -1362,7 +1394,9 @@ index 3481735..688daba 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1374,10 +1408,10 @@ index 3481735..688daba 100644
  
    @AfterAll
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestForwardCompatibility.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestForwardCompatibility.java
-index 84c99a5..58a414b 100644
+index 84c99a5..6f4b2af 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestForwardCompatibility.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestForwardCompatibility.java
-@@ -93,7 +93,18 @@ public class TestForwardCompatibility {
+@@ -93,7 +93,20 @@ public class TestForwardCompatibility {
  
    @BeforeAll
    public static void startSpark() {
@@ -1386,7 +1420,9 @@ index 84c99a5..58a414b 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1398,10 +1434,10 @@ index 84c99a5..58a414b 100644
  
    @AfterAll
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestIcebergSpark.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestIcebergSpark.java
-index 7eff93d..990f386 100644
+index 7eff93d..1774acd 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestIcebergSpark.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestIcebergSpark.java
-@@ -46,7 +46,18 @@ public class TestIcebergSpark {
+@@ -46,7 +46,20 @@ public class TestIcebergSpark {
  
    @BeforeAll
    public static void startSpark() {
@@ -1410,7 +1446,9 @@ index 7eff93d..990f386 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1422,10 +1460,10 @@ index 7eff93d..990f386 100644
  
    @AfterAll
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestPartitionPruning.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestPartitionPruning.java
-index 9464f68..e8d08ec 100644
+index 9464f68..b5dbb88 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestPartitionPruning.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestPartitionPruning.java
-@@ -112,7 +112,18 @@ public class TestPartitionPruning {
+@@ -112,7 +112,20 @@ public class TestPartitionPruning {
  
    @BeforeAll
    public static void startSpark() {
@@ -1434,7 +1472,9 @@ index 9464f68..e8d08ec 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1446,10 +1486,10 @@ index 9464f68..e8d08ec 100644
  
      String optionKey = String.format("fs.%s.impl", 
CountOpenLocalFileSystem.scheme);
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestPartitionValues.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestPartitionValues.java
-index 5c218f2..829b67b 100644
+index 5c218f2..e48e6d1 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestPartitionValues.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestPartitionValues.java
-@@ -107,7 +107,18 @@ public class TestPartitionValues {
+@@ -107,7 +107,20 @@ public class TestPartitionValues {
  
    @BeforeAll
    public static void startSpark() {
@@ -1458,7 +1498,9 @@ index 5c218f2..829b67b 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1470,10 +1512,10 @@ index 5c218f2..829b67b 100644
  
    @AfterAll
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSnapshotSelection.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSnapshotSelection.java
-index a7334a5..abe55f2 100644
+index a7334a5..dca7fdd 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSnapshotSelection.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSnapshotSelection.java
-@@ -87,7 +87,18 @@ public class TestSnapshotSelection {
+@@ -87,7 +87,20 @@ public class TestSnapshotSelection {
  
    @BeforeAll
    public static void startSpark() {
@@ -1482,7 +1524,9 @@ index a7334a5..abe55f2 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1494,10 +1538,10 @@ index a7334a5..abe55f2 100644
  
    @AfterAll
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkDataFile.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkDataFile.java
-index 182b1ef..ffceac5 100644
+index 182b1ef..5da4a3c 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkDataFile.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkDataFile.java
-@@ -120,7 +120,18 @@ public class TestSparkDataFile {
+@@ -120,7 +120,20 @@ public class TestSparkDataFile {
  
    @BeforeAll
    public static void startSpark() {
@@ -1506,7 +1550,9 @@ index 182b1ef..ffceac5 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1518,10 +1564,10 @@ index 182b1ef..ffceac5 100644
    }
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkDataWrite.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkDataWrite.java
-index fb2b312..c3f4e14 100644
+index fb2b312..5927a40 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkDataWrite.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkDataWrite.java
-@@ -96,7 +96,18 @@ public class TestSparkDataWrite {
+@@ -96,7 +96,20 @@ public class TestSparkDataWrite {
  
    @BeforeAll
    public static void startSpark() {
@@ -1530,7 +1576,9 @@ index fb2b312..c3f4e14 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1541,7 +1589,7 @@ index fb2b312..c3f4e14 100644
    }
  
    @AfterEach
-@@ -140,7 +151,7 @@ public class TestSparkDataWrite {
+@@ -140,7 +153,7 @@ public class TestSparkDataWrite {
      Dataset<Row> result = spark.read().format("iceberg").load(targetLocation);
  
      List<SimpleRecord> actual =
@@ -1550,7 +1598,7 @@ index fb2b312..c3f4e14 100644
      assertThat(actual).as("Number of rows should 
match").hasSameSizeAs(expected);
      assertThat(actual).as("Result rows should match").isEqualTo(expected);
      for (ManifestFile manifest :
-@@ -210,7 +221,7 @@ public class TestSparkDataWrite {
+@@ -210,7 +223,7 @@ public class TestSparkDataWrite {
      Dataset<Row> result = spark.read().format("iceberg").load(targetLocation);
  
      List<SimpleRecord> actual =
@@ -1559,7 +1607,7 @@ index fb2b312..c3f4e14 100644
      assertThat(actual).as("Number of rows should 
match").hasSameSizeAs(expected);
      assertThat(actual).as("Result rows should match").isEqualTo(expected);
    }
-@@ -256,7 +267,7 @@ public class TestSparkDataWrite {
+@@ -256,7 +269,7 @@ public class TestSparkDataWrite {
      Dataset<Row> result = spark.read().format("iceberg").load(targetLocation);
  
      List<SimpleRecord> actual =
@@ -1568,7 +1616,7 @@ index fb2b312..c3f4e14 100644
      assertThat(actual).as("Number of rows should 
match").hasSameSizeAs(expected);
      assertThat(actual).as("Result rows should match").isEqualTo(expected);
    }
-@@ -309,7 +320,7 @@ public class TestSparkDataWrite {
+@@ -309,7 +322,7 @@ public class TestSparkDataWrite {
      Dataset<Row> result = spark.read().format("iceberg").load(targetLocation);
  
      List<SimpleRecord> actual =
@@ -1577,7 +1625,7 @@ index fb2b312..c3f4e14 100644
      assertThat(actual).as("Number of rows should 
match").hasSameSizeAs(expected);
      assertThat(actual).as("Result rows should match").isEqualTo(expected);
    }
-@@ -352,7 +363,7 @@ public class TestSparkDataWrite {
+@@ -352,7 +365,7 @@ public class TestSparkDataWrite {
      Dataset<Row> result = spark.read().format("iceberg").load(targetLocation);
  
      List<SimpleRecord> actual =
@@ -1586,7 +1634,7 @@ index fb2b312..c3f4e14 100644
      assertThat(actual).as("Number of rows should 
match").hasSameSizeAs(expected);
      assertThat(actual).as("Result rows should match").isEqualTo(expected);
    }
-@@ -392,7 +403,7 @@ public class TestSparkDataWrite {
+@@ -392,7 +405,7 @@ public class TestSparkDataWrite {
      Dataset<Row> result = spark.read().format("iceberg").load(targetLocation);
  
      List<SimpleRecord> actual =
@@ -1595,7 +1643,7 @@ index fb2b312..c3f4e14 100644
      assertThat(actual).as("Number of rows should 
match").hasSameSizeAs(expected);
      assertThat(actual).as("Result rows should match").isEqualTo(expected);
  
-@@ -458,7 +469,7 @@ public class TestSparkDataWrite {
+@@ -458,7 +471,7 @@ public class TestSparkDataWrite {
      Dataset<Row> result = spark.read().format("iceberg").load(targetLocation);
  
      List<SimpleRecord> actual =
@@ -1604,7 +1652,7 @@ index fb2b312..c3f4e14 100644
      assertThat(actual).as("Number of rows should 
match").hasSameSizeAs(expected);
      assertThat(actual).as("Result rows should match").isEqualTo(expected);
    }
-@@ -622,7 +633,7 @@ public class TestSparkDataWrite {
+@@ -622,7 +635,7 @@ public class TestSparkDataWrite {
      Dataset<Row> result = spark.read().format("iceberg").load(targetLocation);
  
      List<SimpleRecord> actual =
@@ -1613,7 +1661,7 @@ index fb2b312..c3f4e14 100644
      assertThat(actual).as("Number of rows should 
match").hasSameSizeAs(expected);
      assertThat(actual).as("Result rows should match").isEqualTo(expected);
  
-@@ -708,7 +719,7 @@ public class TestSparkDataWrite {
+@@ -708,7 +721,7 @@ public class TestSparkDataWrite {
      // Since write and commit succeeded, the rows should be readable
      Dataset<Row> result = spark.read().format("iceberg").load(targetLocation);
      List<SimpleRecord> actual =
@@ -1623,10 +1671,10 @@ index fb2b312..c3f4e14 100644
      assertThat(actual)
          .describedAs("Result rows should match")
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
-index becf6a0..b98c2f6 100644
+index becf6a0..beb7c80 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReadProjection.java
-@@ -83,7 +83,18 @@ public class TestSparkReadProjection extends 
TestReadProjection {
+@@ -83,7 +83,20 @@ public class TestSparkReadProjection extends 
TestReadProjection {
  
    @BeforeAll
    public static void startSpark() {
@@ -1635,7 +1683,9 @@ index becf6a0..b98c2f6 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1647,15 +1697,17 @@ index becf6a0..b98c2f6 100644
          ImmutableMap.of(
              "type", "hive",
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReaderDeletes.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReaderDeletes.java
-index 4f1cef5..f31c9b1 100644
+index 4f1cef5..0e34e08 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReaderDeletes.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReaderDeletes.java
-@@ -136,6 +136,14 @@ public class TestSparkReaderDeletes extends 
DeleteReadTests {
+@@ -136,6 +136,16 @@ public class TestSparkReaderDeletes extends 
DeleteReadTests {
              .config("spark.ui.liveUpdate.period", 0)
              .config(SQLConf.PARTITION_OVERWRITE_MODE().key(), "dynamic")
              .config("spark.hadoop." + METASTOREURIS.varname, 
hiveConf.get(METASTOREURIS.varname))
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1666,15 +1718,17 @@ index 4f1cef5..f31c9b1 100644
              .getOrCreate();
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReaderWithBloomFilter.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReaderWithBloomFilter.java
-index baf7fa8..02cf1c8 100644
+index baf7fa8..509c5de 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReaderWithBloomFilter.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestSparkReaderWithBloomFilter.java
-@@ -182,6 +182,14 @@ public class TestSparkReaderWithBloomFilter {
+@@ -182,6 +182,16 @@ public class TestSparkReaderWithBloomFilter {
          SparkSession.builder()
              .master("local[2]")
              .config("spark.hadoop." + METASTOREURIS.varname, 
hiveConf.get(METASTOREURIS.varname))
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1685,15 +1739,17 @@ index baf7fa8..02cf1c8 100644
              .getOrCreate();
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestStructuredStreaming.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestStructuredStreaming.java
-index 17db46b..e6afced 100644
+index 17db46b..d68b638 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestStructuredStreaming.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestStructuredStreaming.java
-@@ -65,6 +65,14 @@ public class TestStructuredStreaming {
+@@ -65,6 +65,16 @@ public class TestStructuredStreaming {
          SparkSession.builder()
              .master("local[2]")
              .config("spark.sql.shuffle.partitions", 4)
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1704,10 +1760,10 @@ index 17db46b..e6afced 100644
    }
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestTimestampWithoutZone.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestTimestampWithoutZone.java
-index 306444b..525ddb0 100644
+index 306444b..0ed52fd 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestTimestampWithoutZone.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestTimestampWithoutZone.java
-@@ -75,7 +75,18 @@ public class TestTimestampWithoutZone extends TestBase {
+@@ -75,7 +75,20 @@ public class TestTimestampWithoutZone extends TestBase {
  
    @BeforeAll
    public static void startSpark() {
@@ -1716,7 +1772,9 @@ index 306444b..525ddb0 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1728,10 +1786,10 @@ index 306444b..525ddb0 100644
  
    @AfterAll
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestWriteMetricsConfig.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestWriteMetricsConfig.java
-index 841268a..8da9ea2 100644
+index 841268a..e290c24 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestWriteMetricsConfig.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestWriteMetricsConfig.java
-@@ -80,7 +80,18 @@ public class TestWriteMetricsConfig {
+@@ -80,7 +80,20 @@ public class TestWriteMetricsConfig {
  
    @BeforeAll
    public static void startSpark() {
@@ -1740,7 +1798,9 @@ index 841268a..8da9ea2 100644
 +        SparkSession.builder()
 +            .master("local[2]")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1752,15 +1812,17 @@ index 841268a..8da9ea2 100644
    }
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/sql/TestAggregatePushDown.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/sql/TestAggregatePushDown.java
-index 6e09252..f92dad4 100644
+index 6e09252..438bc4d 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/sql/TestAggregatePushDown.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/sql/TestAggregatePushDown.java
-@@ -60,6 +60,14 @@ public class TestAggregatePushDown extends CatalogTestBase {
+@@ -60,6 +60,16 @@ public class TestAggregatePushDown extends CatalogTestBase {
          SparkSession.builder()
              .master("local[2]")
              .config("spark.sql.iceberg.aggregate_pushdown", "true")
 +            .config("spark.plugins", "org.apache.spark.CometPlugin")
-+            .config("spark.comet.exec.shuffle.enabled", "false")
++            .config(
++                "spark.shuffle.manager",
++                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +            .config("spark.comet.explainFallback.enabled", "true")
 +            .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +            .config("spark.memory.offHeap.enabled", "true")
@@ -1785,3 +1847,16 @@ index 9d2ce2b..5e23368 100644
      } else {
        assertThat(planAsString).as("Should be no post scan 
filter").doesNotContain("Filter (");
      }
+diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/sql/TestStoragePartitionedJoins.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/sql/TestStoragePartitionedJoins.java
+index 6719c45..2515454 100644
+--- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/sql/TestStoragePartitionedJoins.java
++++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/sql/TestStoragePartitionedJoins.java
+@@ -616,7 +616,7 @@ public class TestStoragePartitionedJoins extends 
TestBaseWithCatalog {
+             + "FROM %s t1 "
+             + "INNER JOIN %s t2 "
+             + "ON t1.id = t2.id AND t1.%s = t2.%s "
+-            + "ORDER BY t1.id, t1.%s",
++            + "ORDER BY t1.id, t1.%s, t1.salary",
+         sourceColumnName,
+         tableName,
+         tableName(OTHER_TABLE_NAME),


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@datafusion.apache.org
For additional commands, e-mail: commits-h...@datafusion.apache.org


Reply via email to