This is an automated email from the ASF dual-hosted git repository.

agrove pushed a commit to branch branch-0.10
in repository https://gitbox.apache.org/repos/asf/datafusion-comet.git


The following commit(s) were added to refs/heads/branch-0.10 by this push:
     new 7ae31c989 fix merge conflict (#2535)
7ae31c989 is described below

commit 7ae31c9897fa70b04c5830f878cb0f2d456f7827
Author: Andy Grove <[email protected]>
AuthorDate: Mon Oct 6 11:29:58 2025 -0600

    fix merge conflict (#2535)
    
    Co-authored-by: hsiang-c <[email protected]>
---
 dev/diffs/iceberg/1.8.1.diff                 |  11 ++-
 dev/diffs/iceberg/{1.8.1.diff => 1.9.1.diff} | 124 +++++++++++++--------------
 2 files changed, 67 insertions(+), 68 deletions(-)

diff --git a/dev/diffs/iceberg/1.8.1.diff b/dev/diffs/iceberg/1.8.1.diff
index 938a2af08..6af0ea6bb 100644
--- a/dev/diffs/iceberg/1.8.1.diff
+++ b/dev/diffs/iceberg/1.8.1.diff
@@ -17,7 +17,7 @@ index 7327b38905d..7967109f039 100644
        exclude group: 'org.apache.avro', module: 'avro'
        // already shaded by Parquet
 diff --git a/gradle/libs.versions.toml b/gradle/libs.versions.toml
-index 04ffa8f4edc..cc0099ccc93 100644
+index 04ffa8f4edc..a909cd552c1 100644
 --- a/gradle/libs.versions.toml
 +++ b/gradle/libs.versions.toml
 @@ -34,6 +34,7 @@ azuresdk-bom = "1.2.31"
@@ -300,10 +300,10 @@ index 00000000000..ddf6c7de5ae
 +}
 diff --git 
a/parquet/src/main/java/org/apache/iceberg/parquet/CometVectorizedParquetReader.java
 
b/parquet/src/main/java/org/apache/iceberg/parquet/CometVectorizedParquetReader.java
 new file mode 100644
-index 00000000000..88b195b76a2
+index 00000000000..a3cba401827
 --- /dev/null
 +++ 
b/parquet/src/main/java/org/apache/iceberg/parquet/CometVectorizedParquetReader.java
-@@ -0,0 +1,255 @@
+@@ -0,0 +1,260 @@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one
 + * or more contributor license agreements.  See the NOTICE file
@@ -446,6 +446,7 @@ index 00000000000..88b195b76a2
 +    private long valuesRead = 0;
 +    private T last = null;
 +    private final FileReader cometReader;
++    private ReadConf conf;
 +
 +    FileIterator(
 +        ReadConf conf,
@@ -470,6 +471,7 @@ index 00000000000..88b195b76a2
 +              length,
 +              fileEncryptionKey,
 +              fileAADPrefix);
++      this.conf = conf;
 +    }
 +
 +    private FileReader newCometReader(
@@ -556,6 +558,9 @@ index 00000000000..88b195b76a2
 +    public void close() throws IOException {
 +      model.close();
 +      cometReader.close();
++      if (conf != null && conf.reader() != null) {
++        conf.reader().close();
++      }
 +    }
 +  }
 +}
diff --git a/dev/diffs/iceberg/1.8.1.diff b/dev/diffs/iceberg/1.9.1.diff
similarity index 96%
copy from dev/diffs/iceberg/1.8.1.diff
copy to dev/diffs/iceberg/1.9.1.diff
index 938a2af08..52e499874 100644
--- a/dev/diffs/iceberg/1.8.1.diff
+++ b/dev/diffs/iceberg/1.9.1.diff
@@ -1,8 +1,8 @@
 diff --git a/build.gradle b/build.gradle
-index 7327b38905d..7967109f039 100644
+index 998f2ee9ea6..017e61be98c 100644
 --- a/build.gradle
 +++ b/build.gradle
-@@ -780,6 +780,13 @@ project(':iceberg-parquet') {
+@@ -814,6 +814,13 @@ project(':iceberg-parquet') {
      implementation project(':iceberg-core')
      implementation project(':iceberg-common')
  
@@ -17,26 +17,17 @@ index 7327b38905d..7967109f039 100644
        exclude group: 'org.apache.avro', module: 'avro'
        // already shaded by Parquet
 diff --git a/gradle/libs.versions.toml b/gradle/libs.versions.toml
-index 04ffa8f4edc..cc0099ccc93 100644
+index c50991c5fc6..f7ad00f0b78 100644
 --- a/gradle/libs.versions.toml
 +++ b/gradle/libs.versions.toml
-@@ -34,6 +34,7 @@ azuresdk-bom = "1.2.31"
- awssdk-s3accessgrants = "2.3.0"
+@@ -36,6 +36,7 @@ awssdk-s3accessgrants = "2.3.0"
+ bson-ver = "4.11.5"
  caffeine = "2.9.3"
- calcite = "1.10.0"
-+comet = "0.10.1"
+ calcite = "1.39.0"
++comet = "0.11.0-SNAPSHOT"
  datasketches = "6.2.0"
- delta-standalone = "3.3.0"
- delta-spark = "3.3.0"
-@@ -81,7 +82,7 @@ slf4j = "2.0.16"
- snowflake-jdbc = "3.22.0"
- spark-hive33 = "3.3.4"
- spark-hive34 = "3.4.4"
--spark-hive35 = "3.5.4"
-+spark-hive35 = "3.5.6"
- sqlite-jdbc = "3.48.0.0"
- testcontainers = "1.20.4"
- tez010 = "0.10.4"
+ delta-standalone = "3.3.1"
+ delta-spark = "3.3.1"
 diff --git 
a/parquet/src/main/java/org/apache/iceberg/parquet/CometTypeUtils.java 
b/parquet/src/main/java/org/apache/iceberg/parquet/CometTypeUtils.java
 new file mode 100644
 index 00000000000..ddf6c7de5ae
@@ -300,10 +291,10 @@ index 00000000000..ddf6c7de5ae
 +}
 diff --git 
a/parquet/src/main/java/org/apache/iceberg/parquet/CometVectorizedParquetReader.java
 
b/parquet/src/main/java/org/apache/iceberg/parquet/CometVectorizedParquetReader.java
 new file mode 100644
-index 00000000000..88b195b76a2
+index 00000000000..a3cba401827
 --- /dev/null
 +++ 
b/parquet/src/main/java/org/apache/iceberg/parquet/CometVectorizedParquetReader.java
-@@ -0,0 +1,255 @@
+@@ -0,0 +1,260 @@
 +/*
 + * Licensed to the Apache Software Foundation (ASF) under one
 + * or more contributor license agreements.  See the NOTICE file
@@ -446,6 +437,7 @@ index 00000000000..88b195b76a2
 +    private long valuesRead = 0;
 +    private T last = null;
 +    private final FileReader cometReader;
++    private ReadConf conf;
 +
 +    FileIterator(
 +        ReadConf conf,
@@ -470,6 +462,7 @@ index 00000000000..88b195b76a2
 +              length,
 +              fileEncryptionKey,
 +              fileAADPrefix);
++      this.conf = conf;
 +    }
 +
 +    private FileReader newCometReader(
@@ -556,14 +549,17 @@ index 00000000000..88b195b76a2
 +    public void close() throws IOException {
 +      model.close();
 +      cometReader.close();
++      if (conf != null && conf.reader() != null) {
++        conf.reader().close();
++      }
 +    }
 +  }
 +}
 diff --git a/parquet/src/main/java/org/apache/iceberg/parquet/Parquet.java 
b/parquet/src/main/java/org/apache/iceberg/parquet/Parquet.java
-index 2c37a52449e..3442cfc4375 100644
+index 31f9e2a80a6..520f142c212 100644
 --- a/parquet/src/main/java/org/apache/iceberg/parquet/Parquet.java
 +++ b/parquet/src/main/java/org/apache/iceberg/parquet/Parquet.java
-@@ -1075,6 +1075,7 @@ public class Parquet {
+@@ -1124,6 +1124,7 @@ public class Parquet {
      private NameMapping nameMapping = null;
      private ByteBuffer fileEncryptionKey = null;
      private ByteBuffer fileAADPrefix = null;
@@ -571,7 +567,7 @@ index 2c37a52449e..3442cfc4375 100644
  
      private ReadBuilder(InputFile file) {
        this.file = file;
-@@ -1172,6 +1173,11 @@ public class Parquet {
+@@ -1168,6 +1169,11 @@ public class Parquet {
        return this;
      }
  
@@ -580,19 +576,19 @@ index 2c37a52449e..3442cfc4375 100644
 +      return this;
 +    }
 +
-     public ReadBuilder withFileEncryptionKey(ByteBuffer encryptionKey) {
-       this.fileEncryptionKey = encryptionKey;
-       return this;
-@@ -1182,7 +1188,7 @@ public class Parquet {
-       return this;
+     /**
+      * @deprecated will be removed in 2.0.0; use {@link 
#createReaderFunc(Function)} instead
+      */
+@@ -1263,7 +1269,7 @@ public class Parquet {
      }
  
+     @Override
 -    @SuppressWarnings({"unchecked", "checkstyle:CyclomaticComplexity"})
 +    @SuppressWarnings({"unchecked", "checkstyle:CyclomaticComplexity", 
"MethodLength"})
      public <D> CloseableIterable<D> build() {
        FileDecryptionProperties fileDecryptionProperties = null;
        if (fileEncryptionKey != null) {
-@@ -1234,16 +1240,35 @@ public class Parquet {
+@@ -1315,16 +1321,35 @@ public class Parquet {
          }
  
          if (batchedReaderFunc != null) {
@@ -636,8 +632,8 @@ index 2c37a52449e..3442cfc4375 100644
 +                maxRecordsPerBatch);
 +          }
          } else {
-           return new org.apache.iceberg.parquet.ParquetReader<>(
-               file, schema, options, readerFunc, mapping, filter, 
reuseContainers, caseSensitive);
+           Function<MessageType, ParquetValueReader<?>> readBuilder =
+               readerFuncWithSchema != null
 diff --git a/parquet/src/main/java/org/apache/iceberg/parquet/ReadConf.java 
b/parquet/src/main/java/org/apache/iceberg/parquet/ReadConf.java
 index 1fb2372ba56..142e5fbadf1 100644
 --- a/parquet/src/main/java/org/apache/iceberg/parquet/ReadConf.java
@@ -658,7 +654,7 @@ index 1fb2372ba56..142e5fbadf1 100644
      return model;
    }
 diff --git a/spark/v3.5/build.gradle b/spark/v3.5/build.gradle
-index e2d2c7a7ac0..f64232dc57f 100644
+index 572c32f9292..d155f634a4d 100644
 --- a/spark/v3.5/build.gradle
 +++ b/spark/v3.5/build.gradle
 @@ -75,7 +75,7 @@ 
project(":iceberg-spark:iceberg-spark-${sparkMajorVersion}_${scalaVersion}") {
@@ -670,7 +666,7 @@ index e2d2c7a7ac0..f64232dc57f 100644
  
      implementation libs.parquet.column
      implementation libs.parquet.hadoop
-@@ -183,7 +183,7 @@ 
project(":iceberg-spark:iceberg-spark-extensions-${sparkMajorVersion}_${scalaVer
+@@ -184,7 +184,7 @@ 
project(":iceberg-spark:iceberg-spark-extensions-${sparkMajorVersion}_${scalaVer
      testImplementation libs.avro.avro
      testImplementation libs.parquet.hadoop
      testImplementation libs.awaitility
@@ -679,7 +675,7 @@ index e2d2c7a7ac0..f64232dc57f 100644
  
      // Required because we remove antlr plugin dependencies from the compile 
configuration, see note above
      runtimeOnly libs.antlr.runtime
-@@ -263,6 +263,7 @@ 
project(":iceberg-spark:iceberg-spark-runtime-${sparkMajorVersion}_${scalaVersio
+@@ -265,6 +265,7 @@ 
project(":iceberg-spark:iceberg-spark-runtime-${sparkMajorVersion}_${scalaVersio
      integrationImplementation project(path: ':iceberg-hive-metastore', 
configuration: 'testArtifacts')
      integrationImplementation project(path: 
":iceberg-spark:iceberg-spark-${sparkMajorVersion}_${scalaVersion}", 
configuration: 'testArtifacts')
      integrationImplementation project(path: 
":iceberg-spark:iceberg-spark-extensions-${sparkMajorVersion}_${scalaVersion}", 
configuration: 'testArtifacts')
@@ -687,7 +683,7 @@ index e2d2c7a7ac0..f64232dc57f 100644
  
      // runtime dependencies for running Hive Catalog based integration test
      integrationRuntimeOnly project(':iceberg-hive-metastore')
-@@ -300,8 +301,8 @@ 
project(":iceberg-spark:iceberg-spark-runtime-${sparkMajorVersion}_${scalaVersio
+@@ -302,8 +303,8 @@ 
project(":iceberg-spark:iceberg-spark-runtime-${sparkMajorVersion}_${scalaVersio
      relocate 'org.apache.avro', 'org.apache.iceberg.shaded.org.apache.avro'
      relocate 'avro.shaded', 'org.apache.iceberg.shaded.org.apache.avro.shaded'
      relocate 'com.thoughtworks.paranamer', 
'org.apache.iceberg.shaded.com.thoughtworks.paranamer'
@@ -720,7 +716,7 @@ index 578845e3da2..0118b30683d 100644
              .getOrCreate();
  
 diff --git 
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestCallStatementParser.java
 
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestCallStatementParser.java
-index ade19de36fe..150a2cddbc8 100644
+index ecf9e6f8a59..0f8cced69aa 100644
 --- 
a/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestCallStatementParser.java
 +++ 
b/spark/v3.5/spark-extensions/src/test/java/org/apache/iceberg/spark/extensions/TestCallStatementParser.java
 @@ -56,6 +56,16 @@ public class TestCallStatementParser {
@@ -850,13 +846,13 @@ index 68c537e34a4..f66be2f3896 100644
        builder
            .config("parquet.dictionary.page.size", "1")
 diff --git 
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/vectorized/CometColumnReader.java
 
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/vectorized/CometColumnReader.java
-index 4794863ab1b..8bb508f19f8 100644
+index 16159dcbdff..eba1a2a0fb1 100644
 --- 
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/vectorized/CometColumnReader.java
 +++ 
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/vectorized/CometColumnReader.java
-@@ -20,21 +20,25 @@ package org.apache.iceberg.spark.data.vectorized;
+@@ -19,18 +19,22 @@
+ package org.apache.iceberg.spark.data.vectorized;
  
  import java.io.IOException;
- import java.util.Map;
 +import org.apache.comet.CometConf;
 +import org.apache.comet.CometSchemaImporter;
  import org.apache.comet.parquet.AbstractColumnReader;
@@ -873,15 +869,12 @@ index 4794863ab1b..8bb508f19f8 100644
  import org.apache.iceberg.spark.SparkSchemaUtil;
  import org.apache.iceberg.types.Types;
  import org.apache.parquet.column.ColumnDescriptor;
- import org.apache.parquet.column.page.PageReadStore;
 -import org.apache.parquet.column.page.PageReader;
- import org.apache.parquet.hadoop.metadata.ColumnChunkMetaData;
- import org.apache.parquet.hadoop.metadata.ColumnPath;
 +import org.apache.spark.sql.internal.SQLConf;
  import org.apache.spark.sql.types.DataType;
  import org.apache.spark.sql.types.Metadata;
  import org.apache.spark.sql.types.StructField;
-@@ -46,23 +50,28 @@ class CometColumnReader implements 
VectorizedReader<ColumnVector> {
+@@ -42,23 +46,28 @@ class CometColumnReader implements 
VectorizedReader<ColumnVector> {
  
    private final ColumnDescriptor descriptor;
    private final DataType sparkType;
@@ -912,7 +905,7 @@ index 4794863ab1b..8bb508f19f8 100644
    }
  
    public AbstractColumnReader delegate() {
-@@ -96,7 +105,26 @@ class CometColumnReader implements 
VectorizedReader<ColumnVector> {
+@@ -92,7 +101,26 @@ class CometColumnReader implements 
VectorizedReader<ColumnVector> {
      }
  
      this.importer = new CometSchemaImporter(new RootAllocator());
@@ -940,7 +933,7 @@ index 4794863ab1b..8bb508f19f8 100644
      this.initialized = true;
    }
  
-@@ -115,9 +143,9 @@ class CometColumnReader implements 
VectorizedReader<ColumnVector> {
+@@ -111,9 +139,9 @@ class CometColumnReader implements 
VectorizedReader<ColumnVector> {
     * <p>NOTE: this should be called before reading a new Parquet column 
chunk, and after {@link
     * CometColumnReader#reset} is called.
     */
@@ -953,7 +946,7 @@ index 4794863ab1b..8bb508f19f8 100644
  
    @Override
 diff --git 
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/vectorized/CometColumnarBatchReader.java
 
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/vectorized/CometColumnarBatchReader.java
-index 1440e5d1d3f..85cca62e90f 100644
+index 04ac69476ad..916face2bf2 100644
 --- 
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/vectorized/CometColumnarBatchReader.java
 +++ 
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/data/vectorized/CometColumnarBatchReader.java
 @@ -22,8 +22,12 @@ import java.io.IOException;
@@ -990,7 +983,7 @@ index 1440e5d1d3f..85cca62e90f 100644
    }
  
    @Override
-@@ -85,19 +87,22 @@ class CometColumnarBatchReader implements 
VectorizedReader<ColumnarBatch> {
+@@ -79,19 +81,22 @@ class CometColumnarBatchReader implements 
VectorizedReader<ColumnarBatch> {
              && !(readers[i] instanceof CometPositionColumnReader)
              && !(readers[i] instanceof CometDeleteColumnReader)) {
            readers[i].reset();
@@ -1016,7 +1009,7 @@ index 1440e5d1d3f..85cca62e90f 100644
              .getRowIndexOffset()
              .orElseThrow(
                  () ->
-@@ -154,9 +159,17 @@ class CometColumnarBatchReader implements 
VectorizedReader<ColumnarBatch> {
+@@ -148,9 +153,17 @@ class CometColumnarBatchReader implements 
VectorizedReader<ColumnarBatch> {
          Pair<int[], Integer> pair = buildRowIdMapping(vectors);
          if (pair != null) {
            int[] rowIdMapping = pair.first();
@@ -1126,10 +1119,10 @@ index 780e1750a52..57892ac4c59 100644
    }
  
 diff --git 
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkBatch.java
 
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkBatch.java
-index 11f054b1171..f443042ba48 100644
+index 45bd48aea2e..5362578861c 100644
 --- 
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkBatch.java
 +++ 
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkBatch.java
-@@ -172,17 +172,32 @@ class SparkBatch implements Batch {
+@@ -172,11 +172,11 @@ class SparkBatch implements Batch {
      return field.type().isPrimitiveType() || 
MetadataColumns.isMetadataColumn(field.fieldId());
    }
  
@@ -1143,7 +1136,8 @@ index 11f054b1171..f443042ba48 100644
    }
  
    private boolean supportsCometBatchReads(Types.NestedField field) {
-     return field.type().isPrimitiveType() && 
!field.type().typeId().equals(Type.TypeID.UUID);
+@@ -186,6 +186,21 @@ class SparkBatch implements Batch {
+         && field.fieldId() != 
MetadataColumns.LAST_UPDATED_SEQUENCE_NUMBER.fieldId();
    }
  
 +  private boolean supportsCometBatchReads(ScanTask task) {
@@ -1165,18 +1159,18 @@ index 11f054b1171..f443042ba48 100644
    // - ORC vectorization is enabled
    // - all tasks are of type FileScanTask and read only ORC files with no 
delete files
 diff --git 
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkScan.java 
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkScan.java
-index 019f3919dc5..656e0600ac0 100644
+index 106b296de09..967b0d41d08 100644
 --- 
a/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkScan.java
 +++ 
b/spark/v3.5/spark/src/main/java/org/apache/iceberg/spark/source/SparkScan.java
-@@ -23,6 +23,7 @@ import java.util.List;
- import java.util.Map;
+@@ -24,6 +24,7 @@ import java.util.Map;
+ import java.util.Optional;
  import java.util.function.Supplier;
  import java.util.stream.Collectors;
 +import org.apache.comet.parquet.SupportsComet;
  import org.apache.iceberg.BlobMetadata;
  import org.apache.iceberg.ScanTask;
  import org.apache.iceberg.ScanTaskGroup;
-@@ -94,7 +95,7 @@ import org.apache.spark.sql.types.StructType;
+@@ -95,7 +96,7 @@ import org.apache.spark.sql.types.StructType;
  import org.slf4j.Logger;
  import org.slf4j.LoggerFactory;
  
@@ -1185,7 +1179,7 @@ index 019f3919dc5..656e0600ac0 100644
    private static final Logger LOG = LoggerFactory.getLogger(SparkScan.class);
    private static final String NDV_KEY = "ndv";
  
-@@ -348,4 +349,10 @@ abstract class SparkScan implements Scan, 
SupportsReportStatistics {
+@@ -351,4 +352,10 @@ abstract class SparkScan implements Scan, 
SupportsReportStatistics {
        return splitSize;
      }
    }
@@ -1197,7 +1191,7 @@ index 019f3919dc5..656e0600ac0 100644
 +  }
  }
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/SparkDistributedDataScanTestBase.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/SparkDistributedDataScanTestBase.java
-index 404ba728460..19d0cd5c827 100644
+index 404ba728460..26d6f9b613f 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/SparkDistributedDataScanTestBase.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/SparkDistributedDataScanTestBase.java
 @@ -90,6 +90,16 @@ public abstract class SparkDistributedDataScanTestBase
@@ -1206,8 +1200,8 @@ index 404ba728460..19d0cd5c827 100644
          .config(SQLConf.SHUFFLE_PARTITIONS().key(), "4")
 +        .config("spark.plugins", "org.apache.spark.CometPlugin")
 +        .config(
-+                "spark.shuffle.manager",
-+                
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
++            "spark.shuffle.manager",
++            
"org.apache.spark.sql.comet.execution.shuffle.CometShuffleManager")
 +        .config("spark.comet.explainFallback.enabled", "true")
 +        .config("spark.sql.iceberg.parquet.reader-type", "COMET")
 +        .config("spark.memory.offHeap.enabled", "true")
@@ -1281,7 +1275,7 @@ index 2665d7ba8d3..306e859ce1a 100644
    }
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/TestBase.java 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/TestBase.java
-index de68351f6e3..3ab0d3db477 100644
+index 3e9f3334ef6..ecf855986b7 100644
 --- a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/TestBase.java
 +++ b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/TestBase.java
 @@ -77,6 +77,17 @@ public abstract class TestBase extends SparkTestHelperBase {
@@ -1329,10 +1323,10 @@ index bc4e722bc86..1a40d8edcb3 100644
  
    @AfterAll
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/ScanTestBase.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/ScanTestBase.java
-index 3a269740b70..59592eb2ad7 100644
+index 0886df957db..ba2ee6b1c2c 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/ScanTestBase.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/ScanTestBase.java
-@@ -54,7 +54,20 @@ public abstract class ScanTestBase extends AvroDataTest {
+@@ -57,7 +57,20 @@ public abstract class ScanTestBase extends AvroDataTest {
  
    @BeforeAll
    public static void startSpark() {
@@ -1355,7 +1349,7 @@ index 3a269740b70..59592eb2ad7 100644
    }
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestCompressionSettings.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestCompressionSettings.java
-index f411920a5dc..367c2e2cb45 100644
+index 6b7d861364f..f9805fa8df5 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestCompressionSettings.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestCompressionSettings.java
 @@ -144,7 +144,20 @@ public class TestCompressionSettings extends 
CatalogTestBase {
@@ -1764,10 +1758,10 @@ index baf7fa8f88a..509c5deba51 100644
              .getOrCreate();
  
 diff --git 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestStructuredStreaming.java
 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestStructuredStreaming.java
-index 17db46b85c3..d68b638ce09 100644
+index c84a65cbe95..42ee74280c1 100644
 --- 
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestStructuredStreaming.java
 +++ 
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/source/TestStructuredStreaming.java
-@@ -65,6 +65,16 @@ public class TestStructuredStreaming {
+@@ -67,6 +67,16 @@ public class TestStructuredStreaming {
          SparkSession.builder()
              .master("local[2]")
              .config("spark.sql.shuffle.partitions", 4)


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to