This is an automated email from the ASF dual-hosted git repository. agrove pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/datafusion-comet.git
The following commit(s) were added to refs/heads/main by this push: new 5875cff1a chore: Enable `native_iceberg_compat` Spark SQL tests (for real, this time) (#1910) 5875cff1a is described below commit 5875cff1a92abbba69ade3c75fc2fc66f261387b Author: Andy Grove <agr...@apache.org> AuthorDate: Wed Jun 18 19:05:21 2025 -0600 chore: Enable `native_iceberg_compat` Spark SQL tests (for real, this time) (#1910) --- .github/workflows/spark_sql_test.yml | 2 +- dev/diffs/3.5.6.diff | 20 +++++++++++++++----- 2 files changed, 16 insertions(+), 6 deletions(-) diff --git a/.github/workflows/spark_sql_test.yml b/.github/workflows/spark_sql_test.yml index 7bd51fa67..34e63f38e 100644 --- a/.github/workflows/spark_sql_test.yml +++ b/.github/workflows/spark_sql_test.yml @@ -114,6 +114,6 @@ jobs: run: | cd apache-spark rm -rf /root/.m2/repository/org/apache/parquet # somehow parquet cache requires cleanups - ENABLE_COMET=true ENABLE_COMET_SHUFFLE=true COMET_SPARK_PARQUET_IMPL=native_iceberg_compat build/sbt -Dsbt.log.noformat=true ${{ matrix.module.args1 }} "${{ matrix.module.args2 }}" + ENABLE_COMET=true ENABLE_COMET_SHUFFLE=true COMET_PARQUET_SCAN_IMPL=native_iceberg_compat build/sbt -Dsbt.log.noformat=true ${{ matrix.module.args1 }} "${{ matrix.module.args2 }}" env: LC_ALL: "C.UTF-8" diff --git a/dev/diffs/3.5.6.diff b/dev/diffs/3.5.6.diff index 0d35cbf7c..e7d953bbf 100644 --- a/dev/diffs/3.5.6.diff +++ b/dev/diffs/3.5.6.diff @@ -1877,7 +1877,7 @@ index 07e2849ce6f..3e73645b638 100644 ParquetOutputFormat.WRITER_VERSION -> ParquetProperties.WriterVersion.PARQUET_2_0.toString ) diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala -index 8e88049f51e..d3c0737d52e 100644 +index 8e88049f51e..8f3cf8a0f80 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala @@ -1095,7 +1095,11 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared @@ -1938,7 +1938,17 @@ index 8e88049f51e..d3c0737d52e 100644 import testImplicits._ // keep() should take effect on StartsWith/EndsWith/Contains Seq( -@@ -1984,7 +1997,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared +@@ -1743,7 +1756,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared + } + } + +- test("SPARK-17091: Convert IN predicate to Parquet filter push-down") { ++ test("SPARK-17091: Convert IN predicate to Parquet filter push-down", ++ IgnoreCometNativeScan("Comet has different push-down behavior")) { + val schema = StructType(Seq( + StructField("a", IntegerType, nullable = false) + )) +@@ -1984,7 +1998,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared } } @@ -1948,7 +1958,7 @@ index 8e88049f51e..d3c0737d52e 100644 // block 1: // null count min max // page-0 0 0 99 -@@ -2044,7 +2058,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared +@@ -2044,7 +2059,8 @@ abstract class ParquetFilterSuite extends QueryTest with ParquetTest with Shared } } @@ -1958,7 +1968,7 @@ index 8e88049f51e..d3c0737d52e 100644 withTempPath { dir => val path = dir.getCanonicalPath spark.range(100).selectExpr("id * 2 AS id") -@@ -2276,7 +2291,11 @@ class ParquetV1FilterSuite extends ParquetFilterSuite { +@@ -2276,7 +2292,11 @@ class ParquetV1FilterSuite extends ParquetFilterSuite { assert(pushedParquetFilters.exists(_.getClass === filterClass), s"${pushedParquetFilters.map(_.getClass).toList} did not contain ${filterClass}.") @@ -1971,7 +1981,7 @@ index 8e88049f51e..d3c0737d52e 100644 } else { assert(selectedFilters.isEmpty, "There is filter pushed down") } -@@ -2336,7 +2355,11 @@ class ParquetV2FilterSuite extends ParquetFilterSuite { +@@ -2336,7 +2356,11 @@ class ParquetV2FilterSuite extends ParquetFilterSuite { assert(pushedParquetFilters.exists(_.getClass === filterClass), s"${pushedParquetFilters.map(_.getClass).toList} did not contain ${filterClass}.") --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@datafusion.apache.org For additional commands, e-mail: commits-h...@datafusion.apache.org