This is an automated email from the ASF dual-hosted git repository.

zhli pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/incubator-gluten.git


The following commit(s) were added to refs/heads/main by this push:
     new b798ef5f1 [GLUTEN-5844][CORE] Refactor the usage of 
spark.gluten.enabled (#5845)
b798ef5f1 is described below

commit b798ef5f1bab3a6674d3c69473b5ec781ffe2ca3
Author: sharmaplkt <[email protected]>
AuthorDate: Thu May 23 04:51:02 2024 +0530

    [GLUTEN-5844][CORE] Refactor the usage of spark.gluten.enabled (#5845)
    
    [GLUTEN-5844][CORE] Refactor the usage of spark.gluten.enabled.
    Co-authored-by: Pulkit Sharma <[email protected]>
---
 .../GlutenClickHouseMergeTreeOptimizeSuite.scala   | 50 +++++++++++-----------
 .../GlutenClickHouseNativeWriteTableSuite.scala    | 20 ++++-----
 ...ckHouseTPCHColumnarShuffleParquetAQESuite.scala |  3 +-
 .../GlutenClickHouseTPCHSaltNullParquetSuite.scala |  3 +-
 .../GlutenClickHouseTableAfterRestart.scala        |  6 ++-
 .../execution/GlutenClickhouseFunctionSuite.scala  |  2 +-
 .../scala/org/apache/gluten/s3/S3AuthSuite.scala   |  2 +-
 .../benchmarks/CHAggAndShuffleBenchmark.scala      |  3 +-
 .../benchmarks/CHParquetReadBenchmark.scala        |  3 +-
 .../execution/WholeStageTransformerSuite.scala     |  3 +-
 .../benchmarks/ParquetReadBenchmark.scala          |  2 +-
 .../extension/GlutenCustomerExtensionSuite.scala   |  4 +-
 .../benchmarks/ParquetReadBenchmark.scala          |  2 +-
 .../extension/GlutenCustomerExtensionSuite.scala   |  4 +-
 .../benchmarks/ParquetReadBenchmark.scala          |  2 +-
 .../extension/GlutenCustomerExtensionSuite.scala   |  4 +-
 .../benchmarks/ParquetReadBenchmark.scala          |  2 +-
 .../extension/GlutenCustomerExtensionSuite.scala   |  4 +-
 18 files changed, 68 insertions(+), 51 deletions(-)

diff --git 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseMergeTreeOptimizeSuite.scala
 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseMergeTreeOptimizeSuite.scala
index ae0cd170d..d4302193f 100644
--- 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseMergeTreeOptimizeSuite.scala
+++ 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseMergeTreeOptimizeSuite.scala
@@ -16,6 +16,8 @@
  */
 package org.apache.gluten.execution
 
+import org.apache.gluten.GlutenConfig
+
 import org.apache.spark.SparkConf
 import org.apache.spark.sql.SaveMode
 import org.apache.spark.sql.execution.adaptive.AdaptiveSparkPlanHelper
@@ -122,11 +124,11 @@ class GlutenClickHouseMergeTreeOptimizeSuite
     val ret = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p").collect()
     assert(ret.apply(0).get(0) == 600572)
 
-    spark.sql("set spark.gluten.enabled=false")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
     assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p")) == 
22728)
     spark.sql("VACUUM lineitem_mergetree_optimize_p RETAIN 0 HOURS")
     assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p")) == 
22728)
-    spark.sql("set spark.gluten.enabled=true")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
 
     val ret2 = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p").collect()
     assert(ret2.apply(0).get(0) == 600572)
@@ -154,14 +156,14 @@ class GlutenClickHouseMergeTreeOptimizeSuite
     val ret = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p2").collect()
     assert(ret.apply(0).get(0) == 600572)
 
-    spark.sql("set spark.gluten.enabled=false")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
     assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p2")) 
== 812)
     spark.sql("VACUUM lineitem_mergetree_optimize_p2 RETAIN 0 HOURS")
     assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p2")) 
== 232)
     spark.sql("VACUUM lineitem_mergetree_optimize_p2 RETAIN 0 HOURS")
     // the second VACUUM will remove some empty folders
     assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p2")) 
== 220)
-    spark.sql("set spark.gluten.enabled=true")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
 
     val ret2 = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p2").collect()
     assert(ret2.apply(0).get(0) == 600572)
@@ -185,13 +187,13 @@ class GlutenClickHouseMergeTreeOptimizeSuite
       val ret = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p3").collect()
       assert(ret.apply(0).get(0) == 600572)
 
-      spark.sql("set spark.gluten.enabled=false")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
       assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p3")) 
== 398)
       spark.sql("VACUUM lineitem_mergetree_optimize_p3 RETAIN 0 HOURS")
       assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p3")) 
== 286)
       spark.sql("VACUUM lineitem_mergetree_optimize_p3 RETAIN 0 HOURS")
       assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p3")) 
== 270)
-      spark.sql("set spark.gluten.enabled=true")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
 
       val ret2 = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p3").collect()
       assert(ret2.apply(0).get(0) == 600572)
@@ -216,13 +218,13 @@ class GlutenClickHouseMergeTreeOptimizeSuite
       val ret = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p4").collect()
       assert(ret.apply(0).get(0) == 600572)
 
-      spark.sql("set spark.gluten.enabled=false")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
       assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p4")) 
== 398)
       spark.sql("VACUUM lineitem_mergetree_optimize_p4 RETAIN 0 HOURS")
       assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p4")) 
== 286)
       spark.sql("VACUUM lineitem_mergetree_optimize_p4 RETAIN 0 HOURS")
       assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p4")) 
== 270)
-      spark.sql("set spark.gluten.enabled=true")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
 
       val ret2 = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p4").collect()
       assert(ret2.apply(0).get(0) == 600572)
@@ -246,11 +248,11 @@ class GlutenClickHouseMergeTreeOptimizeSuite
 
       spark.sql("optimize lineitem_mergetree_optimize_p5")
 
-      spark.sql("set spark.gluten.enabled=false")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
       spark.sql("VACUUM lineitem_mergetree_optimize_p5 RETAIN 0 HOURS")
       spark.sql("VACUUM lineitem_mergetree_optimize_p5 RETAIN 0 HOURS")
       assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p5")) 
== 99)
-      spark.sql("set spark.gluten.enabled=true")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
 
       val ret = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p5").collect()
       assert(ret.apply(0).get(0) == 600572)
@@ -266,11 +268,11 @@ class GlutenClickHouseMergeTreeOptimizeSuite
 
       spark.sql("optimize lineitem_mergetree_optimize_p5")
 
-      spark.sql("set spark.gluten.enabled=false")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
       spark.sql("VACUUM lineitem_mergetree_optimize_p5 RETAIN 0 HOURS")
       spark.sql("VACUUM lineitem_mergetree_optimize_p5 RETAIN 0 HOURS")
       assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p5")) 
== 93)
-      spark.sql("set spark.gluten.enabled=true")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
 
       val ret = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p5").collect()
       assert(ret.apply(0).get(0) == 600572)
@@ -279,11 +281,11 @@ class GlutenClickHouseMergeTreeOptimizeSuite
     // now merge all parts (testing merging from merged parts)
     spark.sql("optimize lineitem_mergetree_optimize_p5")
 
-    spark.sql("set spark.gluten.enabled=false")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
     spark.sql("VACUUM lineitem_mergetree_optimize_p5 RETAIN 0 HOURS")
     spark.sql("VACUUM lineitem_mergetree_optimize_p5 RETAIN 0 HOURS")
     assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p5")) 
== 77)
-    spark.sql("set spark.gluten.enabled=true")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
 
     val ret = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p5").collect()
     assert(ret.apply(0).get(0) == 600572)
@@ -309,7 +311,7 @@ class GlutenClickHouseMergeTreeOptimizeSuite
     val ret = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p6").collect()
     assert(ret.apply(0).get(0) == 600572)
 
-    spark.sql("set spark.gluten.enabled=false")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
     assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p6")) 
== {
       if (sparkVersion.equals("3.2")) 931 else 1014
     })
@@ -318,7 +320,7 @@ class GlutenClickHouseMergeTreeOptimizeSuite
     assert(countFiles(new File(s"$basePath/lineitem_mergetree_optimize_p6")) 
== {
       if (sparkVersion.equals("3.2")) 439 else 445
     })
-    spark.sql("set spark.gluten.enabled=true")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
 
     val ret2 = spark.sql("select count(*) from 
lineitem_mergetree_optimize_p6").collect()
     assert(ret2.apply(0).get(0) == 600572)
@@ -341,9 +343,9 @@ class GlutenClickHouseMergeTreeOptimizeSuite
                    |""".stripMargin)
 
       spark.sql("optimize lineitem_mergetree_index")
-      spark.sql("set spark.gluten.enabled=false")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
       spark.sql("vacuum lineitem_mergetree_index")
-      spark.sql("set spark.gluten.enabled=true")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
 
       val df = spark
         .sql(s"""
@@ -387,10 +389,10 @@ class GlutenClickHouseMergeTreeOptimizeSuite
       val clickhouseTable = ClickhouseTable.forPath(spark, dataPath)
       clickhouseTable.optimize().executeCompaction()
 
-      spark.sql("set spark.gluten.enabled=false")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
       clickhouseTable.vacuum(0.0)
       clickhouseTable.vacuum(0.0)
-      spark.sql("set spark.gluten.enabled=true")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
       assert(countFiles(new File(dataPath)) == 99)
 
       val ret = spark.sql(s"select count(*) from 
clickhouse.`$dataPath`").collect()
@@ -408,10 +410,10 @@ class GlutenClickHouseMergeTreeOptimizeSuite
       val clickhouseTable = ClickhouseTable.forPath(spark, dataPath)
       clickhouseTable.optimize().executeCompaction()
 
-      spark.sql("set spark.gluten.enabled=false")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
       clickhouseTable.vacuum(0.0)
       clickhouseTable.vacuum(0.0)
-      spark.sql("set spark.gluten.enabled=true")
+      spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
       assert(countFiles(new File(dataPath)) == 93)
 
       val ret = spark.sql(s"select count(*) from 
clickhouse.`$dataPath`").collect()
@@ -422,10 +424,10 @@ class GlutenClickHouseMergeTreeOptimizeSuite
     val clickhouseTable = ClickhouseTable.forPath(spark, dataPath)
     clickhouseTable.optimize().executeCompaction()
 
-    spark.sql("set spark.gluten.enabled=false")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
     clickhouseTable.vacuum(0.0)
     clickhouseTable.vacuum(0.0)
-    spark.sql("set spark.gluten.enabled=true")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
     assert(countFiles(new File(dataPath)) == 77)
 
     val ret = spark.sql(s"select count(*) from 
clickhouse.`$dataPath`").collect()
diff --git 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseNativeWriteTableSuite.scala
 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseNativeWriteTableSuite.scala
index 6b5924297..9269303d9 100644
--- 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseNativeWriteTableSuite.scala
+++ 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseNativeWriteTableSuite.scala
@@ -173,7 +173,7 @@ class GlutenClickHouseNativeWriteTableSuite
   test("test insert into dir") {
     withSQLConf(
       ("spark.gluten.sql.native.writer.enabled", "true"),
-      ("spark.gluten.enabled", "true")) {
+      (GlutenConfig.GLUTEN_ENABLED.key, "true")) {
 
       val originDF = spark.createDataFrame(genTestData())
       originDF.createOrReplaceTempView("origin_table")
@@ -209,7 +209,7 @@ class GlutenClickHouseNativeWriteTableSuite
     withSQLConf(
       ("spark.gluten.sql.native.writer.enabled", "true"),
       ("spark.sql.orc.compression.codec", "lz4"),
-      ("spark.gluten.enabled", "true")) {
+      (GlutenConfig.GLUTEN_ENABLED.key, "true")) {
 
       val originDF = spark.createDataFrame(genTestData())
       originDF.createOrReplaceTempView("origin_table")
@@ -260,7 +260,7 @@ class GlutenClickHouseNativeWriteTableSuite
   test("test CTAS") {
     withSQLConf(
       ("spark.gluten.sql.native.writer.enabled", "true"),
-      ("spark.gluten.enabled", "true")) {
+      (GlutenConfig.GLUTEN_ENABLED.key, "true")) {
 
       val originDF = spark.createDataFrame(genTestData())
       originDF.createOrReplaceTempView("origin_table")
@@ -311,7 +311,7 @@ class GlutenClickHouseNativeWriteTableSuite
       ("spark.gluten.sql.native.writer.enabled", "true"),
       ("spark.sql.hive.convertMetastoreParquet", "false"),
       ("spark.sql.hive.convertMetastoreOrc", "false"),
-      ("spark.gluten.enabled", "true")
+      (GlutenConfig.GLUTEN_ENABLED.key, "true")
     ) {
 
       val originDF = spark.createDataFrame(genTestData())
@@ -429,7 +429,7 @@ class GlutenClickHouseNativeWriteTableSuite
   test("test 2-col partitioned table") {
     withSQLConf(
       ("spark.gluten.sql.native.writer.enabled", "true"),
-      ("spark.gluten.enabled", "true")) {
+      (GlutenConfig.GLUTEN_ENABLED.key, "true")) {
 
       val fields: ListMap[String, String] = ListMap(
         ("string_field", "string"),
@@ -467,7 +467,7 @@ class GlutenClickHouseNativeWriteTableSuite
       " ignore because takes too long") {
     withSQLConf(
       ("spark.gluten.sql.native.writer.enabled", "true"),
-      ("spark.gluten.enabled", "true")) {
+      (GlutenConfig.GLUTEN_ENABLED.key, "true")) {
 
       val fields: ListMap[String, String] = ListMap(
         ("date_field", "date"),
@@ -508,7 +508,7 @@ class GlutenClickHouseNativeWriteTableSuite
   ignore("test hive parquet/orc table, all columns being partitioned. ") {
     withSQLConf(
       ("spark.gluten.sql.native.writer.enabled", "true"),
-      ("spark.gluten.enabled", "true")) {
+      (GlutenConfig.GLUTEN_ENABLED.key, "true")) {
 
       val fields: ListMap[String, String] = ListMap(
         ("date_field", "date"),
@@ -547,7 +547,7 @@ class GlutenClickHouseNativeWriteTableSuite
   test(("test hive parquet/orc table with aggregated results")) {
     withSQLConf(
       ("spark.gluten.sql.native.writer.enabled", "true"),
-      ("spark.gluten.enabled", "true")) {
+      (GlutenConfig.GLUTEN_ENABLED.key, "true")) {
 
       val fields: ListMap[String, String] = ListMap(
         ("sum(int_field)", "bigint")
@@ -573,7 +573,7 @@ class GlutenClickHouseNativeWriteTableSuite
   test("test 1-col partitioned + 1-col bucketed table") {
     withSQLConf(
       ("spark.gluten.sql.native.writer.enabled", "true"),
-      ("spark.gluten.enabled", "true")) {
+      (GlutenConfig.GLUTEN_ENABLED.key, "true")) {
 
       val fields: ListMap[String, String] = ListMap(
         ("string_field", "string"),
@@ -911,7 +911,7 @@ class GlutenClickHouseNativeWriteTableSuite
   test("GLUTEN-4316: fix crash on dynamic partition inserting") {
     withSQLConf(
       ("spark.gluten.sql.native.writer.enabled", "true"),
-      ("spark.gluten.enabled", "true")) {
+      (GlutenConfig.GLUTEN_ENABLED.key, "true")) {
       formats.foreach(
         format => {
           val tbl = "t_" + format
diff --git 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTPCHColumnarShuffleParquetAQESuite.scala
 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTPCHColumnarShuffleParquetAQESuite.scala
index 1b3929dcc..6caac9918 100644
--- 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTPCHColumnarShuffleParquetAQESuite.scala
+++ 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTPCHColumnarShuffleParquetAQESuite.scala
@@ -16,6 +16,7 @@
  */
 package org.apache.gluten.execution
 
+import org.apache.gluten.GlutenConfig
 import org.apache.gluten.extension.GlutenPlan
 
 import org.apache.spark.SparkConf
@@ -298,7 +299,7 @@ class GlutenClickHouseTPCHColumnarShuffleParquetAQESuite
   }
 
   test("Test 'spark.gluten.enabled' false") {
-    withSQLConf(("spark.gluten.enabled", "false")) {
+    withSQLConf((GlutenConfig.GLUTEN_ENABLED.key, "false")) {
       runTPCHQuery(2, noFallBack = false) {
         df =>
           val glutenPlans = collect(df.queryExecution.executedPlan) {
diff --git 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTPCHSaltNullParquetSuite.scala
 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTPCHSaltNullParquetSuite.scala
index a1bba300e..eec0ad874 100644
--- 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTPCHSaltNullParquetSuite.scala
+++ 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTPCHSaltNullParquetSuite.scala
@@ -16,6 +16,7 @@
  */
 package org.apache.gluten.execution
 
+import org.apache.gluten.GlutenConfig
 import org.apache.gluten.extension.GlutenPlan
 
 import org.apache.spark.{SparkConf, SparkException}
@@ -1231,7 +1232,7 @@ class GlutenClickHouseTPCHSaltNullParquetSuite extends 
GlutenClickHouseTPCHAbstr
   }
 
   test("Test 'spark.gluten.enabled' false") {
-    withSQLConf(("spark.gluten.enabled", "false")) {
+    withSQLConf((GlutenConfig.GLUTEN_ENABLED.key, "false")) {
       runTPCHQuery(2, noFallBack = false) {
         df =>
           val glutenPlans = df.queryExecution.executedPlan.collect {
diff --git 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTableAfterRestart.scala
 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTableAfterRestart.scala
index a673d4ba3..9e55df0fa 100644
--- 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTableAfterRestart.scala
+++ 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickHouseTableAfterRestart.scala
@@ -16,6 +16,8 @@
  */
 package org.apache.gluten.execution
 
+import org.apache.gluten.GlutenConfig
+
 import org.apache.spark.SparkConf
 import org.apache.spark.sql.SparkSession
 import org.apache.spark.sql.SparkSession.{getActiveSession, getDefaultSession}
@@ -248,9 +250,9 @@ class GlutenClickHouseTableAfterRestart
 
     restartSpark()
 
-    spark.sql("set spark.gluten.enabled=false")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=false")
     spark.sql("vacuum table_restart_vacuum")
-    spark.sql("set spark.gluten.enabled=true")
+    spark.sql(s"set ${GlutenConfig.GLUTEN_ENABLED.key}=true")
 
     assert(spark.sql("select count(*) from 
table_restart_vacuum").collect().apply(0).get(0) == 4)
   }
diff --git 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickhouseFunctionSuite.scala
 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickhouseFunctionSuite.scala
index c90e9131a..63e105e12 100644
--- 
a/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickhouseFunctionSuite.scala
+++ 
b/backends-clickhouse/src/test/scala/org/apache/gluten/execution/GlutenClickhouseFunctionSuite.scala
@@ -121,7 +121,7 @@ class GlutenClickhouseFunctionSuite extends 
GlutenClickHouseTPCHAbstractSuite {
   test("test uuid - write and read") {
     withSQLConf(
       ("spark.gluten.sql.native.writer.enabled", "true"),
-      ("spark.gluten.enabled", "true")) {
+      (GlutenConfig.GLUTEN_ENABLED.key, "true")) {
 
       spark.sql("drop table if exists uuid_test")
       spark.sql("create table if not exists uuid_test (id string) stored as 
parquet")
diff --git 
a/backends-clickhouse/src/test/scala/org/apache/gluten/s3/S3AuthSuite.scala 
b/backends-clickhouse/src/test/scala/org/apache/gluten/s3/S3AuthSuite.scala
index b2ac3672a..32c64c78a 100644
--- a/backends-clickhouse/src/test/scala/org/apache/gluten/s3/S3AuthSuite.scala
+++ b/backends-clickhouse/src/test/scala/org/apache/gluten/s3/S3AuthSuite.scala
@@ -102,7 +102,7 @@ class S3AuthSuite extends AnyFunSuite {
     }
 
     def withGluten(enable: Boolean): Builder = {
-      builder.config("spark.gluten.enabled", enable.toString)
+      builder.config(GlutenConfig.GLUTEN_ENABLED.key, enable.toString)
     }
   }
 
diff --git 
a/backends-clickhouse/src/test/scala/org/apache/spark/sql/execution/benchmarks/CHAggAndShuffleBenchmark.scala
 
b/backends-clickhouse/src/test/scala/org/apache/spark/sql/execution/benchmarks/CHAggAndShuffleBenchmark.scala
index 19ced9517..358b78518 100644
--- 
a/backends-clickhouse/src/test/scala/org/apache/spark/sql/execution/benchmarks/CHAggAndShuffleBenchmark.scala
+++ 
b/backends-clickhouse/src/test/scala/org/apache/spark/sql/execution/benchmarks/CHAggAndShuffleBenchmark.scala
@@ -16,6 +16,7 @@
  */
 package org.apache.spark.sql.execution.benchmarks
 
+import org.apache.gluten.GlutenConfig
 import org.apache.gluten.execution.{FileSourceScanExecTransformer, 
ProjectExecTransformer, WholeStageTransformer}
 import org.apache.gluten.sql.shims.SparkShimLoader
 
@@ -303,7 +304,7 @@ object CHAggAndShuffleBenchmark extends SqlBasedBenchmark 
with CHSqlBasedBenchma
       // Get the file partitions for generating the `FileScanRDD`
       val filePartitions = fileScan.getPartitions
         .map(_.asInstanceOf[FilePartition])
-      spark.conf.set("spark.gluten.enabled", "false")
+      spark.conf.set(GlutenConfig.GLUTEN_ENABLED.key, "false")
       val sparkExecutedPlan = allStages.queryExecution.executedPlan
 
       // Get the `FileSourceScanExec`
diff --git 
a/backends-clickhouse/src/test/scala/org/apache/spark/sql/execution/benchmarks/CHParquetReadBenchmark.scala
 
b/backends-clickhouse/src/test/scala/org/apache/spark/sql/execution/benchmarks/CHParquetReadBenchmark.scala
index 1cc8c8383..5e802eeed 100644
--- 
a/backends-clickhouse/src/test/scala/org/apache/spark/sql/execution/benchmarks/CHParquetReadBenchmark.scala
+++ 
b/backends-clickhouse/src/test/scala/org/apache/spark/sql/execution/benchmarks/CHParquetReadBenchmark.scala
@@ -16,6 +16,7 @@
  */
 package org.apache.spark.sql.execution.benchmarks
 
+import org.apache.gluten.GlutenConfig
 import org.apache.gluten.backendsapi.BackendsApiManager
 import org.apache.gluten.execution.{FileSourceScanExecTransformer, 
WholeStageTransformContext}
 import org.apache.gluten.expression.ConverterUtils
@@ -189,7 +190,7 @@ object CHParquetReadBenchmark extends SqlBasedBenchmark 
with CHSqlBasedBenchmark
     }
 
     if (executedVanilla) {
-      spark.conf.set("spark.gluten.enabled", "false")
+      spark.conf.set(GlutenConfig.GLUTEN_ENABLED.key, "false")
 
       val vanillaParquet = spark.sql(s"""
                                         |select $scanSchema from 
parquet.`$parquetDir`
diff --git 
a/gluten-core/src/test/scala/org/apache/gluten/execution/WholeStageTransformerSuite.scala
 
b/gluten-core/src/test/scala/org/apache/gluten/execution/WholeStageTransformerSuite.scala
index 8e8743857..5f60de27b 100644
--- 
a/gluten-core/src/test/scala/org/apache/gluten/execution/WholeStageTransformerSuite.scala
+++ 
b/gluten-core/src/test/scala/org/apache/gluten/execution/WholeStageTransformerSuite.scala
@@ -16,6 +16,7 @@
  */
 package org.apache.gluten.execution
 
+import org.apache.gluten.GlutenConfig
 import org.apache.gluten.extension.GlutenPlan
 import org.apache.gluten.test.FallbackUtil
 import org.apache.gluten.utils.Arm
@@ -320,7 +321,7 @@ abstract class WholeStageTransformerSuite
       noFallBack)
 
   protected def vanillaSparkConfs(): Seq[(String, String)] = {
-    List(("spark.gluten.enabled", "false"))
+    List((GlutenConfig.GLUTEN_ENABLED.key, "false"))
   }
 
   protected def checkDataFrame(
diff --git 
a/gluten-ut/spark32/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
 
b/gluten-ut/spark32/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
index ad08318bb..ef9699458 100644
--- 
a/gluten-ut/spark32/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
+++ 
b/gluten-ut/spark32/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
@@ -165,7 +165,7 @@ object ParquetReadBenchmark extends SqlBasedBenchmark {
     }
 
     if (executedVanilla) {
-      spark.conf.set("spark.gluten.enabled", "false")
+      spark.conf.set(GlutenConfig.GLUTEN_ENABLED.key, "false")
 
       val vanillaParquet = spark.sql(s"""
                                         |select $scanSchema from 
parquet.`$parquetDir`
diff --git 
a/gluten-ut/spark32/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
 
b/gluten-ut/spark32/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
index 18a9f5c9f..5e6c66265 100644
--- 
a/gluten-ut/spark32/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
+++ 
b/gluten-ut/spark32/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
@@ -16,6 +16,8 @@
  */
 package org.apache.spark.sql.extension
 
+import org.apache.gluten.GlutenConfig
+
 import org.apache.spark.SparkConf
 import org.apache.spark.sql.GlutenSQLTestsTrait
 
@@ -32,7 +34,7 @@ class GlutenCustomerExtensionSuite extends 
GlutenSQLTestsTrait {
   }
 
   testGluten("test customer column rules") {
-    withSQLConf(("spark.gluten.enabled", "false")) {
+    withSQLConf((GlutenConfig.GLUTEN_ENABLED.key, "false")) {
       sql("create table my_parquet(id int) using parquet")
       sql("insert into my_parquet values (1)")
       sql("insert into my_parquet values (2)")
diff --git 
a/gluten-ut/spark33/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
 
b/gluten-ut/spark33/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
index 7d8a29204..f61f7ebba 100644
--- 
a/gluten-ut/spark33/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
+++ 
b/gluten-ut/spark33/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
@@ -165,7 +165,7 @@ object ParquetReadBenchmark extends SqlBasedBenchmark {
     }
 
     if (executedVanilla) {
-      spark.conf.set("spark.gluten.enabled", "false")
+      spark.conf.set(GlutenConfig.GLUTEN_ENABLED.key, "false")
 
       val vanillaParquet = spark.sql(s"""
                                         |select $scanSchema from 
parquet.`$parquetDir`
diff --git 
a/gluten-ut/spark33/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
 
b/gluten-ut/spark33/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
index 18a9f5c9f..5e6c66265 100644
--- 
a/gluten-ut/spark33/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
+++ 
b/gluten-ut/spark33/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
@@ -16,6 +16,8 @@
  */
 package org.apache.spark.sql.extension
 
+import org.apache.gluten.GlutenConfig
+
 import org.apache.spark.SparkConf
 import org.apache.spark.sql.GlutenSQLTestsTrait
 
@@ -32,7 +34,7 @@ class GlutenCustomerExtensionSuite extends 
GlutenSQLTestsTrait {
   }
 
   testGluten("test customer column rules") {
-    withSQLConf(("spark.gluten.enabled", "false")) {
+    withSQLConf((GlutenConfig.GLUTEN_ENABLED.key, "false")) {
       sql("create table my_parquet(id int) using parquet")
       sql("insert into my_parquet values (1)")
       sql("insert into my_parquet values (2)")
diff --git 
a/gluten-ut/spark34/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
 
b/gluten-ut/spark34/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
index b5481f4d8..471bdf179 100644
--- 
a/gluten-ut/spark34/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
+++ 
b/gluten-ut/spark34/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
@@ -166,7 +166,7 @@ object ParquetReadBenchmark extends SqlBasedBenchmark {
     }
 
     if (executedVanilla) {
-      spark.conf.set("spark.gluten.enabled", "false")
+      spark.conf.set(GlutenConfig.GLUTEN_ENABLED.key, "false")
 
       val vanillaParquet = spark.sql(s"""
                                         |select $scanSchema from 
parquet.`$parquetDir`
diff --git 
a/gluten-ut/spark34/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
 
b/gluten-ut/spark34/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
index 18a9f5c9f..5e6c66265 100644
--- 
a/gluten-ut/spark34/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
+++ 
b/gluten-ut/spark34/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
@@ -16,6 +16,8 @@
  */
 package org.apache.spark.sql.extension
 
+import org.apache.gluten.GlutenConfig
+
 import org.apache.spark.SparkConf
 import org.apache.spark.sql.GlutenSQLTestsTrait
 
@@ -32,7 +34,7 @@ class GlutenCustomerExtensionSuite extends 
GlutenSQLTestsTrait {
   }
 
   testGluten("test customer column rules") {
-    withSQLConf(("spark.gluten.enabled", "false")) {
+    withSQLConf((GlutenConfig.GLUTEN_ENABLED.key, "false")) {
       sql("create table my_parquet(id int) using parquet")
       sql("insert into my_parquet values (1)")
       sql("insert into my_parquet values (2)")
diff --git 
a/gluten-ut/spark35/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
 
b/gluten-ut/spark35/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
index 7d8a29204..f61f7ebba 100644
--- 
a/gluten-ut/spark35/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
+++ 
b/gluten-ut/spark35/src/test/scala/org/apache/spark/sql/execution/benchmarks/ParquetReadBenchmark.scala
@@ -165,7 +165,7 @@ object ParquetReadBenchmark extends SqlBasedBenchmark {
     }
 
     if (executedVanilla) {
-      spark.conf.set("spark.gluten.enabled", "false")
+      spark.conf.set(GlutenConfig.GLUTEN_ENABLED.key, "false")
 
       val vanillaParquet = spark.sql(s"""
                                         |select $scanSchema from 
parquet.`$parquetDir`
diff --git 
a/gluten-ut/spark35/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
 
b/gluten-ut/spark35/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
index 18a9f5c9f..5e6c66265 100644
--- 
a/gluten-ut/spark35/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
+++ 
b/gluten-ut/spark35/src/test/scala/org/apache/spark/sql/extension/GlutenCustomerExtensionSuite.scala
@@ -16,6 +16,8 @@
  */
 package org.apache.spark.sql.extension
 
+import org.apache.gluten.GlutenConfig
+
 import org.apache.spark.SparkConf
 import org.apache.spark.sql.GlutenSQLTestsTrait
 
@@ -32,7 +34,7 @@ class GlutenCustomerExtensionSuite extends 
GlutenSQLTestsTrait {
   }
 
   testGluten("test customer column rules") {
-    withSQLConf(("spark.gluten.enabled", "false")) {
+    withSQLConf((GlutenConfig.GLUTEN_ENABLED.key, "false")) {
       sql("create table my_parquet(id int) using parquet")
       sql("insert into my_parquet values (1)")
       sql("insert into my_parquet values (2)")


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to