Hi, All. I'd like to highlight this discussion because this is more important and tricky in a way.
As already mentioned in the mailing list and PRs, there was an obvious mistake which missed an improper configuration name, `spark.databricks.*`. https://github.com/apache/spark/blob/a6f220d951742f4074b37772485ee0ec7a774e7d/sql/catalyst/src/main/scala/org/apache/spark/sql/internal/SQLConf.scala#L3424 `spark.databricks.sql.optimizer.pruneFiltersCanPruneStreamingSubplan` In fact, Apache Spark committers have been preventing this repetitive mistake pattern during the review stages successfully until we slip the following backportings at Apache Spark 3.5.4. https://github.com/apache/spark/pull/45649 https://github.com/apache/spark/pull/48149 https://github.com/apache/spark/pull/49121 At this point of writing, `spark.databricks.*` was removed successfully from `master` and `branch-4.0` and a new ScalaStyle rule was added to protect Apache Spark repository from future mistakes. SPARK-51172 Rename to spark.sql.optimizer.pruneFiltersCanPruneStreamingSubplan SPARK-51173 Add `configName` Scalastyle rule What I proposed is to release Apache Spark 3.5.5 next week with the deprecation in order to make Apache Spark 4.0 be free of `spark.databricks.*` configuration. Apache Spark 3.5.5 (2025 February, with deprecation warning with alternative) Apache Spark 4.0.0 (2025 March, without `spark.databricks.*` config) In addition, I'd like to volunteer as a release manager of Apache Spark 3.5.5 for a swift release. WDYT? FYI, `branch-3.5` has 37 patches currently. $ git log --oneline v3.5.4..HEAD | wc -l 37 Best Regards, Dongjoon.