[ https://issues.apache.org/jira/browse/SPARK-40309?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Xinrong Meng deleted SPARK-40309: --------------------------------- > Introduce sql_conf context manager for pyspark.sql > -------------------------------------------------- > > Key: SPARK-40309 > URL: https://issues.apache.org/jira/browse/SPARK-40309 > Project: Spark > Issue Type: Sub-task > Reporter: Xinrong Meng > Priority: Major > Labels: release-notes > > That would simplify the control of Spark SQL configuration as below > from > {code:java} > original_value = spark.conf.get("key") > spark.conf.set("key", "value") > ... > spark.conf.set("key", original_value){code} > to > {code:java} > with sql_conf({"key": "value"}): > ... > {code} > [Here|https://github.com/apache/spark/blob/master/python/pyspark/pandas/utils.py#L490] > is such a context manager is in Pandas API on Spark. > We should introduce one in `pyspark.sql`, and deduplicate code if possible. > -- This message was sent by Atlassian Jira (v8.20.10#820010) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org