Ángel, https://spark.apache.org/docs/latest/configuration.html search through `decommission` in this page.
The config you may have found from Synapse is spark."yarn".executor.decommission.enabled. And the question was even "k8s" and none of the information for the vendor was mentioned from the question. I don't even think these configs are on the internet - simply saying, google them (be sure to wrap with double quotes). Interestingly, I wouldn't even expect the config for graceful shutdown for decommission. The functionality Spark provides for "decommission" is basically a "graceful shutdown" of the executor. It sounds to me as redundant. On Thu, Oct 10, 2024 at 11:11 AM Ángel <angel.alvarez.pas...@gmail.com> wrote: > Looks like it actually exists ... but only for the Spark Synapse > implementation ... > > > https://learn.microsoft.com/en-us/answers/questions/1496283/purpose-of-spark-yarn-executor-decommission-enable > > > Jay Han was asking for some config on k8s, so .... we shouldn't bring this > config to the table, should we? > > El jue, 10 oct 2024 a las 2:55, Sean Owen (<sro...@gmail.com>) escribió: > >> Mich: you can set any key-value pair you want in Spark config. It doesn't >> mean it is a real flag that code reads. >> >> spark.conf.set("ham", "sandwich") >> print(spark.conf.get("ham")) >> >> prints "sandwich" >> >> forceKillTimeout is a real config: >> >> https://github.com/apache/spark/blob/fed9a8da3d4187794161e0be325aa96be8487783/core/src/main/scala/org/apache/spark/internal/config/package.scala#L2394 >> >> The others I cannot find, as in: >> >> https://github.com/search?q=repo%3Aapache%2Fspark%20spark.executor.decommission.gracefulShutdown&type=code >> >> If you're continuing to suggest these are real configs, where are you >> finding those two in any docs or source? >> Or what config were you thinking of if it's a typo? >> >> On Wed, Oct 9, 2024 at 5:14 PM Mich Talebzadeh <mich.talebza...@gmail.com> >> wrote: >> >>> Let us take this for a ride using these so called non-existent >>> configuration settings >>> >>> spark.executor.decommission.enabled=true >>> spark.executor.decommission.gracefulShutdown=true >>> >>> Tested on Spark 3.4 >>> >>> from pyspark.sql import SparkSession >>> # Initialize a Spark session >>> spark = SparkSession.builder \ >>> .appName("Verifying Spark Configurations") \ >>> .config("spark.executor.decommission.enabled", "true") \ >>> .config("spark.executor.decommission.forceKillTimeout", "100s") \ >>> .getOrCreate() >>> >>> # Access Spark context >>> sc = spark.sparkContext >>> # Set the log level to ERROR to reduce verbosity >>> sc.setLogLevel("ERROR") >>> print(f"\n\nSpark version: ", sc.version) >>> >>> # Verify the configuration for executor decommissioning >>> decommission_enabled = >>> sc.getConf().get("spark.executor.decommission.enabled", "false") >>> force_kill_timeout = >>> sc.getConf().get("spark.executor.decommission.forceKillTimeout", >>> "default_value") >>> >>> # Print the values >>> print(f"spark.executor.decommission.enabled: {decommission_enabled}") >>> print(f"spark.executor.decommission.forceKillTimeout: >>> {force_kill_timeout}") >>> >>> The output >>> >>> Spark version: 3.4.0 >>> spark.executor.decommission.enabled: true >>> spark.executor.decommission.forceKillTimeout: 100s >>> >>> By creating a simple Spark application and verifying the configuration >>> values, I trust it is shown that these two parameters are valid and are >>> applied by Spark >>> >>> HTH >>> >>> Mich Talebzadeh, >>> >>> Architect | Data Engineer | Data Science | Financial Crime >>> PhD <https://en.wikipedia.org/wiki/Doctor_of_Philosophy> Imperial >>> College London <https://en.wikipedia.org/wiki/Imperial_College_London> >>> London, United Kingdom >>> >>> >>> view my Linkedin profile >>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/> >>> >>> >>> https://en.everybodywiki.com/Mich_Talebzadeh >>> >>> >>> >>> *Disclaimer:* The information provided is correct to the best of my >>> knowledge but of course cannot be guaranteed . It is essential to note >>> that, as with any advice, quote "one test result is worth one-thousand >>> expert opinions (Werner >>> <https://en.wikipedia.org/wiki/Wernher_von_Braun>Von Braun >>> <https://en.wikipedia.org/wiki/Wernher_von_Braun>)". >>> >>> >>> On Wed, 9 Oct 2024 at 16:51, Mich Talebzadeh <mich.talebza...@gmail.com> >>> wrote: >>> >>>> Do you have a better recommendation? >>>> >>>> Or trying to waste time as usual. >>>> >>>> It is far easier to throw than catch. >>>> >>>> Do your homework and stop throwing spanners at work. >>>> >>>> Mich Talebzadeh, >>>> >>>> Architect | Data Engineer | Data Science | Financial Crime >>>> PhD <https://en.wikipedia.org/wiki/Doctor_of_Philosophy> Imperial >>>> College London <https://en.wikipedia.org/wiki/Imperial_College_London> >>>> London, United Kingdom >>>> >>>> >>>> view my Linkedin profile >>>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/> >>>> >>>> >>>> https://en.everybodywiki.com/Mich_Talebzadeh >>>> >>>> >>>> >>>> *Disclaimer:* The information provided is correct to the best of my >>>> knowledge but of course cannot be guaranteed . It is essential to note >>>> that, as with any advice, quote "one test result is worth one-thousand >>>> expert opinions (Werner >>>> <https://en.wikipedia.org/wiki/Wernher_von_Braun>Von Braun >>>> <https://en.wikipedia.org/wiki/Wernher_von_Braun>)". >>>> >>>> >>>> On Wed, 9 Oct 2024 at 16:43, Nicholas Chammas < >>>> nicholas.cham...@gmail.com> wrote: >>>> >>>>> Mich, >>>>> >>>>> Can you please share with the list where *exactly* you are citing >>>>> these configs from? >>>>> >>>>> As far as I can tell, these two configs don’t exist and have never >>>>> existed in the Spark codebase: >>>>> >>>>> spark.executor.decommission.enabled=true >>>>> spark.executor.decommission.gracefulShutdown=true >>>>> >>>>> Where exactly are you getting this information from (and then posting >>>>> it to the list as advice)? Please be clear and provide specific >>>>> references. >>>>> >>>>> Nick >>>>> >>>>> >>>>> On Oct 9, 2024, at 1:20 PM, Mich Talebzadeh <mich.talebza...@gmail.com> >>>>> wrote: >>>>> >>>>> Before responding, what configuration parameters are you using to make >>>>> this work? >>>>> >>>>> spark.executor.decommission.enabled=true >>>>> spark.executor.decommission.gracefulShutdown=true >>>>> spark.executor.decommission.forceKillTimeout=100s >>>>> >>>>> HTH >>>>> >>>>> Mich Talebzadeh, >>>>> >>>>> Architect | Data Engineer | Data Science | Financial Crime >>>>> PhD <https://en.wikipedia.org/wiki/Doctor_of_Philosophy> Imperial >>>>> College London <https://en.wikipedia.org/wiki/Imperial_College_London> >>>>> >>>>> London, United Kingdom >>>>> >>>>> view my Linkedin profile >>>>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/> >>>>> >>>>> >>>>> https://en.everybodywiki.com/Mich_Talebzadeh >>>>> >>>>> >>>>> >>>>> *Disclaimer:* The information provided is correct to the best of my >>>>> knowledge but of course cannot be guaranteed . It is essential to note >>>>> that, as with any advice, quote "one test result is worth one-thousand >>>>> expert opinions (Werner >>>>> <https://en.wikipedia.org/wiki/Wernher_von_Braun>Von Braun >>>>> <https://en.wikipedia.org/wiki/Wernher_von_Braun>)". >>>>> >>>>> >>>>> On Wed, 9 Oct 2024 at 11:05, Jay Han <tunyu...@gmail.com> wrote: >>>>> >>>>>> Hi spark community, >>>>>> I have such a question: Why driver doesn't shutdown executors >>>>>> gracefully on k8s. For instance, >>>>>> kubernetesClient.pods().withGracePeriod(100).delete(). >>>>>> >>>>>> >>>>>> -- >>>>>> Best, >>>>>> Jay >>>>>> >>>>> >>>>>