Let us take this for a ride using these so called non-existent
configuration settings
spark.executor.decommission.enabled=true
spark.executor.decommission.gracefulShutdown=true
Tested on Spark 3.4
from pyspark.sql import SparkSession
# Initialize a Spark session
spark = SparkSession.builder \
.appName("Verifying Spark Configurations") \
.config("spark.executor.decommission.enabled", "true") \
.config("spark.executor.decommission.forceKillTimeout", "100s") \
.getOrCreate()
# Access Spark context
sc = spark.sparkContext
# Set the log level to ERROR to reduce verbosity
sc.setLogLevel("ERROR")
print(f"\n\nSpark version: ", sc.version)
# Verify the configuration for executor decommissioning
decommission_enabled =
sc.getConf().get("spark.executor.decommission.enabled", "false")
force_kill_timeout =
sc.getConf().get("spark.executor.decommission.forceKillTimeout",
"default_value")
# Print the values
print(f"spark.executor.decommission.enabled: {decommission_enabled}")
print(f"spark.executor.decommission.forceKillTimeout: {force_kill_timeout}")
The output
Spark version: 3.4.0
spark.executor.decommission.enabled: true
spark.executor.decommission.forceKillTimeout: 100s
By creating a simple Spark application and verifying the configuration
values, I trust it is shown that these two parameters are valid and are
applied by Spark
HTH
Mich Talebzadeh,
Architect | Data Engineer | Data Science | Financial Crime
PhD <https://en.wikipedia.org/wiki/Doctor_of_Philosophy> Imperial College
London <https://en.wikipedia.org/wiki/Imperial_College_London>
London, United Kingdom
view my Linkedin profile
<https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
https://en.everybodywiki.com/Mich_Talebzadeh
*Disclaimer:* The information provided is correct to the best of my
knowledge but of course cannot be guaranteed . It is essential to note
that, as with any advice, quote "one test result is worth one-thousand
expert opinions (Werner <https://en.wikipedia.org/wiki/Wernher_von_Braun>Von
Braun <https://en.wikipedia.org/wiki/Wernher_von_Braun>)".
On Wed, 9 Oct 2024 at 16:51, Mich Talebzadeh <[email protected]>
wrote:
> Do you have a better recommendation?
>
> Or trying to waste time as usual.
>
> It is far easier to throw than catch.
>
> Do your homework and stop throwing spanners at work.
>
> Mich Talebzadeh,
>
> Architect | Data Engineer | Data Science | Financial Crime
> PhD <https://en.wikipedia.org/wiki/Doctor_of_Philosophy> Imperial College
> London <https://en.wikipedia.org/wiki/Imperial_College_London>
> London, United Kingdom
>
>
> view my Linkedin profile
> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>
>
> https://en.everybodywiki.com/Mich_Talebzadeh
>
>
>
> *Disclaimer:* The information provided is correct to the best of my
> knowledge but of course cannot be guaranteed . It is essential to note
> that, as with any advice, quote "one test result is worth one-thousand
> expert opinions (Werner <https://en.wikipedia.org/wiki/Wernher_von_Braun>Von
> Braun <https://en.wikipedia.org/wiki/Wernher_von_Braun>)".
>
>
> On Wed, 9 Oct 2024 at 16:43, Nicholas Chammas <[email protected]>
> wrote:
>
>> Mich,
>>
>> Can you please share with the list where *exactly* you are citing these
>> configs from?
>>
>> As far as I can tell, these two configs don’t exist and have never
>> existed in the Spark codebase:
>>
>> spark.executor.decommission.enabled=true
>> spark.executor.decommission.gracefulShutdown=true
>>
>> Where exactly are you getting this information from (and then posting it
>> to the list as advice)? Please be clear and provide specific references.
>>
>> Nick
>>
>>
>> On Oct 9, 2024, at 1:20 PM, Mich Talebzadeh <[email protected]>
>> wrote:
>>
>> Before responding, what configuration parameters are you using to make
>> this work?
>>
>> spark.executor.decommission.enabled=true
>> spark.executor.decommission.gracefulShutdown=true
>> spark.executor.decommission.forceKillTimeout=100s
>>
>> HTH
>>
>> Mich Talebzadeh,
>>
>> Architect | Data Engineer | Data Science | Financial Crime
>> PhD <https://en.wikipedia.org/wiki/Doctor_of_Philosophy> Imperial
>> College London <https://en.wikipedia.org/wiki/Imperial_College_London>
>> London, United Kingdom
>>
>> view my Linkedin profile
>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>
>>
>> https://en.everybodywiki.com/Mich_Talebzadeh
>>
>>
>>
>> *Disclaimer:* The information provided is correct to the best of my
>> knowledge but of course cannot be guaranteed . It is essential to note
>> that, as with any advice, quote "one test result is worth one-thousand
>> expert opinions (Werner
>> <https://en.wikipedia.org/wiki/Wernher_von_Braun>Von Braun
>> <https://en.wikipedia.org/wiki/Wernher_von_Braun>)".
>>
>>
>> On Wed, 9 Oct 2024 at 11:05, Jay Han <[email protected]> wrote:
>>
>>> Hi spark community,
>>> I have such a question: Why driver doesn't shutdown executors
>>> gracefully on k8s. For instance,
>>> kubernetesClient.pods().withGracePeriod(100).delete().
>>>
>>>
>>> --
>>> Best,
>>> Jay
>>>
>>
>>