Ruochen Zou created SPARK-50186:
-----------------------------------

             Summary: Remove Hardcoded OnOutOfMemoryError Setting in Executor 
Startup Script
                 Key: SPARK-50186
                 URL: https://issues.apache.org/jira/browse/SPARK-50186
             Project: Spark
          Issue Type: Improvement
          Components: Spark Core
    Affects Versions: 3.3.1, 4.0.0
            Reporter: Ruochen Zou


Currently, the Executor startup script hardcodes the addition of 
{{{}-XX:OnOutOfMemoryError='kill %p'{}}}, which causes the process to be killed 
when the Executor encounters an OOM (Out Of Memory) error. As a result, the 
YarnAllocator receives an exit code of 143 and is unable to accurately 
determine the reason for the Executor's termination based on this exit code. 
Moreover, the CoarseGrainedExecutorBackend cannot guarantee that StatusUpdate 
messages are sent to the Driver before the process is killed.
Could we remove this setting, since users can set it via the 
{{spark.executor.extraJavaOptions}} parameter if necessary?

Executor log:
!image-2024-10-31-14-02-10-261.png!
 
Driver log:
!image-2024-10-31-14-11-22-952.png!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to