[ 
https://issues.apache.org/jira/browse/SPARK-11058?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16219282#comment-16219282
 ] 

Marcelo Vanzin commented on SPARK-11058:
----------------------------------------

Nobody has expressed enough interest in changing Spark's public API to add the 
necessary information, and personally I don't think it's worth it if the only 
thing it would fix is this.

It would also require changes in every Spark application to actually fix this, 
which makes it even less interesting.

> failed spark job reports on YARN as successful
> ----------------------------------------------
>
>                 Key: SPARK-11058
>                 URL: https://issues.apache.org/jira/browse/SPARK-11058
>             Project: Spark
>          Issue Type: Bug
>          Components: YARN
>    Affects Versions: 1.3.0
>         Environment: CDH 5.4
>            Reporter: Lan Jiang
>            Priority: Minor
>
> I have a spark batch job running on CDH5.4 + Spark 1.3.0. Job is submitted in 
> “yarn-client” mode. The job itself failed due to YARN kills several executor 
> containers because the containers exceeded the memory limit posed by YARN. 
> However, when I went to the YARN resource manager site, it displayed the job 
> as successful. I found there was an issue reported in JIRA 
> https://issues.apache.org/jira/browse/SPARK-3627, but it says it was fixed in 
> Spark 1.2. On Spark history server, it shows the job as “Incomplete”. 



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to