[ 
https://issues.apache.org/jira/browse/SPARK-18898?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15755475#comment-15755475
 ] 

Sean Owen commented on SPARK-18898:
-----------------------------------

Probably the same as SPARK-15955

> Exception not failing Scala applications (in yarn)
> --------------------------------------------------
>
>                 Key: SPARK-18898
>                 URL: https://issues.apache.org/jira/browse/SPARK-18898
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Submit, YARN
>    Affects Versions: 2.0.2
>            Reporter: Raphael
>
> I am submitting my scala applications with SparkLauncher which uses 
> spark-submit project.
> When I throw some error in my spark job, the final status of this job in YARN 
> is FINISHED, after viewing the source code of SparkSubmit:
> {code:title=SparkSubmit.scala|borderStyle=solid}
> ...
> try {
>       mainMethod.invoke(null, childArgs.toArray)
>     } catch {
>       case t: Throwable =>
>         findCause(t) match {
>           case SparkUserAppException(exitCode) =>
>             System.exit(exitCode)
>           case t: Throwable =>
>             throw t
>         }
>     }
> ...
> {code}
>  
> Its seems we have in our code to throw the SparkUserAppException, but this 
> exception is a private case class inside the SparkException class.
> Please make this case class avaliable or give us a way to launch errors 
> inside our applications.
> More details at: 
> http://stackoverflow.com/questions/41184158/how-to-throw-an-exception-in-spark
> In the past, the same issue with pyspark was opened here:
> https://issues.apache.org/jira/browse/SPARK-7736
> And resolved here:
> https://github.com/apache/spark/pull/8258
> Best Regards 
> Raphael.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to