[ 
https://issues.apache.org/jira/browse/SPARK-26389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16750223#comment-16750223
 ] 

Gabor Somogyi commented on SPARK-26389:
---------------------------------------

Good to hear with HDFS it's working.

Prohibiting the user from using temp checkpoints when their frameworks are 
non-local is not a good idea because for example console sink uses that.
I think it's a valid use-case to print to console even in cluster mode with 
several nodes.

What we can do is to add a message which explains that temp checkpoint is used 
and its consequences.
Maybe a force clean-up flag for temp checkpoints can be introduced but have to 
think about it a bit more...


> temp checkpoint folder at executor should be deleted on graceful shutdown
> -------------------------------------------------------------------------
>
>                 Key: SPARK-26389
>                 URL: https://issues.apache.org/jira/browse/SPARK-26389
>             Project: Spark
>          Issue Type: Bug
>          Components: Structured Streaming
>    Affects Versions: 2.4.0
>            Reporter: Fengyu Cao
>            Priority: Major
>
> {{spark-submit --master mesos://<mesos> -conf 
> spark.streaming.stopGracefullyOnShutdown=true <structured streaming 
> framework>}}
> CTRL-C, framework shutdown
> {{18/12/18 10:27:36 ERROR MicroBatchExecution: Query [id = 
> f512e17a-df88-4414-a5cd-a23550cf1e7f, runId = 
> 24d99723-8d61-48c0-beab-af432f7a19d3] terminated with error 
> org.apache.spark.SparkException: Writing job aborted.}}
> {{/tmp/temporary-<uuid> on executor not deleted due to 
> org.apache.spark.SparkException: Writing job aborted., and this temp 
> checkpoint can't used to recovery.}}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to