[
https://issues.apache.org/jira/browse/SPARK-26389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16750223#comment-16750223
]
Gabor Somogyi commented on SPARK-26389:
---------------------------------------
Good to hear with HDFS it's working.
Prohibiting the user from using temp checkpoints when their frameworks are
non-local is not a good idea because for example console sink uses that.
I think it's a valid use-case to print to console even in cluster mode with
several nodes.
What we can do is to add a message which explains that temp checkpoint is used
and its consequences.
Maybe a force clean-up flag for temp checkpoints can be introduced but have to
think about it a bit more...
> temp checkpoint folder at executor should be deleted on graceful shutdown
> -------------------------------------------------------------------------
>
> Key: SPARK-26389
> URL: https://issues.apache.org/jira/browse/SPARK-26389
> Project: Spark
> Issue Type: Bug
> Components: Structured Streaming
> Affects Versions: 2.4.0
> Reporter: Fengyu Cao
> Priority: Major
>
> {{spark-submit --master mesos://<mesos> -conf
> spark.streaming.stopGracefullyOnShutdown=true <structured streaming
> framework>}}
> CTRL-C, framework shutdown
> {{18/12/18 10:27:36 ERROR MicroBatchExecution: Query [id =
> f512e17a-df88-4414-a5cd-a23550cf1e7f, runId =
> 24d99723-8d61-48c0-beab-af432f7a19d3] terminated with error
> org.apache.spark.SparkException: Writing job aborted.}}
> {{/tmp/temporary-<uuid> on executor not deleted due to
> org.apache.spark.SparkException: Writing job aborted., and this temp
> checkpoint can't used to recovery.}}
>
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]