[
https://issues.apache.org/jira/browse/SPARK-26389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16750667#comment-16750667
]
Fengyu Cao commented on SPARK-26389:
------------------------------------
a force clean-up flag maybe help if not use hdfs
the size of temp checkpints out of hdfs is not acceptable for me:
# nginx log group by uid (1h window, 5min slide window)
# run about 4 hours on 2 executor hosts (default trigger)
# more than 1gb on each host
seems hdfs state store clean logic not work well on non-hdfs file system(xfs)
thanks anyway, shoud I close this issue or change type/prioriy?
> temp checkpoint folder at executor should be deleted on graceful shutdown
> -------------------------------------------------------------------------
>
> Key: SPARK-26389
> URL: https://issues.apache.org/jira/browse/SPARK-26389
> Project: Spark
> Issue Type: Bug
> Components: Structured Streaming
> Affects Versions: 2.4.0
> Reporter: Fengyu Cao
> Priority: Major
>
> {{spark-submit --master mesos://<mesos> -conf
> spark.streaming.stopGracefullyOnShutdown=true <structured streaming
> framework>}}
> CTRL-C, framework shutdown
> {{18/12/18 10:27:36 ERROR MicroBatchExecution: Query [id =
> f512e17a-df88-4414-a5cd-a23550cf1e7f, runId =
> 24d99723-8d61-48c0-beab-af432f7a19d3] terminated with error
> org.apache.spark.SparkException: Writing job aborted.}}
> {{/tmp/temporary-<uuid> on executor not deleted due to
> org.apache.spark.SparkException: Writing job aborted., and this temp
> checkpoint can't used to recovery.}}
>
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]