[ 
https://issues.apache.org/jira/browse/SPARK-5869?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean Owen resolved SPARK-5869.
------------------------------
    Resolution: Duplicate

Just fixed as a follow on to SPARK-5841

> Exception when deleting Spark local dirs when shutting down DiskBlockManager
> ----------------------------------------------------------------------------
>
>                 Key: SPARK-5869
>                 URL: https://issues.apache.org/jira/browse/SPARK-5869
>             Project: Spark
>          Issue Type: Bug
>    Affects Versions: 1.3.0
>            Reporter: Kai Zeng
>
> Running Spark on a ec2 cluster deployed using spark-ec2 scripts.
> Got this error when applications are finishing:
> 15/02/17 19:23:44 ERROR util.Utils: Uncaught exception in thread delete Spark 
> local dirs
> java.lang.IllegalStateException: Shutdown in progress
>         at 
> java.lang.ApplicationShutdownHooks.remove(ApplicationShutdownHooks.java:82)
>         at java.lang.Runtime.removeShutdownHook(Runtime.java:239)
>         at 
> org.apache.spark.storage.DiskBlockManager.stop(DiskBlockManager.scala:151)
>         at 
> org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply$mcV$sp(DiskBlockManager.scala:141)
>         at 
> org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply(DiskBlockManager.scala:139)
>         at 
> org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply(DiskBlockManager.scala:139)
>         at 
> org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1613)
>         at 
> org.apache.spark.storage.DiskBlockManager$$anon$1.run(DiskBlockManager.scala:139)
> Exception in thread "delete Spark local dirs" 
> java.lang.IllegalStateException: Shutdown in progress
>         at 
> java.lang.ApplicationShutdownHooks.remove(ApplicationShutdownHooks.java:82)
>         at java.lang.Runtime.removeShutdownHook(Runtime.java:239)
>         at 
> org.apache.spark.storage.DiskBlockManager.stop(DiskBlockManager.scala:151)
>         at 
> org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply$mcV$sp(DiskBlockManager.scala:141)
>         at 
> org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply(DiskBlockManager.scala:139)
>         at 
> org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply(DiskBlockManager.scala:139)
>         at 
> org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1613)
>         at 
> org.apache.spark.storage.DiskBlockManager$$anon$1.run(DiskBlockManager.scala:139)
> It seems to have this issue after commit "SPARK-5841: remove DiskBlockManager 
> shutdown hook on stop"



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to