[ https://issues.apache.org/jira/browse/SPARK-5869?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Sean Owen resolved SPARK-5869. ------------------------------ Resolution: Duplicate Just fixed as a follow on to SPARK-5841 > Exception when deleting Spark local dirs when shutting down DiskBlockManager > ---------------------------------------------------------------------------- > > Key: SPARK-5869 > URL: https://issues.apache.org/jira/browse/SPARK-5869 > Project: Spark > Issue Type: Bug > Affects Versions: 1.3.0 > Reporter: Kai Zeng > > Running Spark on a ec2 cluster deployed using spark-ec2 scripts. > Got this error when applications are finishing: > 15/02/17 19:23:44 ERROR util.Utils: Uncaught exception in thread delete Spark > local dirs > java.lang.IllegalStateException: Shutdown in progress > at > java.lang.ApplicationShutdownHooks.remove(ApplicationShutdownHooks.java:82) > at java.lang.Runtime.removeShutdownHook(Runtime.java:239) > at > org.apache.spark.storage.DiskBlockManager.stop(DiskBlockManager.scala:151) > at > org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply$mcV$sp(DiskBlockManager.scala:141) > at > org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply(DiskBlockManager.scala:139) > at > org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply(DiskBlockManager.scala:139) > at > org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1613) > at > org.apache.spark.storage.DiskBlockManager$$anon$1.run(DiskBlockManager.scala:139) > Exception in thread "delete Spark local dirs" > java.lang.IllegalStateException: Shutdown in progress > at > java.lang.ApplicationShutdownHooks.remove(ApplicationShutdownHooks.java:82) > at java.lang.Runtime.removeShutdownHook(Runtime.java:239) > at > org.apache.spark.storage.DiskBlockManager.stop(DiskBlockManager.scala:151) > at > org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply$mcV$sp(DiskBlockManager.scala:141) > at > org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply(DiskBlockManager.scala:139) > at > org.apache.spark.storage.DiskBlockManager$$anon$1$$anonfun$run$1.apply(DiskBlockManager.scala:139) > at > org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1613) > at > org.apache.spark.storage.DiskBlockManager$$anon$1.run(DiskBlockManager.scala:139) > It seems to have this issue after commit "SPARK-5841: remove DiskBlockManager > shutdown hook on stop" -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org