[ https://issues.apache.org/jira/browse/SPARK-52536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Hyukjin Kwon reassigned SPARK-52536: ------------------------------------ Assignee: Zhen Wang > Specify AsyncProfilerLoader.extractionDir to spark local dir > ------------------------------------------------------------ > > Key: SPARK-52536 > URL: https://issues.apache.org/jira/browse/SPARK-52536 > Project: Spark > Issue Type: Improvement > Components: Spark Core > Affects Versions: 4.0.0 > Reporter: Zhen Wang > Assignee: Zhen Wang > Priority: Major > > AsyncProfilerLoader uses `user.home` by default to store the extracted > libraries: > [https://github.com/jvm-profiling-tools/ap-loader/blob/main/src/main/java/one/profiler/AsyncProfilerLoader.java#L139-L152] > The `user.home` directory of the datanodes in our yarn cluster was not > initialized, causing the executor startup to fail: > {code:java} > 25/06/20 11:54:26 ERROR YarnCoarseGrainedExecutorBackend: Executor > self-exiting due to : Unable to create executor due to /home/pilot > java.nio.file.AccessDeniedException: /home/pilot > at > sun.nio.fs.UnixException.translateToIOException(UnixException.java:84) > at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:102) > at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:107) > at > sun.nio.fs.UnixFileSystemProvider.createDirectory(UnixFileSystemProvider.java:384) > at java.nio.file.Files.createDirectory(Files.java:674) > at java.nio.file.Files.createAndCheckIsDirectory(Files.java:781) > at java.nio.file.Files.createDirectories(Files.java:767) > at > one.profiler.AsyncProfilerLoader.getExtractionDirectory(AsyncProfilerLoader.java:133) > at > one.profiler.AsyncProfilerLoader.getAsyncProfilerPath(AsyncProfilerLoader.java:562) > at one.profiler.AsyncProfilerLoader.load(AsyncProfilerLoader.java:861) > at > org.apache.spark.profiler.SparkAsyncProfiler.<init>(SparkAsyncProfiler.scala:70) > at > org.apache.spark.profiler.ProfilerExecutorPlugin.init(ProfilerPlugin.scala:82) > at > org.apache.spark.internal.plugin.ExecutorPluginContainer.$anonfun$executorPlugins$1(PluginContainer.scala:125) > at > scala.collection.TraversableLike.$anonfun$flatMap$1(TraversableLike.scala:293) > at > scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) > at > scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) > at scala.collection.TraversableLike.flatMap(TraversableLike.scala:293) > at scala.collection.TraversableLike.flatMap$(TraversableLike.scala:290) > at scala.collection.AbstractTraversable.flatMap(Traversable.scala:108) > at > org.apache.spark.internal.plugin.ExecutorPluginContainer.<init>(PluginContainer.scala:113) > at > org.apache.spark.internal.plugin.PluginContainer$.apply(PluginContainer.scala:211) > at > org.apache.spark.internal.plugin.PluginContainer$.apply(PluginContainer.scala:199) > at > org.apache.spark.executor.Executor.$anonfun$plugins$1(Executor.scala:337) > at org.apache.spark.util.Utils$.withContextClassLoader(Utils.scala:178) > at org.apache.spark.executor.Executor.<init>(Executor.scala:337) > at > org.apache.spark.executor.CoarseGrainedExecutorBackend$$anonfun$receive$1.applyOrElse(CoarseGrainedExecutorBackend.scala:181) > at org.apache.spark.rpc.netty.Inbox.$anonfun$process$1(Inbox.scala:115) > at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:213) > at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:100) > at > org.apache.spark.rpc.netty.MessageLoop.org$apache$spark$rpc$netty$MessageLoop$$receiveLoop(MessageLoop.scala:75) > at > org.apache.spark.rpc.netty.MessageLoop$$anon$1.run(MessageLoop.scala:41) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > 25/06/20 11:54:26 INFO YarnCoarseGrainedExecutorBackend: Driver commanded a > shutdown {code} > > We can specify `AsyncProfilerLoader.extractionDir` to spark local dir to > avoid this issue. > -- This message was sent by Atlassian Jira (v8.20.10#820010) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org