Zhen Wang created SPARK-52536:
---------------------------------
Summary: Specify AsyncProfilerLoader.extractionDir to spark temp
dir
Key: SPARK-52536
URL: https://issues.apache.org/jira/browse/SPARK-52536
Project: Spark
Issue Type: Improvement
Components: Spark Core
Affects Versions: 4.0.0
Reporter: Zhen Wang
AsyncProfilerLoader uses `user.home` by default to store the extracted
libraries:
https://github.com/jvm-profiling-tools/ap-loader/blob/main/src/main/java/one/profiler/AsyncProfilerLoader.java#L139-L152
The `user.home` directory of the datanodes in our yarn cluster was not
initialized, causing the executor startup to fail:
```
25/06/20 11:54:26 ERROR YarnCoarseGrainedExecutorBackend: Executor self-exiting
due to : Unable to create executor due to /home/test
java.nio.file.AccessDeniedException: /home/pilot at
sun.nio.fs.UnixException.translateToIOException(UnixException.java:84) at
sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:102) at
sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:107) at
sun.nio.fs.UnixFileSystemProvider.createDirectory(UnixFileSystemProvider.java:384)
at java.nio.file.Files.createDirectory(Files.java:674) at
java.nio.file.Files.createAndCheckIsDirectory(Files.java:781) at
java.nio.file.Files.createDirectories(Files.java:767) at
one.profiler.AsyncProfilerLoader.getExtractionDirectory(AsyncProfilerLoader.java:133)
at
one.profiler.AsyncProfilerLoader.getAsyncProfilerPath(AsyncProfilerLoader.java:562)
at one.profiler.AsyncProfilerLoader.load(AsyncProfilerLoader.java:861) at
org.apache.spark.profiler.SparkAsyncProfiler.<init>(SparkAsyncProfiler.scala:70)
at
org.apache.spark.profiler.ProfilerExecutorPlugin.init(ProfilerPlugin.scala:82)
at
org.apache.spark.internal.plugin.ExecutorPluginContainer.$anonfun$executorPlugins$1(PluginContainer.scala:125)
at
scala.collection.TraversableLike.$anonfun$flatMap$1(TraversableLike.scala:293)
at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) at
scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) at
scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) at
scala.collection.TraversableLike.flatMap(TraversableLike.scala:293) at
scala.collection.TraversableLike.flatMap$(TraversableLike.scala:290) at
scala.collection.AbstractTraversable.flatMap(Traversable.scala:108) at
org.apache.spark.internal.plugin.ExecutorPluginContainer.<init>(PluginContainer.scala:113)
at
org.apache.spark.internal.plugin.PluginContainer$.apply(PluginContainer.scala:211)
at
org.apache.spark.internal.plugin.PluginContainer$.apply(PluginContainer.scala:199)
at org.apache.spark.executor.Executor.$anonfun$plugins$1(Executor.scala:337)
at org.apache.spark.util.Utils$.withContextClassLoader(Utils.scala:178) at
org.apache.spark.executor.Executor.<init>(Executor.scala:337) at
org.apache.spark.executor.CoarseGrainedExecutorBackend$$anonfun$receive$1.applyOrElse(CoarseGrainedExecutorBackend.scala:181)
at org.apache.spark.rpc.netty.Inbox.$anonfun$process$1(Inbox.scala:115) at
org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:213) at
org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:100) at
org.apache.spark.rpc.netty.MessageLoop.org$apache$spark$rpc$netty$MessageLoop$$receiveLoop(MessageLoop.scala:75)
at org.apache.spark.rpc.netty.MessageLoop$$anon$1.run(MessageLoop.scala:41) at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748) 25/06/20 11:54:26 INFO
YarnCoarseGrainedExecutorBackend: Driver commanded a shut
```
We can specify `AsyncProfilerLoader.extractionDir` to spark temp dir to avoid
this issue.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]