yew1eb commented on issue #1961:
URL: https://github.com/apache/auron/issues/1961#issuecomment-3818481989
```
23:37:34.273 Executor task launch worker for task 0.0 in stage 0.0 (TID 0)
ERROR ExecutorClassLoader: Failed to check existence of class
org.apache.spark.sql.catalyst.expressions.Object on REPL class server at
spark://10.147.102.206:50531/artifacts/bc960bcf-da02-4ba1-9cc0-d8fa5716233e/classes/
java.io.IOException: Failed to connect to /10.147.102.206:50531
at
org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:304)
at
org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:224)
at
org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:236)
at
org.apache.spark.rpc.netty.NettyRpcEnv.downloadClient(NettyRpcEnv.scala:406)
at
org.apache.spark.rpc.netty.NettyRpcEnv.$anonfun$openChannel$4(NettyRpcEnv.scala:370)
at
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
at
org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1337)
at
org.apache.spark.rpc.netty.NettyRpcEnv.openChannel(NettyRpcEnv.scala:369)
at
org.apache.spark.executor.ExecutorClassLoader.getClassFileInputStreamFromSparkRPC(ExecutorClassLoader.scala:132)
at
org.apache.spark.executor.ExecutorClassLoader.$anonfun$fetchFn$1(ExecutorClassLoader.scala:63)
at
org.apache.spark.executor.ExecutorClassLoader.findClassLocally(ExecutorClassLoader.scala:173)
at
org.apache.spark.executor.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:110)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:593)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:580)
at
org.apache.spark.util.ParentClassLoader.loadClass(ParentClassLoader.java:40)
at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:526)
at
org.codehaus.janino.ClassLoaderIClassLoader.findIClass(ClassLoaderIClassLoader.java:75)
at org.codehaus.janino.IClassLoader.loadIClass(IClassLoader.java:317)
at
org.codehaus.janino.UnitCompiler.findTypeByName(UnitCompiler.java:8895)
at
org.codehaus.janino.UnitCompiler.getRawReferenceType(UnitCompiler.java:7065)
at
org.codehaus.janino.UnitCompiler.getReferenceType(UnitCompiler.java:6974)
at
org.codehaus.janino.UnitCompiler.getReferenceType(UnitCompiler.java:6855)
at org.codehaus.janino.UnitCompiler.getType2(UnitCompiler.java:6826)
at org.codehaus.janino.UnitCompiler.access$14800(UnitCompiler.java:236)
at
org.codehaus.janino.UnitCompiler$24.visitReferenceType(UnitCompiler.java:6724)
at
org.codehaus.janino.UnitCompiler$24.visitReferenceType(UnitCompiler.java:6721)
at org.codehaus.janino.Java$ReferenceType.accept(Java.java:4289)
at org.codehaus.janino.UnitCompiler.getType(UnitCompiler.java:6721)
at org.codehaus.janino.UnitCompiler.getRawType(UnitCompiler.java:6717)
at org.codehaus.janino.UnitCompiler.getType2(UnitCompiler.java:7189)
at org.codehaus.janino.UnitCompiler.access$14600(UnitCompiler.java:236)
at
org.codehaus.janino.UnitCompiler$24.visitArrayType(UnitCompiler.java:6722)
at
org.codehaus.janino.UnitCompiler$24.visitArrayType(UnitCompiler.java:6721)
at org.codehaus.janino.Java$ArrayType.accept(Java.java:4382)
at org.codehaus.janino.UnitCompiler.getType(UnitCompiler.java:6721)
at org.codehaus.janino.UnitCompiler.getRawType(UnitCompiler.java:6717)
at org.codehaus.janino.UnitCompiler.access$1300(UnitCompiler.java:236)
at
org.codehaus.janino.UnitCompiler$41.getParameterTypes2(UnitCompiler.java:10775)
at
org.codehaus.janino.IClass$IInvocable.getParameterTypes(IClass.java:903)
at org.codehaus.janino.IClass$IMethod.getDescriptor2(IClass.java:1170)
at org.codehaus.janino.IClass$IInvocable.getDescriptor(IClass.java:926)
at org.codehaus.janino.IClass.getIMethods(IClass.java:270)
at org.codehaus.janino.IClass.getIMethods(IClass.java:259)
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:502)
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:442)
at org.codehaus.janino.UnitCompiler.access$400(UnitCompiler.java:236)
at
org.codehaus.janino.UnitCompiler$3.visitPackageMemberClassDeclaration(UnitCompiler.java:422)
at
org.codehaus.janino.UnitCompiler$3.visitPackageMemberClassDeclaration(UnitCompiler.java:418)
at
org.codehaus.janino.Java$PackageMemberClassDeclaration.accept(Java.java:1688)
at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:418)
at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:392)
at org.codehaus.janino.UnitCompiler.access$000(UnitCompiler.java:236)
at
org.codehaus.janino.UnitCompiler$2.visitCompilationUnit(UnitCompiler.java:363)
at
org.codehaus.janino.UnitCompiler$2.visitCompilationUnit(UnitCompiler.java:361)
at org.codehaus.janino.Java$CompilationUnit.accept(Java.java:371)
at org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:361)
at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:264)
at
org.codehaus.janino.ClassBodyEvaluator.cook(ClassBodyEvaluator.java:294)
at
org.codehaus.janino.ClassBodyEvaluator.cook(ClassBodyEvaluator.java:288)
at
org.codehaus.janino.ClassBodyEvaluator.cook(ClassBodyEvaluator.java:267)
at org.codehaus.commons.compiler.Cookable.cook(Cookable.java:82)
at
org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$.doCompile(CodeGenerator.scala:1549)
at
org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$.$anonfun$cache$1(CodeGenerator.scala:1638)
at
org.apache.spark.util.NonFateSharingCache$$anon$1.load(NonFateSharingCache.scala:68)
at
org.sparkproject.guava.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3551)
at
org.sparkproject.guava.cache.LocalCache$Segment.loadSync(LocalCache.java:2302)
at
org.sparkproject.guava.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2177)
at
org.sparkproject.guava.cache.LocalCache$Segment.get(LocalCache.java:2068)
at org.sparkproject.guava.cache.LocalCache.get(LocalCache.java:3986)
at
org.sparkproject.guava.cache.LocalCache.getOrLoad(LocalCache.java:4007)
at
org.sparkproject.guava.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:4946)
at
org.apache.spark.util.NonFateSharingLoadingCache.$anonfun$get$2(NonFateSharingCache.scala:108)
at org.apache.spark.util.KeyLock.withLock(KeyLock.scala:64)
at
org.apache.spark.util.NonFateSharingLoadingCache.get(NonFateSharingCache.scala:108)
at
org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$.compile(CodeGenerator.scala:1490)
at
org.apache.spark.sql.catalyst.expressions.codegen.GenerateUnsafeProjection$.create(GenerateUnsafeProjection.scala:378)
at
org.apache.spark.sql.catalyst.expressions.codegen.GenerateUnsafeProjection$.generate(GenerateUnsafeProjection.scala:327)
at
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.createCodeGeneratedObject(Projection.scala:125)
at
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.createCodeGeneratedObject(Projection.scala:121)
at
org.apache.spark.sql.catalyst.expressions.CodeGeneratorWithInterpretedFallback.createObject(CodeGeneratorWithInterpretedFallback.scala:45)
at
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.create(Projection.scala:152)
at
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.create(Projection.scala:145)
at
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.create(Projection.scala:137)
at
org.apache.spark.sql.auron.NativeHelper$$anon$1.$anonfun$hasNext$1(NativeHelper.scala:137)
at
org.apache.auron.jni.AuronCallNativeWrapper.importBatch(AuronCallNativeWrapper.java:153)
at org.apache.auron.jni.JniBridge.nextBatch(Native Method)
at
org.apache.auron.jni.AuronCallNativeWrapper.loadNextBatch(AuronCallNativeWrapper.java:118)
at
org.apache.spark.sql.auron.NativeHelper$$anon$1.hasNext(NativeHelper.scala:133)
at
org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31)
at
org.apache.spark.sql.execution.datasources.FileFormatDataWriter.writeWithIterator(FileFormatDataWriter.scala:110)
at
org.apache.spark.sql.execution.datasources.FileFormatWriter$.$anonfun$executeTask$1(FileFormatWriter.scala:406)
at
org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1337)
at
org.apache.spark.sql.execution.datasources.FileFormatWriter$.executeTask(FileFormatWriter.scala:418)
at
org.apache.spark.sql.execution.datasources.WriteFilesExec.$anonfun$doExecuteWrite$1(WriteFiles.scala:107)
at
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2(RDD.scala:901)
at
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2$adapted(RDD.scala:901)
at
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:374)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:338)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:93)
at
org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:180)
at org.apache.spark.scheduler.Task.run(Task.scala:147)
at
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$5(Executor.scala:716)
at
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:86)
at
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:83)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:97)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:719)
at
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144)
at
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642)
at java.base/java.lang.Thread.run(Thread.java:1583)
Caused by: java.util.concurrent.RejectedExecutionException: event executor
terminated
at
io.netty.util.concurrent.SingleThreadEventExecutor.reject(SingleThreadEventExecutor.java:1123)
at
io.netty.util.concurrent.SingleThreadEventExecutor.offerTask(SingleThreadEventExecutor.java:413)
at
io.netty.util.concurrent.SingleThreadEventExecutor.addTask(SingleThreadEventExecutor.java:406)
at
io.netty.util.concurrent.SingleThreadEventExecutor.execute(SingleThreadEventExecutor.java:1025)
at
io.netty.util.concurrent.SingleThreadEventExecutor.execute0(SingleThreadEventExecutor.java:991)
at
io.netty.util.concurrent.SingleThreadEventExecutor.execute(SingleThreadEventExecutor.java:981)
at
io.netty.channel.AbstractChannel$AbstractUnsafe.register(AbstractChannel.java:349)
at
io.netty.channel.SingleThreadEventLoop.register(SingleThreadEventLoop.java:119)
at
io.netty.channel.SingleThreadEventLoop.register(SingleThreadEventLoop.java:113)
at
io.netty.channel.MultithreadEventLoopGroup.register(MultithreadEventLoopGroup.java:86)
at
io.netty.bootstrap.AbstractBootstrap.initAndRegister(AbstractBootstrap.java:339)
at io.netty.bootstrap.Bootstrap.doResolveAndConnect(Bootstrap.java:164)
at io.netty.bootstrap.Bootstrap.connect(Bootstrap.java:148)
at
org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:289)
... 109 more
```
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]