yew1eb commented on issue #1961:
URL: https://github.com/apache/auron/issues/1961#issuecomment-3818481989

   ```
   23:37:34.273 Executor task launch worker for task 0.0 in stage 0.0 (TID 0) 
ERROR ExecutorClassLoader: Failed to check existence of class 
org.apache.spark.sql.catalyst.expressions.Object on REPL class server at 
spark://10.147.102.206:50531/artifacts/bc960bcf-da02-4ba1-9cc0-d8fa5716233e/classes/
   java.io.IOException: Failed to connect to /10.147.102.206:50531
        at 
org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:304)
        at 
org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:224)
        at 
org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:236)
        at 
org.apache.spark.rpc.netty.NettyRpcEnv.downloadClient(NettyRpcEnv.scala:406)
        at 
org.apache.spark.rpc.netty.NettyRpcEnv.$anonfun$openChannel$4(NettyRpcEnv.scala:370)
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
        at 
org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1337)
        at 
org.apache.spark.rpc.netty.NettyRpcEnv.openChannel(NettyRpcEnv.scala:369)
        at 
org.apache.spark.executor.ExecutorClassLoader.getClassFileInputStreamFromSparkRPC(ExecutorClassLoader.scala:132)
        at 
org.apache.spark.executor.ExecutorClassLoader.$anonfun$fetchFn$1(ExecutorClassLoader.scala:63)
        at 
org.apache.spark.executor.ExecutorClassLoader.findClassLocally(ExecutorClassLoader.scala:173)
        at 
org.apache.spark.executor.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:110)
        at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:593)
        at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:580)
        at 
org.apache.spark.util.ParentClassLoader.loadClass(ParentClassLoader.java:40)
        at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:526)
        at 
org.codehaus.janino.ClassLoaderIClassLoader.findIClass(ClassLoaderIClassLoader.java:75)
        at org.codehaus.janino.IClassLoader.loadIClass(IClassLoader.java:317)
        at 
org.codehaus.janino.UnitCompiler.findTypeByName(UnitCompiler.java:8895)
        at 
org.codehaus.janino.UnitCompiler.getRawReferenceType(UnitCompiler.java:7065)
        at 
org.codehaus.janino.UnitCompiler.getReferenceType(UnitCompiler.java:6974)
        at 
org.codehaus.janino.UnitCompiler.getReferenceType(UnitCompiler.java:6855)
        at org.codehaus.janino.UnitCompiler.getType2(UnitCompiler.java:6826)
        at org.codehaus.janino.UnitCompiler.access$14800(UnitCompiler.java:236)
        at 
org.codehaus.janino.UnitCompiler$24.visitReferenceType(UnitCompiler.java:6724)
        at 
org.codehaus.janino.UnitCompiler$24.visitReferenceType(UnitCompiler.java:6721)
        at org.codehaus.janino.Java$ReferenceType.accept(Java.java:4289)
        at org.codehaus.janino.UnitCompiler.getType(UnitCompiler.java:6721)
        at org.codehaus.janino.UnitCompiler.getRawType(UnitCompiler.java:6717)
        at org.codehaus.janino.UnitCompiler.getType2(UnitCompiler.java:7189)
        at org.codehaus.janino.UnitCompiler.access$14600(UnitCompiler.java:236)
        at 
org.codehaus.janino.UnitCompiler$24.visitArrayType(UnitCompiler.java:6722)
        at 
org.codehaus.janino.UnitCompiler$24.visitArrayType(UnitCompiler.java:6721)
        at org.codehaus.janino.Java$ArrayType.accept(Java.java:4382)
        at org.codehaus.janino.UnitCompiler.getType(UnitCompiler.java:6721)
        at org.codehaus.janino.UnitCompiler.getRawType(UnitCompiler.java:6717)
        at org.codehaus.janino.UnitCompiler.access$1300(UnitCompiler.java:236)
        at 
org.codehaus.janino.UnitCompiler$41.getParameterTypes2(UnitCompiler.java:10775)
        at 
org.codehaus.janino.IClass$IInvocable.getParameterTypes(IClass.java:903)
        at org.codehaus.janino.IClass$IMethod.getDescriptor2(IClass.java:1170)
        at org.codehaus.janino.IClass$IInvocable.getDescriptor(IClass.java:926)
        at org.codehaus.janino.IClass.getIMethods(IClass.java:270)
        at org.codehaus.janino.IClass.getIMethods(IClass.java:259)
        at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:502)
        at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:442)
        at org.codehaus.janino.UnitCompiler.access$400(UnitCompiler.java:236)
        at 
org.codehaus.janino.UnitCompiler$3.visitPackageMemberClassDeclaration(UnitCompiler.java:422)
        at 
org.codehaus.janino.UnitCompiler$3.visitPackageMemberClassDeclaration(UnitCompiler.java:418)
        at 
org.codehaus.janino.Java$PackageMemberClassDeclaration.accept(Java.java:1688)
        at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:418)
        at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:392)
        at org.codehaus.janino.UnitCompiler.access$000(UnitCompiler.java:236)
        at 
org.codehaus.janino.UnitCompiler$2.visitCompilationUnit(UnitCompiler.java:363)
        at 
org.codehaus.janino.UnitCompiler$2.visitCompilationUnit(UnitCompiler.java:361)
        at org.codehaus.janino.Java$CompilationUnit.accept(Java.java:371)
        at org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:361)
        at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:264)
        at 
org.codehaus.janino.ClassBodyEvaluator.cook(ClassBodyEvaluator.java:294)
        at 
org.codehaus.janino.ClassBodyEvaluator.cook(ClassBodyEvaluator.java:288)
        at 
org.codehaus.janino.ClassBodyEvaluator.cook(ClassBodyEvaluator.java:267)
        at org.codehaus.commons.compiler.Cookable.cook(Cookable.java:82)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$.doCompile(CodeGenerator.scala:1549)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$.$anonfun$cache$1(CodeGenerator.scala:1638)
        at 
org.apache.spark.util.NonFateSharingCache$$anon$1.load(NonFateSharingCache.scala:68)
        at 
org.sparkproject.guava.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3551)
        at 
org.sparkproject.guava.cache.LocalCache$Segment.loadSync(LocalCache.java:2302)
        at 
org.sparkproject.guava.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2177)
        at 
org.sparkproject.guava.cache.LocalCache$Segment.get(LocalCache.java:2068)
        at org.sparkproject.guava.cache.LocalCache.get(LocalCache.java:3986)
        at 
org.sparkproject.guava.cache.LocalCache.getOrLoad(LocalCache.java:4007)
        at 
org.sparkproject.guava.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:4946)
        at 
org.apache.spark.util.NonFateSharingLoadingCache.$anonfun$get$2(NonFateSharingCache.scala:108)
        at org.apache.spark.util.KeyLock.withLock(KeyLock.scala:64)
        at 
org.apache.spark.util.NonFateSharingLoadingCache.get(NonFateSharingCache.scala:108)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$.compile(CodeGenerator.scala:1490)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.GenerateUnsafeProjection$.create(GenerateUnsafeProjection.scala:378)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.GenerateUnsafeProjection$.generate(GenerateUnsafeProjection.scala:327)
        at 
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.createCodeGeneratedObject(Projection.scala:125)
        at 
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.createCodeGeneratedObject(Projection.scala:121)
        at 
org.apache.spark.sql.catalyst.expressions.CodeGeneratorWithInterpretedFallback.createObject(CodeGeneratorWithInterpretedFallback.scala:45)
        at 
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.create(Projection.scala:152)
        at 
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.create(Projection.scala:145)
        at 
org.apache.spark.sql.catalyst.expressions.UnsafeProjection$.create(Projection.scala:137)
        at 
org.apache.spark.sql.auron.NativeHelper$$anon$1.$anonfun$hasNext$1(NativeHelper.scala:137)
        at 
org.apache.auron.jni.AuronCallNativeWrapper.importBatch(AuronCallNativeWrapper.java:153)
        at org.apache.auron.jni.JniBridge.nextBatch(Native Method)
        at 
org.apache.auron.jni.AuronCallNativeWrapper.loadNextBatch(AuronCallNativeWrapper.java:118)
        at 
org.apache.spark.sql.auron.NativeHelper$$anon$1.hasNext(NativeHelper.scala:133)
        at 
org.apache.spark.util.CompletionIterator.hasNext(CompletionIterator.scala:31)
        at 
org.apache.spark.sql.execution.datasources.FileFormatDataWriter.writeWithIterator(FileFormatDataWriter.scala:110)
        at 
org.apache.spark.sql.execution.datasources.FileFormatWriter$.$anonfun$executeTask$1(FileFormatWriter.scala:406)
        at 
org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1337)
        at 
org.apache.spark.sql.execution.datasources.FileFormatWriter$.executeTask(FileFormatWriter.scala:418)
        at 
org.apache.spark.sql.execution.datasources.WriteFilesExec.$anonfun$doExecuteWrite$1(WriteFiles.scala:107)
        at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2(RDD.scala:901)
        at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsInternal$2$adapted(RDD.scala:901)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:374)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:338)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:93)
        at 
org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:180)
        at org.apache.spark.scheduler.Task.run(Task.scala:147)
        at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$5(Executor.scala:716)
        at 
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:86)
        at 
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:83)
        at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:97)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:719)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642)
        at java.base/java.lang.Thread.run(Thread.java:1583)
   Caused by: java.util.concurrent.RejectedExecutionException: event executor 
terminated
        at 
io.netty.util.concurrent.SingleThreadEventExecutor.reject(SingleThreadEventExecutor.java:1123)
        at 
io.netty.util.concurrent.SingleThreadEventExecutor.offerTask(SingleThreadEventExecutor.java:413)
        at 
io.netty.util.concurrent.SingleThreadEventExecutor.addTask(SingleThreadEventExecutor.java:406)
        at 
io.netty.util.concurrent.SingleThreadEventExecutor.execute(SingleThreadEventExecutor.java:1025)
        at 
io.netty.util.concurrent.SingleThreadEventExecutor.execute0(SingleThreadEventExecutor.java:991)
        at 
io.netty.util.concurrent.SingleThreadEventExecutor.execute(SingleThreadEventExecutor.java:981)
        at 
io.netty.channel.AbstractChannel$AbstractUnsafe.register(AbstractChannel.java:349)
        at 
io.netty.channel.SingleThreadEventLoop.register(SingleThreadEventLoop.java:119)
        at 
io.netty.channel.SingleThreadEventLoop.register(SingleThreadEventLoop.java:113)
        at 
io.netty.channel.MultithreadEventLoopGroup.register(MultithreadEventLoopGroup.java:86)
        at 
io.netty.bootstrap.AbstractBootstrap.initAndRegister(AbstractBootstrap.java:339)
        at io.netty.bootstrap.Bootstrap.doResolveAndConnect(Bootstrap.java:164)
        at io.netty.bootstrap.Bootstrap.connect(Bootstrap.java:148)
        at 
org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:289)
        ... 109 more
   
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to