yew1eb opened a new issue, #1961:
URL: https://github.com/apache/auron/issues/1961

   
https://github.com/apache/auron/actions/runs/21383141822/job/61554030266?pr=1958
   
   
   ```
   - log function with negative input *** FAILED ***
     org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 
in stage 119.0 failed 1 times, most recent failure: Lost task 0.0 in stage 
119.0 (TID 288) 
(runnervmymu0l.lcg2avo1p2berav5sjxbnkio3f.bx.internal.cloudapp.net executor 
driver): org.apache.spark.executor.RemoteClassLoaderError: 
org.apache.spark.sql.catalyst.expressions.Object
        at 
org.apache.spark.executor.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:117)
        at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:593)
        at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:580)
        at 
org.apache.spark.util.ParentClassLoader.loadClass(ParentClassLoader.java:40)
        at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:526)
        at 
org.codehaus.janino.ClassLoaderIClassLoader.findIClass(ClassLoaderIClassLoader.java:75)
        at org.codehaus.janino.IClassLoader.loadIClass(IClassLoader.java:317)
        at 
org.codehaus.janino.UnitCompiler.findTypeByName(UnitCompiler.java:8895)
        at 
org.codehaus.janino.UnitCompiler.getRawReferenceType(UnitCompiler.java:7065)
        at 
org.codehaus.janino.UnitCompiler.getReferenceType(UnitCompiler.java:6974)
        at 
org.codehaus.janino.UnitCompiler.getReferenceType(UnitCompiler.java:6855)
        at org.codehaus.janino.UnitCompiler.getType2(UnitCompiler.java:6826)
        at org.codehaus.janino.UnitCompiler.access$14800(UnitCompiler.java:236)
        at 
org.codehaus.janino.UnitCompiler$24.visitReferenceType(UnitCompiler.java:6724)
        at 
org.codehaus.janino.UnitCompiler$24.visitReferenceType(UnitCompiler.java:6721)
        at org.codehaus.janino.Java$ReferenceType.accept(Java.java:4289)
        at org.codehaus.janino.UnitCompiler.getType(UnitCompiler.java:6721)
        at org.codehaus.janino.UnitCompiler.getRawType(UnitCompiler.java:6717)
        at org.codehaus.janino.UnitCompiler.getType2(UnitCompiler.java:7189)
        at org.codehaus.janino.UnitCompiler.access$14600(UnitCompiler.java:236)
        at 
org.codehaus.janino.UnitCompiler$24.visitArrayType(UnitCompiler.java:6722)
        at 
org.codehaus.janino.UnitCompiler$24.visitArrayType(UnitCompiler.java:6721)
        at org.codehaus.janino.Java$ArrayType.accept(Java.java:4382)
        at org.codehaus.janino.UnitCompiler.getType(UnitCompiler.java:6721)
        at org.codehaus.janino.UnitCompiler.getRawType(UnitCompiler.java:6717)
        at org.codehaus.janino.UnitCompiler.access$1300(UnitCompiler.java:236)
        at 
org.codehaus.janino.UnitCompiler$41.getParameterTypes2(UnitCompiler.java:10775)
        at 
org.codehaus.janino.IClass$IInvocable.getParameterTypes(IClass.java:903)
        at org.codehaus.janino.IClass$IMethod.getDescriptor2(IClass.java:1170)
        at org.codehaus.janino.IClass$IInvocable.getDescriptor(IClass.java:926)
        at org.codehaus.janino.IClass.getIMethods(IClass.java:270)
        at org.codehaus.janino.IClass.getIMethods(IClass.java:259)
        at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:502)
        at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:442)
        at org.codehaus.janino.UnitCompiler.access$400(UnitCompiler.java:236)
        at 
org.codehaus.janino.UnitCompiler$3.visitPackageMemberClassDeclaration(UnitCompiler.java:422)
        at 
org.codehaus.janino.UnitCompiler$3.visitPackageMemberClassDeclaration(UnitCompiler.java:418)
        at 
org.codehaus.janino.Java$PackageMemberClassDeclaration.accept(Java.java:1688)
        at org.codehaus.janino.UnitCompiler.compile(UnitCompiler.java:418)
        at org.codehaus.janino.UnitCompiler.compile2(UnitCompiler.java:392)
        at org.codehaus.janino.UnitCompiler.access$000(UnitCompiler.java:236)
        at 
org.codehaus.janino.UnitCompiler$2.visitCompilationUnit(UnitCompiler.java:363)
        at 
org.codehaus.janino.UnitCompiler$2.visitCompilationUnit(UnitCompiler.java:361)
        at org.codehaus.janino.Java$CompilationUnit.accept(Java.java:371)
        at org.codehaus.janino.UnitCompiler.compileUnit(UnitCompiler.java:361)
        at org.codehaus.janino.SimpleCompiler.cook(SimpleCompiler.java:264)
        at 
org.codehaus.janino.ClassBodyEvaluator.cook(ClassBodyEvaluator.java:294)
        at 
org.codehaus.janino.ClassBodyEvaluator.cook(ClassBodyEvaluator.java:288)
        at 
org.codehaus.janino.ClassBodyEvaluator.cook(ClassBodyEvaluator.java:267)
        at org.codehaus.commons.compiler.Cookable.cook(Cookable.java:82)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$.doCompile(CodeGenerator.scala:1549)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$.$anonfun$cache$1(CodeGenerator.scala:1638)
        at 
org.apache.spark.util.NonFateSharingCache$$anon$1.load(NonFateSharingCache.scala:68)
        at 
org.sparkproject.guava.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3551)
        at 
org.sparkproject.guava.cache.LocalCache$Segment.loadSync(LocalCache.java:2302)
        at 
org.sparkproject.guava.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2177)
        at 
org.sparkproject.guava.cache.LocalCache$Segment.get(LocalCache.java:2068)
        at org.sparkproject.guava.cache.LocalCache.get(LocalCache.java:3986)
        at 
org.sparkproject.guava.cache.LocalCache.getOrLoad(LocalCache.java:4007)
        at 
org.sparkproject.guava.cache.LocalCache$LocalLoadingCache.get(LocalCache.java:4946)
        at 
org.apache.spark.util.NonFateSharingLoadingCache.$anonfun$get$2(NonFateSharingCache.scala:108)
        at org.apache.spark.util.KeyLock.withLock(KeyLock.scala:64)
        at 
org.apache.spark.util.NonFateSharingLoadingCache.get(NonFateSharingCache.scala:108)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$.compile(CodeGenerator.scala:1490)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.GenerateSafeProjection$.create(GenerateSafeProjection.scala:205)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.GenerateSafeProjection$.create(GenerateSafeProjection.scala:39)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator.generate(CodeGenerator.scala:1415)
        at 
org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator.generate(CodeGenerator.scala:1412)
        at 
org.apache.spark.sql.execution.DeserializeToObjectExec.$anonfun$doExecute$1(objects.scala:95)
        at 
org.apache.spark.sql.execution.DeserializeToObjectExec.$anonfun$doExecute$1$adapted(objects.scala:94)
        at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndexInternal$2(RDD.scala:888)
        at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndexInternal$2$adapted(RDD.scala:888)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:374)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:338)
        at 
org.apache.spark.sql.execution.SQLExecutionRDD.$anonfun$compute$1(SQLExecutionRDD.scala:52)
        at 
org.apache.spark.sql.internal.SQLConf$.withExistingConf(SQLConf.scala:162)
        at 
org.apache.spark.sql.execution.SQLExecutionRDD.compute(SQLExecutionRDD.scala:52)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:374)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:338)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:374)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:338)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:93)
        at 
org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:180)
        at org.apache.spark.scheduler.Task.run(Task.scala:147)
        at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$5(Executor.scala:716)
        at 
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:86)
        at 
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:83)
        at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:97)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:719)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642)
        at java.base/java.lang.Thread.run(Thread.java:1583)
   Caused by: java.io.IOException: Failed to connect to 
runnervmymu0l.lcg2avo1p2berav5sjxbnkio3f.bx.internal.cloudapp.net/10.1.0.153:46807
        at 
org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:304)
        at 
org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:224)
        at 
org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:236)
        at 
org.apache.spark.rpc.netty.NettyRpcEnv.downloadClient(NettyRpcEnv.scala:406)
        at 
org.apache.spark.rpc.netty.NettyRpcEnv.$anonfun$openChannel$4(NettyRpcEnv.scala:370)
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
        at 
org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1337)
        at 
org.apache.spark.rpc.netty.NettyRpcEnv.openChannel(NettyRpcEnv.scala:369)
        at 
org.apache.spark.executor.ExecutorClassLoader.getClassFileInputStreamFromSparkRPC(ExecutorClassLoader.scala:132)
        at 
org.apache.spark.executor.ExecutorClassLoader.$anonfun$fetchFn$1(ExecutorClassLoader.scala:63)
        at 
org.apache.spark.executor.ExecutorClassLoader.findClassLocally(ExecutorClassLoader.scala:173)
        at 
org.apache.spark.executor.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:110)
        ... 93 more
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to