zhouyuan opened a new issue, #10195: URL: https://github.com/apache/incubator-gluten/issues/10195
### Backend VL (Velox) ### Bug description ``` 2025-07-15T18:23:19.4895514Z - string sentences function *** FAILED *** 2025-07-15T18:23:19.4898068Z org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 230.0 failed 1 times, most recent failure: Lost task 0.0 in stage 230.0 (TID 292) (905469d1b180 executor driver): java.lang.NullPointerException: Cannot invoke "org.apache.spark.sql.catalyst.util.ArrayData.numElements()" because "<local2>" is null 2025-07-15T18:23:19.4900984Z at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificSafeProjection.MapObjects_3$(Unknown Source) 2025-07-15T18:23:19.4902251Z at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificSafeProjection.apply(Unknown Source) 2025-07-15T18:23:19.4902954Z at scala.collection.Iterator$$anon$10.next(Iterator.scala:461) 2025-07-15T18:23:19.4903400Z at scala.collection.Iterator$$anon$10.next(Iterator.scala:461) 2025-07-15T18:23:19.4903910Z at org.apache.spark.util.Iterators$.size(Iterators.scala:31) 2025-07-15T18:23:19.4904519Z at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1787) 2025-07-15T18:23:19.4905670Z at org.apache.spark.rdd.RDD.$anonfun$count$1(RDD.scala:1296) 2025-07-15T18:23:19.4906323Z at org.apache.spark.rdd.RDD.$anonfun$count$1$adapted(RDD.scala:1296) 2025-07-15T18:23:19.4907105Z at org.apache.spark.SparkContext.$anonfun$runJob$5(SparkContext.scala:2433) 2025-07-15T18:23:19.4907859Z at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:93) 2025-07-15T18:23:19.4908666Z at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166) 2025-07-15T18:23:19.4909664Z at org.apache.spark.scheduler.Task.run(Task.scala:141) 2025-07-15T18:23:19.4910423Z at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620) 2025-07-15T18:23:19.4911372Z at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64) 2025-07-15T18:23:19.4912401Z at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61) 2025-07-15T18:23:19.4913257Z at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94) 2025-07-15T18:23:19.4913952Z at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623) 2025-07-15T18:23:19.4915094Z at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) 2025-07-15T18:23:19.4916103Z at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) 2025-07-15T18:23:19.4916863Z at java.base/java.lang.Thread.run(Thread.java:833) 2025-07-15T18:23:19.4917193Z 2025-07-15T18:23:19.4917306Z Driver stacktrace: 2025-07-15T18:23:19.4918051Z at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2856) 2025-07-15T18:23:19.4919172Z at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2792) 2025-07-15T18:23:19.4920417Z at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2791) 2025-07-15T18:23:19.4921436Z at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) 2025-07-15T18:23:19.4922357Z at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) 2025-07-15T18:23:19.4923384Z at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) 2025-07-15T18:23:19.4924280Z at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2791) 2025-07-15T18:23:19.4925321Z at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1247) 2025-07-15T18:23:19.4926523Z at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1247) 2025-07-15T18:23:19.4927354Z at scala.Option.foreach(Option.scala:407) 2025-07-15T18:23:19.4927780Z ... 2025-07-15T18:23:19.4928128Z Cause: java.lang.NullPointerException: 2025-07-15T18:23:19.4929083Z at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificSafeProjection.MapObjects_3$(Unknown Source) 2025-07-15T18:23:19.4930839Z at org.apache.spark.sql.catalyst.expressions.GeneratedClass$SpecificSafeProjection.apply(Unknown Source) 2025-07-15T18:23:19.4932093Z at scala.collection.Iterator$$anon$10.next(Iterator.scala:461) 2025-07-15T18:23:19.4933088Z at scala.collection.Iterator$$anon$10.next(Iterator.scala:461) 2025-07-15T18:23:19.4933915Z at org.apache.spark.util.Iterators$.size(Iterators.scala:31) 2025-07-15T18:23:19.4934737Z at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1787) 2025-07-15T18:23:19.4935535Z at org.apache.spark.rdd.RDD.$anonfun$count$1(RDD.scala:1296) 2025-07-15T18:23:19.4936347Z at org.apache.spark.rdd.RDD.$anonfun$count$1$adapted(RDD.scala:1296) 2025-07-15T18:23:19.4937492Z at org.apache.spark.SparkContext.$anonfun$runJob$5(SparkContext.scala:2433) 2025-07-15T18:23:19.4938252Z at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:93) ``` ### Gluten version _No response_ ### Spark version None ### Spark configurations _No response_ ### System information _No response_ ### Relevant logs ```bash ``` -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
