LuciferYang opened a new pull request, #39385:
URL: https://github.com/apache/spark/pull/39385

   ### What changes were proposed in this pull request?
   This pr adds protection to `SparkPlanGraphWrapperSerializer#serialize` to 
avoid throwing NPE.
   
   
   ### Why are the changes needed?
   When `spark.ui.store.path` is configured to run `TPCDSQueryBenchmark`, found 
the following NPE:
   
   ```
   15:07:05.072 ERROR org.apache.spark.scheduler.AsyncEventQueue: Listener 
SQLAppStatusListener threw an exception
   java.lang.NullPointerException
        at 
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serializeSparkPlanGraphClusterWrapper(SparkPlanGraphWrapperSerializer.scala:111)
        at 
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serializeSparkPlanGraphNodeWrapper(SparkPlanGraphWrapperSerializer.scala:57)
        at 
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.$anonfun$serialize$1(SparkPlanGraphWrapperSerializer.scala:35)
        at 
scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
        at 
scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
        at 
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serialize(SparkPlanGraphWrapperSerializer.scala:34)
        at 
org.apache.spark.status.protobuf.KVStoreProtobufSerializer.serialize(KVStoreProtobufSerializer.scala:29)
        at org.apache.spark.util.kvstore.RocksDB.write(RocksDB.java:188)
        at 
org.apache.spark.status.ElementTrackingStore.write(ElementTrackingStore.scala:123)
        at 
org.apache.spark.sql.execution.ui.SQLAppStatusListener.onExecutionStart(SQLAppStatusListener.scala:354)
        at 
org.apache.spark.sql.execution.ui.SQLAppStatusListener.onOtherEvent(SQLAppStatusListener.scala:433)
        at 
org.apache.spark.scheduler.SparkListenerBus.doPostEvent(SparkListenerBus.scala:100)
        at 
org.apache.spark.scheduler.SparkListenerBus.doPostEvent$(SparkListenerBus.scala:28)
        at 
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
        at 
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
        at org.apache.spark.util.ListenerBus.postToAll(ListenerBus.scala:117)
        at org.apache.spark.util.ListenerBus.postToAll$(ListenerBus.scala:101)
        at 
org.apache.spark.scheduler.AsyncEventQueue.super$postToAll(AsyncEventQueue.scala:105)
        at 
org.apache.spark.scheduler.AsyncEventQueue.$anonfun$dispatch$1(AsyncEventQueue.scala:105)
        at 
scala.runtime.java8.JFunction0$mcJ$sp.apply(JFunction0$mcJ$sp.java:23)
        at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)
        at 
org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$dispatch(AsyncEventQueue.scala:100)
        at 
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.$anonfun$run$1(AsyncEventQueue.scala:96)
        at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1444)
        at 
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.run(AsyncEventQueue.scala:96)
   
   15:10:28.346 ERROR org.apache.spark.scheduler.AsyncEventQueue: Listener 
SQLAppStatusListener threw an exception
   java.lang.NullPointerException
        at 
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serializeSparkPlanGraphNode(SparkPlanGraphWrapperSerializer.scala:90)
        at 
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serializeSparkPlanGraphNodeWrapper(SparkPlanGraphWrapperSerializer.scala:56)
        at 
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.$anonfun$serialize$1(SparkPlanGraphWrapperSerializer.scala:35)
        at 
scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
        at 
scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
        at 
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serialize(SparkPlanGraphWrapperSerializer.scala:34)
        at 
org.apache.spark.status.protobuf.KVStoreProtobufSerializer.serialize(KVStoreProtobufSerializer.scala:29)
        at org.apache.spark.util.kvstore.RocksDB.write(RocksDB.java:188)
        at 
org.apache.spark.status.ElementTrackingStore.write(ElementTrackingStore.scala:123)
        at 
org.apache.spark.sql.execution.ui.SQLAppStatusListener.onAdaptiveExecutionUpdate(SQLAppStatusListener.scala:379)
        at 
org.apache.spark.sql.execution.ui.SQLAppStatusListener.onOtherEvent(SQLAppStatusListener.scala:434)
        at 
org.apache.spark.scheduler.SparkListenerBus.doPostEvent(SparkListenerBus.scala:100)
        at 
org.apache.spark.scheduler.SparkListenerBus.doPostEvent$(SparkListenerBus.scala:28)
        at 
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
        at 
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
        at org.apache.spark.util.ListenerBus.postToAll(ListenerBus.scala:117)
        at org.apache.spark.util.ListenerBus.postToAll$(ListenerBus.scala:101)
        at 
org.apache.spark.scheduler.AsyncEventQueue.super$postToAll(AsyncEventQueue.scala:105)
        at 
org.apache.spark.scheduler.AsyncEventQueue.$anonfun$dispatch$1(AsyncEventQueue.scala:105)
        at 
scala.runtime.java8.JFunction0$mcJ$sp.apply(JFunction0$mcJ$sp.java:23)
        at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)
        at 
org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$dispatch(AsyncEventQueue.scala:100)
        at 
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.$anonfun$run$1(AsyncEventQueue.scala:96)
        at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1444)
        at 
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.run(AsyncEventQueue.scala:96)
   ```
   
   ### Does this PR introduce _any_ user-facing change?
   No
   
   
   ### How was this patch tested?
   
   - Pass GitHub Actions
   - Manually execute `TPCDSQueryBenchmark` with `spark.ui.store.path`, observe 
the log, no NPE 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to