LuciferYang opened a new pull request, #39385:
URL: https://github.com/apache/spark/pull/39385
### What changes were proposed in this pull request?
This pr adds protection to `SparkPlanGraphWrapperSerializer#serialize` to
avoid throwing NPE.
### Why are the changes needed?
When `spark.ui.store.path` is configured to run `TPCDSQueryBenchmark`, found
the following NPE:
```
15:07:05.072 ERROR org.apache.spark.scheduler.AsyncEventQueue: Listener
SQLAppStatusListener threw an exception
java.lang.NullPointerException
at
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serializeSparkPlanGraphClusterWrapper(SparkPlanGraphWrapperSerializer.scala:111)
at
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serializeSparkPlanGraphNodeWrapper(SparkPlanGraphWrapperSerializer.scala:57)
at
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.$anonfun$serialize$1(SparkPlanGraphWrapperSerializer.scala:35)
at
scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
at
scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
at
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serialize(SparkPlanGraphWrapperSerializer.scala:34)
at
org.apache.spark.status.protobuf.KVStoreProtobufSerializer.serialize(KVStoreProtobufSerializer.scala:29)
at org.apache.spark.util.kvstore.RocksDB.write(RocksDB.java:188)
at
org.apache.spark.status.ElementTrackingStore.write(ElementTrackingStore.scala:123)
at
org.apache.spark.sql.execution.ui.SQLAppStatusListener.onExecutionStart(SQLAppStatusListener.scala:354)
at
org.apache.spark.sql.execution.ui.SQLAppStatusListener.onOtherEvent(SQLAppStatusListener.scala:433)
at
org.apache.spark.scheduler.SparkListenerBus.doPostEvent(SparkListenerBus.scala:100)
at
org.apache.spark.scheduler.SparkListenerBus.doPostEvent$(SparkListenerBus.scala:28)
at
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
at
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
at org.apache.spark.util.ListenerBus.postToAll(ListenerBus.scala:117)
at org.apache.spark.util.ListenerBus.postToAll$(ListenerBus.scala:101)
at
org.apache.spark.scheduler.AsyncEventQueue.super$postToAll(AsyncEventQueue.scala:105)
at
org.apache.spark.scheduler.AsyncEventQueue.$anonfun$dispatch$1(AsyncEventQueue.scala:105)
at
scala.runtime.java8.JFunction0$mcJ$sp.apply(JFunction0$mcJ$sp.java:23)
at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)
at
org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$dispatch(AsyncEventQueue.scala:100)
at
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.$anonfun$run$1(AsyncEventQueue.scala:96)
at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1444)
at
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.run(AsyncEventQueue.scala:96)
15:10:28.346 ERROR org.apache.spark.scheduler.AsyncEventQueue: Listener
SQLAppStatusListener threw an exception
java.lang.NullPointerException
at
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serializeSparkPlanGraphNode(SparkPlanGraphWrapperSerializer.scala:90)
at
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serializeSparkPlanGraphNodeWrapper(SparkPlanGraphWrapperSerializer.scala:56)
at
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.$anonfun$serialize$1(SparkPlanGraphWrapperSerializer.scala:35)
at
scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
at
scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
at
org.apache.spark.status.protobuf.sql.SparkPlanGraphWrapperSerializer.serialize(SparkPlanGraphWrapperSerializer.scala:34)
at
org.apache.spark.status.protobuf.KVStoreProtobufSerializer.serialize(KVStoreProtobufSerializer.scala:29)
at org.apache.spark.util.kvstore.RocksDB.write(RocksDB.java:188)
at
org.apache.spark.status.ElementTrackingStore.write(ElementTrackingStore.scala:123)
at
org.apache.spark.sql.execution.ui.SQLAppStatusListener.onAdaptiveExecutionUpdate(SQLAppStatusListener.scala:379)
at
org.apache.spark.sql.execution.ui.SQLAppStatusListener.onOtherEvent(SQLAppStatusListener.scala:434)
at
org.apache.spark.scheduler.SparkListenerBus.doPostEvent(SparkListenerBus.scala:100)
at
org.apache.spark.scheduler.SparkListenerBus.doPostEvent$(SparkListenerBus.scala:28)
at
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
at
org.apache.spark.scheduler.AsyncEventQueue.doPostEvent(AsyncEventQueue.scala:37)
at org.apache.spark.util.ListenerBus.postToAll(ListenerBus.scala:117)
at org.apache.spark.util.ListenerBus.postToAll$(ListenerBus.scala:101)
at
org.apache.spark.scheduler.AsyncEventQueue.super$postToAll(AsyncEventQueue.scala:105)
at
org.apache.spark.scheduler.AsyncEventQueue.$anonfun$dispatch$1(AsyncEventQueue.scala:105)
at
scala.runtime.java8.JFunction0$mcJ$sp.apply(JFunction0$mcJ$sp.java:23)
at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)
at
org.apache.spark.scheduler.AsyncEventQueue.org$apache$spark$scheduler$AsyncEventQueue$$dispatch(AsyncEventQueue.scala:100)
at
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.$anonfun$run$1(AsyncEventQueue.scala:96)
at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1444)
at
org.apache.spark.scheduler.AsyncEventQueue$$anon$2.run(AsyncEventQueue.scala:96)
```
### Does this PR introduce _any_ user-facing change?
No
### How was this patch tested?
- Pass GitHub Actions
- Manually execute `TPCDSQueryBenchmark` with `spark.ui.store.path`, observe
the log, no NPE
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]