xupefei commented on code in PR #45701:
URL: https://github.com/apache/spark/pull/45701#discussion_r1572148735


##########
connector/connect/client/jvm/src/main/scala/org/apache/spark/sql/SparkSession.scala:
##########
@@ -813,6 +823,28 @@ class SparkSession private[sql] (
    * Set to false to prevent client.releaseSession on close() (testing only)
    */
   private[sql] var releaseSessionOnClose = true
+
+  private[sql] def registerObservation(planId: Long, observation: 
Observation): Unit = {
+    // makes this class thread-safe:
+    // only the first thread entering this block can set sparkSession
+    // all other threads will see the exception, as it is only allowed to do 
this once
+    observation.synchronized {
+      if (observationRegistry.contains(planId)) {
+        throw new IllegalArgumentException("An Observation can be used with a 
Dataset only once")
+      }
+      observationRegistry.put(planId, observation)
+    }
+  }
+
+  private[sql] def setMetricsAndUnregisterObservation(
+      planId: Long,
+      metrics: Option[Map[String, Any]]): Unit = {
+    observationRegistry.get(planId).map { observation =>
+      if (observation.setMetricsAndNotify(metrics)) {
+        observationRegistry.remove(planId)

Review Comment:
   I had the same question when I looked at the code. In Spark Core we only 
de-register the Observation when some non-empty metrics are set. I am not sure 
under which circumstance the metrics can be empty.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to