This is an automated email from the ASF dual-hosted git repository.
gurwls223 pushed a commit to branch branch-2.4
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/branch-2.4 by this push:
new 476010a [SPARK-31903][SQL][PYSPARK][2.4] Fix toPandas with Arrow
enabled to show metrics in Query UI
476010a is described below
commit 476010aedd101e1a807c202d71328415109660ae
Author: Takuya UESHIN <[email protected]>
AuthorDate: Sat Jun 6 16:50:40 2020 +0900
[SPARK-31903][SQL][PYSPARK][2.4] Fix toPandas with Arrow enabled to show
metrics in Query UI
### What changes were proposed in this pull request?
This is a backport of #28730.
In `Dataset.collectAsArrowToPython`, since the code block for
`serveToStream` is run in the separate thread, `withAction` finishes as soon as
it starts the thread. As a result, it doesn't collect the metrics of the actual
action and Query UI shows the plan graph without metrics.
We should call `serveToStream` first, then `withAction` in it.
### Why are the changes needed?
When calling toPandas, usually Query UI shows each plan node's metric:
```py
>>> df = spark.createDataFrame([(1, 10, 'abc'), (2, 20, 'def')],
schema=['x', 'y', 'z'])
>>> df.toPandas()
x y z
0 1 10 abc
1 2 20 def
```

but if Arrow execution is enabled, it shows only plan nodes and the
duration is not correct:
```py
>>> spark.conf.set('spark.sql.execution.arrow.enabled', True)
>>> df.toPandas()
x y z
0 1 10 abc
1 2 20 def
```

### Does this PR introduce _any_ user-facing change?
Yes, the Query UI will show the plan with the correct metrics.
### How was this patch tested?
I checked it manually in my local.

Closes #28740 from
ueshin/issues/SPARK-31903/2.4/to_pandas_with_arrow_query_ui.
Authored-by: Takuya UESHIN <[email protected]>
Signed-off-by: HyukjinKwon <[email protected]>
---
sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala
b/sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala
index 6bc0d0b..a755a6f 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/Dataset.scala
@@ -3283,8 +3283,8 @@ class Dataset[T] private[sql](
private[sql] def collectAsArrowToPython(): Array[Any] = {
val timeZoneId = sparkSession.sessionState.conf.sessionLocalTimeZone
- withAction("collectAsArrowToPython", queryExecution) { plan =>
- PythonRDD.serveToStreamWithSync("serve-Arrow") { out =>
+ PythonRDD.serveToStreamWithSync("serve-Arrow") { out =>
+ withAction("collectAsArrowToPython", queryExecution) { plan =>
val batchWriter = new ArrowBatchStreamWriter(schema, out, timeZoneId)
val arrowBatchRdd = toArrowBatchRdd(plan)
val numPartitions = arrowBatchRdd.partitions.length
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]