HyukjinKwon commented on code in PR #41528:
URL: https://github.com/apache/spark/pull/41528#discussion_r1224171092
##########
python/pyspark/pandas/internal.py:
##########
@@ -935,10 +935,12 @@ def attach_sequence_column(sdf: PySparkDataFrame,
column_name: str) -> PySparkDa
)
return sdf.select(sequential_index.alias(column_name), *scols)
- # TODO(SPARK-43610): Enable `InternalFrame.attach_distributed_column` in
Spark Connect.
@staticmethod
def attach_distributed_column(sdf: PySparkDataFrame, column_name: str) ->
PySparkDataFrame:
scols = [scol_for(sdf, column) for column in sdf.columns]
+ # We don't add an alias to avoid adding a dedicated protobuf message
for cosmetic changes.
Review Comment:
Can you say nicely? e.g.,
> Does not add an alias to avoid having some changes in protobuf definition
for now.
> The alias is more for query string in DataFrame.explain, and they are
cosmetic changes.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]