zhengruifeng commented on code in PR #49790:
URL: https://github.com/apache/spark/pull/49790#discussion_r1942159626
##########
python/pyspark/sql/pandas/conversion.py:
##########
@@ -785,8 +798,13 @@ def _create_from_arrow_table(
if not isinstance(schema, StructType):
schema = from_arrow_schema(table.schema,
prefer_timestamp_ntz=prefer_timestamp_ntz)
+ prefers_large_var_types = self._jconf.arrowUseLargeVarTypes()
table = _check_arrow_table_timestamps_localize(table, schema, True,
timezone).cast(
- to_arrow_schema(schema,
error_on_duplicated_field_names_in_struct=True)
+ to_arrow_schema(
+ schema,
+ error_on_duplicated_field_names_in_struct=True,
+ prefers_large_types=prefers_large_var_types,
Review Comment:
I feel we are using more and more configs, probably we can also combine them
into a `runner_conf: Dict[str, str]` later
##########
python/pyspark/sql/pandas/conversion.py:
##########
@@ -715,9 +721,16 @@ def _create_from_pandas_with_arrow(
pdf_slices = (pdf.iloc[start : start + step] for start in range(0,
len(pdf), step))
# Create list of Arrow (columns, arrow_type, spark_type) for
serializer dump_stream
+ prefers_large_var_types = self._jconf.arrowUseLargeVarTypes()
Review Comment:
nit: probably we can also fetch all configs via py4j in batch in the future.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]