AlenkaF commented on code in PR #13591:
URL: https://github.com/apache/arrow/pull/13591#discussion_r920725733
##########
python/pyarrow/parquet/__init__.py:
##########
@@ -3113,9 +3114,14 @@ def file_visitor(written_file):
dataset. The first time each partition directory is encountered
the entire directory will be deleted. This allows you to overwrite
old partitions completely.
+ This option is only supported for use_legacy_dataset=False.
**kwargs : dict,
- Additional kwargs for write_table function. See docstring for
- `write_table` or `ParquetWriter` for more information.
+ When use_legacy_dataset=False, used as additional kwargs for
+ `pyarrow.dataset.write_dataset` function (See docstring for
+ write_dataset or ParquetFileFormat for more information).
Review Comment:
```suggestion
`write_dataset` or `ParquetFileFormat` for more information).
```
##########
python/pyarrow/parquet/__init__.py:
##########
@@ -3063,16 +3064,19 @@ def write_to_dataset(table, root_path,
partition_cols=None,
used determined by the number of available CPU cores.
schema : Schema, optional
partitioning : Partitioning or list[str], optional
+ (This option is used only when `use_legacy_dataset` is False.)
Review Comment:
Hm, I do not think so but can be wrong:
https://github.com/apache/arrow/blob/69b06ecd7da4b4d58a0bc9858068b99bfdfa4876/python/pyarrow/parquet/__init__.py#L3243-L3261
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]