AlenkaF commented on code in PR #12811:
URL: https://github.com/apache/arrow/pull/12811#discussion_r845909851


##########
python/pyarrow/tests/parquet/test_dataset.py:
##########
@@ -1290,7 +1290,7 @@ def _test_write_to_dataset_no_partitions(base_path,
     # Without partitions, append files to root_path
     n = 5
     for i in range(n):
-        pq.write_to_dataset(output_table, base_path,
+        pq.write_to_dataset(output_table, base_path, use_legacy_dataset=True,

Review Comment:
   Hm, thinking aloud: `existing_data_behavior` controls how the dataset will 
handle data that already exists. If I implement a unique way of writing parquet 
files when using the new API in `wrtie_to_dataset` I will also have to set 
`existing_data_behavior` to be `overwrite_or_ignore`. That will then make 
trouble when exposing the same parameter in 
https://issues.apache.org/jira/browse/ARROW-15757.
   
   I could do a check if the parameter is specified or not. But am not sure if 
there will be additional complications.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to