jorisvandenbossche commented on code in PR #12811: URL: https://github.com/apache/arrow/pull/12811#discussion_r851128672
########## python/pyarrow/tests/parquet/test_dataset.py: ########## @@ -1290,7 +1290,7 @@ def _test_write_to_dataset_no_partitions(base_path, # Without partitions, append files to root_path n = 5 for i in range(n): - pq.write_to_dataset(output_table, base_path, + pq.write_to_dataset(output_table, base_path, use_legacy_dataset=True, Review Comment: So what I don't understand here is that the `dataset.write_dataset` function has a default to raise an error if there is existing data? But then why doesn't the above test fail with that error? (instead of failing in the test because we now overwrote the files) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: github-unsubscr...@arrow.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org