Lance Dacey created ARROW-11453: ----------------------------------- Summary: [Python] [Dataset] Unable to use write_dataset() to Azure Blob with adlfs 0.6.0 Key: ARROW-11453 URL: https://issues.apache.org/jira/browse/ARROW-11453 Project: Apache Arrow Issue Type: Bug Components: Python Affects Versions: 3.0.0 Environment: This environment results in an error:
adlfs v0.6.0 fsspec 0.8.5 azure.storage.blob 12.6.0 adal 1.2.6 pandas 1.2.1 pyarrow 3.0.0 Reporter: Lance Dacey https://github.com/dask/adlfs/issues/171 I am unable to save data to Azure Blob using ds.write_dataset() with pyarrow 3.0 and adlfs 0.6.0. Reverting to 0.5.9 fixes the issue, but I am not sure what the cause is - posting this here in case there were filesystem changes in pyarrow recently which are incompatible with changes made in adlfs. {code:java} File "pyarrow/_dataset.pyx", line 2343, in pyarrow._dataset._filesystemdataset_write File "pyarrow/_fs.pyx", line 1032, in pyarrow._fs._cb_create_dir File "/opt/conda/lib/python3.8/site-packages/pyarrow/fs.py", line 259, in create_dir self.fs.mkdir(path, create_parents=recursive) File "/opt/conda/lib/python3.8/site-packages/fsspec/asyn.py", line 121, in wrapper return maybe_sync(func, self, *args, **kwargs) File "/opt/conda/lib/python3.8/site-packages/fsspec/asyn.py", line 100, in maybe_sync return sync(loop, func, *args, **kwargs) File "/opt/conda/lib/python3.8/site-packages/fsspec/asyn.py", line 71, in sync raise exc.with_traceback(tb) File "/opt/conda/lib/python3.8/site-packages/fsspec/asyn.py", line 55, in f result[0] = await future File "/opt/conda/lib/python3.8/site-packages/adlfs/spec.py", line 1033, in _mkdir raise FileExistsError( FileExistsError: Cannot overwrite existing Azure container -- dev already exists. {code} -- This message was sent by Atlassian Jira (v8.3.4#803005)