Wes McKinney created ARROW-2306: ----------------------------------- Summary: [Python] HDFS test failures Key: ARROW-2306 URL: https://issues.apache.org/jira/browse/ARROW-2306 Project: Apache Arrow Issue Type: Bug Components: Python Reporter: Wes McKinney Assignee: Wes McKinney Fix For: 0.9.0
These weren't caught because we aren't running the HDFS tests in Travis CI {code} pyarrow/tests/test_hdfs.py::TestLibHdfs::test_write_to_dataset_no_partitions FAILED >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> traceback >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> self = <pyarrow.tests.test_hdfs.TestLibHdfs testMethod=test_write_to_dataset_no_partitions> @test_parquet.parquet def test_write_to_dataset_no_partitions(self): tmpdir = pjoin(self.tmp_path, 'write-no_partitions-' + guid()) self.hdfs.mkdir(tmpdir) test_parquet._test_write_to_dataset_no_partitions( > tmpdir, filesystem=self.hdfs) pyarrow/tests/test_hdfs.py:367: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ pyarrow/tests/test_parquet.py:1475: in _test_write_to_dataset_no_partitions filesystem=filesystem) pyarrow/parquet.py:1059: in write_to_dataset _mkdir_if_not_exists(fs, root_path) pyarrow/parquet.py:1006: in _mkdir_if_not_exists if fs._isfilestore() and not fs.exists(path): _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <pyarrow.hdfs.HadoopFileSystem object at 0x7f09e87a4c48> def _isfilestore(self): """ Returns True if this FileSystem is a unix-style file store with directories. """ > raise NotImplementedError E NotImplementedError pyarrow/filesystem.py:143: NotImplementedError >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> entering PDB >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > /home/wesm/code/arrow/python/pyarrow/filesystem.py(143)_isfilestore() -> raise NotImplementedError (Pdb) c pyarrow/tests/test_hdfs.py::TestLibHdfs::test_write_to_dataset_with_partitions FAILED >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> traceback >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> self = <pyarrow.tests.test_hdfs.TestLibHdfs testMethod=test_write_to_dataset_with_partitions> @test_parquet.parquet def test_write_to_dataset_with_partitions(self): tmpdir = pjoin(self.tmp_path, 'write-partitions-' + guid()) self.hdfs.mkdir(tmpdir) test_parquet._test_write_to_dataset_with_partitions( > tmpdir, filesystem=self.hdfs) pyarrow/tests/test_hdfs.py:360: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ pyarrow/tests/test_parquet.py:1433: in _test_write_to_dataset_with_partitions filesystem=filesystem) pyarrow/parquet.py:1059: in write_to_dataset _mkdir_if_not_exists(fs, root_path) pyarrow/parquet.py:1006: in _mkdir_if_not_exists if fs._isfilestore() and not fs.exists(path): _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <pyarrow.hdfs.HadoopFileSystem object at 0x7f09e87a4c48> def _isfilestore(self): """ Returns True if this FileSystem is a unix-style file store with directories. """ > raise NotImplementedError E NotImplementedError pyarrow/filesystem.py:143: NotImplementedError >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> entering PDB >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>> > /home/wesm/code/arrow/python/pyarrow/filesystem.py(143)_isfilestore() -> raise NotImplementedError (Pdb) c {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)