mccormickt12 commented on PR #2291:
URL: https://github.com/apache/iceberg-python/pull/2291#issuecomment-3165111733

   this shows that setting the netloc on filesystem creation and having it in 
the path (as is done for the other fs types) doesn't work for hdfs 
   
   ```
   >>> hdfs = 
fs.HadoopFileSystem(host='ltx1-yugioh-cluster01.linkfs.prod-ltx1.atd.prod.linkedin.com',
 port=9000)
   25/08/07 17:21:12 WARN util.NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
   >>> table_base = 
"/jobs/openhouse/cutover_zdt_testing_db/cutover_zdt_testing_table_partitioned_one-f814050d-6416-4fa8-ae85-c63ac74b4567"
   >>> long_table_base = 
"ltx1-yugioh-cluster01.linkfs.prod-ltx1.atd.prod.linkedin.com/jobs/openhouse/cutover_zdt_testing_db/cutover_zdt_testing_table_partitioned_one-f814050d-6416-4fa8-ae85-c63ac74b4567"
   >>> hdfs.get_file_info(fs.FileSelector(table_base))
   25/08/07 17:22:00 WARN shortcircuit.DomainSocketFactory: The short-circuit 
local reads feature cannot be used because libhadoop cannot be loaded.
   [<FileInfo for 
'/jobs/openhouse/cutover_zdt_testing_db/cutover_zdt_testing_table_partitioned_one-f814050d-6416-4fa8-ae85-c63ac74b4567/00000-3ec53886-ceae-46f2-a926-050afb7f95b9.metadata.json':
 type=FileType.File, size=2900>, <FileInfo for 
'/jobs/openhouse/cutover_zdt_testing_db/cutover_zdt_testing_table_partitioned_one-f814050d-6416-4fa8-ae85-c63ac74b4567/00001-fc1f6c92-0449-4deb-8908-097db5f6589a.metadata.json':
 type=FileType.File, size=4366>, <FileInfo for 
'/jobs/openhouse/cutover_zdt_testing_db/cutover_zdt_testing_table_partitioned_one-f814050d-6416-4fa8-ae85-c63ac74b4567/data':
 type=FileType.Directory>, <FileInfo for 
'/jobs/openhouse/cutover_zdt_testing_db/cutover_zdt_testing_table_partitioned_one-f814050d-6416-4fa8-ae85-c63ac74b4567/metadata':
 type=FileType.Directory>]
   >>> hdfs.get_file_info(fs.FileSelector(long_table_base))
   Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File "pyarrow/_fs.pyx", line 582, in pyarrow._fs.FileSystem.get_file_info
     File "pyarrow/error.pxi", line 155, in 
pyarrow.lib.pyarrow_internal_check_status
     File "pyarrow/error.pxi", line 92, in pyarrow.lib.check_status
   FileNotFoundError: [Errno 2] HDFS list directory failed. Detail: [errno 2] 
No such file or directory
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org
For additional commands, e-mail: issues-h...@iceberg.apache.org

Reply via email to