shaofengshi opened a new issue, #76:
URL: https://github.com/apache/gravitino-playground/issues/76

   Start the playground docker images, and then open Jupyter, run the 
gravitino-fileset-example.ipynb.
   
   In the second step, I got this error:
   
   ```
   from hdfs import InsecureClient
   
   # Create a HDFS connector client
   hdfs_client = InsecureClient('http://hive:50070', user='root')
   
   # List HDFS file and directories
   print(hdfs_client.list('/user/datastrato'))
   
   
   HdfsError                                 Traceback (most recent call last)
   Cell In[2], line 7
         4 hdfs_client = InsecureClient('http://hive:50070/', user='root')
         6 # List HDFS file and directories
   ----> 7 print(hdfs_client.list('/user/datastrato'))
         9 # hdfs_client.delete("/user/datastrato")
   
   File /opt/conda/lib/python3.11/site-packages/hdfs/client.py:1118, in 
Client.list(self, hdfs_path, status)
      1116 _logger.info('Listing %r.', hdfs_path)
      1117 hdfs_path = self.resolve(hdfs_path)
   -> 1118 statuses = 
self._list_status(hdfs_path).json()['FileStatuses']['FileStatus']
      1119 if len(statuses) == 1 and (
      1120   not statuses[0]['pathSuffix'] or self.status(hdfs_path)['type'] == 
'FILE'
      1121   # HttpFS behaves incorrectly here, we sometimes need an extra call 
to
      1122   # make sure we always identify if we are dealing with a file.
      1123 ):
      1124   raise HdfsError('%r is not a directory.', hdfs_path)
   
   File /opt/conda/lib/python3.11/site-packages/hdfs/client.py:118, in 
_Request.to_method.<locals>.api_handler(client, hdfs_path, data, strict, 
**params)
       116   if err.exception not in ('RetriableException', 'StandbyException'):
       117     if strict:
   --> 118       raise err
       119     return res
       121 attempted_hosts.add(host)
   
   HdfsError: File /user/datastrato does not exist.
   ```
   
   Then I look into the HDFS, there is no "/user/datastrato" folder; Under 
"/user" there is only "/user/hive"
   
   `root@3c98b46a73b9:/# hadoop fs -ls /user
   OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=256m; support 
was removed in 8.0
   24/09/19 06:50:14 WARN util.NativeCodeLoader: Unable to load native-hadoop 
library for your platform... using builtin-java classes where applicable
   Found 1 items
   drwxr-xr-x   - root hdfs          0 2024-09-19 06:32 /user/hive
   `
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to