dramaticlly opened a new issue, #7975:
URL: https://github.com/apache/iceberg/issues/7975

   ### Feature Request / Improvement
   
   Since #6682 provide major deletion performance improvement on S3 and hadoop 
fileIO, I think we can try to add SupportsBulkOperations for 
https://github.com/apache/iceberg/blob/master/core/src/main/java/org/apache/iceberg/io/ResolvingFileIO.java.
 
   
   Today the ResolvingFileIO do dynamic fileIO loading based on location 
scheme, which are essentially fall to 
   - S3FileIO
   - HadoopFileIO
   and both implement 
[`SupportsBulkOperations`](https://github.com/apache/iceberg/blob/master/api/src/main/java/org/apache/iceberg/io/SupportsBulkOperations.java)
 interface.
   
   However I realized the its method is given a iterable of location instead of 
singular, 
   ```java
   void deleteFiles(Iterable<String> pathsToDelete) throws 
BulkDeletionFailureException;
   ```
   so it does not fit out of box with current hashmap dynamic loading from 
scheme to its fileIO impl. 
   
   ## Questions
   does it make sense to group prefix by scheme and delegate bulk deletion to 
dynamic IO?
   
   
   @rdblue and @nastra do you have any suggestions on this?
   
   
   ### Query engine
   
   None


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to