[
https://issues.apache.org/jira/browse/HDFS-5916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13896902#comment-13896902
]
Sergey Shelukhin commented on HDFS-5916:
----------------------------------------
1-2-3 are both up to you, for the case I have in mind it should operate like a
sequence of regular deletes, for (1) probably best-effort, 2 - no, 3 -
non-atomically. But that could be controlled by parameters.
4 - what do other operations do? As far as I recall some of them can recover.
Can you provide details on how to enforce multiple RPC calls in one for this
case? We currently use FileSystem/DistributedFileSystem interface.
The workaround wouldn't work, due to legacy users as well as due to the fact
that the files/dirs are already in the same path, it's just that we don't want
to delete all of them - e.g. from /path/A, /path/B/, /path/C/ and /path/D we
only want to delete B and D (of course with longer lists)
> provide API to bulk delete directories/files
> --------------------------------------------
>
> Key: HDFS-5916
> URL: https://issues.apache.org/jira/browse/HDFS-5916
> Project: Hadoop HDFS
> Issue Type: Improvement
> Reporter: Sergey Shelukhin
>
> It would be nice to have an API to delete directories and files in bulk - for
> example, when deleting Hive partitions or HBase regions in large numbers, the
> code could avoid many trips to NN.
--
This message was sent by Atlassian JIRA
(v6.1.5#6160)