[
https://issues.apache.org/jira/browse/HADOOP-17244?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17191840#comment-17191840
]
Steve Loughran commented on HADOOP-17244:
-----------------------------------------
This bug is is in {{DeleteOperation.deleteDirectoryTree()}} BTW...not any of
the new code.
Currently we naively list all entries and queue for delete in pages, and
there's clearly a big assumption there: dir markers don't have entries under
them.
What to change?
I'm going to to go with being less incremental
-files get deleted from S3Guard incrementally
-dir markers are not, instead at end of delete the existing cleanup will do its
thing
> ITestS3AFileContextMainOperations#testRenameDirectoryAsNonExistentDirectory
> test failure on -Dauth
> --------------------------------------------------------------------------------------------------
>
> Key: HADOOP-17244
> URL: https://issues.apache.org/jira/browse/HADOOP-17244
> Project: Hadoop Common
> Issue Type: Sub-task
> Components: fs/s3
> Affects Versions: 3.3.1
> Reporter: Steve Loughran
> Assignee: Steve Loughran
> Priority: Blocker
>
> Test failure:
> {{ITestS3AFileContextMainOperations#testRenameDirectoryAsNonExistentDirectory}}
> This is repeatable on -Dauth runs (we haven't been running them, have we?)
> Either its from the recent dir marker changes (initial hypothesis) or its
> been lurking a while and not been picked up.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]