[
https://issues.apache.org/jira/browse/HADOOP-15469?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16478057#comment-16478057
]
Aaron Fabbri commented on HADOOP-15469:
---------------------------------------
Interesting. The argument for this seems to be (1) this is a case that works
with FileOutputCommitter and (2) this does not harm any important uses of job
commit conflict resolution. The current docs seem to be congruent with this:
{quote}
The Directory Committer uses the entire directory tree for conflict resolution.
If any file exists at the destination it will fail in job setup; if the
resolution
mechanism is "replace" then all existing files will be deleted.
{quote}
I didn't notice any docs that really need updating here.
Any risks of this change? I'm not thinking of any.
> S3A directory committer commit job fails if _temporary directory created
> under dest
> -----------------------------------------------------------------------------------
>
> Key: HADOOP-15469
> URL: https://issues.apache.org/jira/browse/HADOOP-15469
> Project: Hadoop Common
> Issue Type: Sub-task
> Components: fs/s3
> Affects Versions: 3.1.0
> Environment: spark test runs
> Reporter: Steve Loughran
> Assignee: Steve Loughran
> Priority: Major
> Attachments: HADOOP-15469-001.patch
>
>
> The directory staging committer fails in commit job if any temporary
> files/dirs have been created. Spark work can create such a dir for placement
> of absolute files.
> This is because commitJob() looks for the dest dir existing, not containing
> non-hidden files.
> As the comment says, "its kind of superfluous". More specifically, it means
> jobs which would commit with the classic committer & overwrite=false will fail
> Proposed fix: remove the check
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]