[
https://issues.apache.org/jira/browse/HADOOP-17139?focusedWorklogId=623591&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-623591
]
ASF GitHub Bot logged work on HADOOP-17139:
-------------------------------------------
Author: ASF GitHub Bot
Created on: 16/Jul/21 13:33
Start Date: 16/Jul/21 13:33
Worklog Time Spent: 10m
Work Description: bogthe commented on a change in pull request #3101:
URL: https://github.com/apache/hadoop/pull/3101#discussion_r671257936
##########
File path:
hadoop-common-project/hadoop-common/src/site/markdown/filesystem/filesystem.md
##########
@@ -1419,6 +1419,75 @@ operations related to the part of the file being
truncated is undefined.
+### `boolean copyFromLocalFile(boolean delSrc, boolean overwrite, Path src,
Path dst)`
+
+The source file or directory at `src` is on the local disk and is copied into
the file system at
+destination `dst`. If the source should be deleted after the move then
`delSrc` flag needs to be
+set to TRUE. If destination already exists, and the destination contents
should be overwritten
+then `overwrite` flag should be set to TRUE.
+
+#### Preconditions
+
+The source file or directory must exist:
+
+ if not exists(FS, src) : raise FileNotFoundException
+
+Directories cannot be copied into files regardless to what the overwrite flag
is set to:
+
+ if isDir(FS, src) && isFile(FS, dst) : raise PathExistsException
+
+If destination exists and the above precondition holds then the overwrite flag
must be set to TRUE
Review comment:
Bad wording on my part; the `above precondition` was referring to the
fact that `Directories cannot be copied into files regardless to what the
overwrite flag is set to`, so what the sentence should say is: `For all cases,
except the one for which the above precondition throws an exception, the
overwrite flag must be set to TRUE for the operation to succeed. This will also
overwrite any files / directories at the destination.`
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
Issue Time Tracking
-------------------
Worklog Id: (was: 623591)
Time Spent: 4h 50m (was: 4h 40m)
> Re-enable optimized copyFromLocal implementation in S3AFileSystem
> -----------------------------------------------------------------
>
> Key: HADOOP-17139
> URL: https://issues.apache.org/jira/browse/HADOOP-17139
> Project: Hadoop Common
> Issue Type: Sub-task
> Components: fs/s3
> Affects Versions: 3.3.0, 3.2.1
> Reporter: Sahil Takiar
> Assignee: Bogdan Stolojan
> Priority: Minor
> Labels: pull-request-available
> Time Spent: 4h 50m
> Remaining Estimate: 0h
>
> It looks like HADOOP-15932 disabled the optimized copyFromLocal
> implementation in S3A for correctness reasons. innerCopyFromLocalFile should
> be fixed and re-enabled. The current implementation uses
> FileSystem.copyFromLocal which will open an input stream from the local fs
> and an output stream to the destination fs, and then call IOUtils.copyBytes.
> With default configs, this will cause S3A to read the file into memory, write
> it back to a file on the local fs, and then when the file is closed, upload
> it to S3.
> The optimized version of copyFromLocal in innerCopyFromLocalFile, directly
> creates a PutObjectRequest request with the local file as the input.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]