[ 
https://issues.apache.org/jira/browse/HADOOP-17139?focusedWorklogId=623592&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-623592
 ]

ASF GitHub Bot logged work on HADOOP-17139:
-------------------------------------------

                Author: ASF GitHub Bot
            Created on: 16/Jul/21 13:33
            Start Date: 16/Jul/21 13:33
    Worklog Time Spent: 10m 
      Work Description: bogthe commented on a change in pull request #3101:
URL: https://github.com/apache/hadoop/pull/3101#discussion_r671257936



##########
File path: 
hadoop-common-project/hadoop-common/src/site/markdown/filesystem/filesystem.md
##########
@@ -1419,6 +1419,75 @@ operations related to the part of the file being 
truncated is undefined.
 
 
 
+### `boolean copyFromLocalFile(boolean delSrc, boolean overwrite, Path src, 
Path dst)`
+
+The source file or directory at `src` is on the local disk and is copied into 
the file system at
+destination `dst`. If the source should be deleted after the move then 
`delSrc` flag needs to be
+set to TRUE. If destination already exists, and the destination contents 
should be overwritten
+then `overwrite` flag should be set to TRUE.
+
+#### Preconditions
+
+The source file or directory must exist:
+
+    if not exists(FS, src) : raise FileNotFoundException
+
+Directories cannot be copied into files regardless to what the overwrite flag 
is set to:
+
+    if isDir(FS, src) && isFile(FS, dst) : raise PathExistsException
+
+If destination exists and the above precondition holds then the overwrite flag 
must be set to TRUE

Review comment:
       Bad wording on my part; the `above precondition` was referring to the 
fact that `Directories cannot be copied into files regardless to what the 
overwrite flag is set to`, so what the sentence should say is: `For all cases, 
except the one for which the above precondition throws, the overwrite flag must 
be set to TRUE for the operation to succeed. This will also overwrite any files 
/ directories at the destination.`




-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


Issue Time Tracking
-------------------

    Worklog Id:     (was: 623592)
    Time Spent: 5h  (was: 4h 50m)

> Re-enable optimized copyFromLocal implementation in S3AFileSystem
> -----------------------------------------------------------------
>
>                 Key: HADOOP-17139
>                 URL: https://issues.apache.org/jira/browse/HADOOP-17139
>             Project: Hadoop Common
>          Issue Type: Sub-task
>          Components: fs/s3
>    Affects Versions: 3.3.0, 3.2.1
>            Reporter: Sahil Takiar
>            Assignee: Bogdan Stolojan
>            Priority: Minor
>              Labels: pull-request-available
>          Time Spent: 5h
>  Remaining Estimate: 0h
>
> It looks like HADOOP-15932 disabled the optimized copyFromLocal 
> implementation in S3A for correctness reasons.  innerCopyFromLocalFile should 
> be fixed and re-enabled. The current implementation uses 
> FileSystem.copyFromLocal which will open an input stream from the local fs 
> and an output stream to the destination fs, and then call IOUtils.copyBytes. 
> With default configs, this will cause S3A to read the file into memory, write 
> it back to a file on the local fs, and then when the file is closed, upload 
> it to S3.
> The optimized version of copyFromLocal in innerCopyFromLocalFile, directly 
> creates a PutObjectRequest request with the local file as the input.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to