[ 
https://issues.apache.org/jira/browse/HADOOP-15209?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16391493#comment-16391493
 ] 

Steve Loughran commented on HADOOP-15209:
-----------------------------------------

FWIW, managed to fail distcp if you run it against an S3 store w/ simulated 
inconsistency turned on (no s3guard); operation saw duplicate entries in the 
directory listing at the destination.
{code}
2018-03-08 16:33:17,517 [Thread-131] WARN  mapred.LocalJobRunner 
(LocalJobRunner.java:run(590)) - job_local148600535_0001
org.apache.hadoop.tools.CopyListing$DuplicateFileException: File 
s3a://hwdev-steve-frankfurt-new/SLOW/hadoop-auth/src and 
s3a://hwdev-steve-frankfurt-new/SLOW/hadoop-auth/src would cause duplicates. 
Aborting
        at 
org.apache.hadoop.tools.CopyListing.validateFinalListing(CopyListing.java:175)
        at org.apache.hadoop.tools.CopyListing.buildListing(CopyListing.java:93)
        at 
org.apache.hadoop.tools.GlobbedCopyListing.doBuildListing(GlobbedCopyListing.java:89)
        at org.apache.hadoop.tools.CopyListing.buildListing(CopyListing.java:86)
        at 
org.apache.hadoop.tools.mapred.CopyCommitter.listTargetFiles(CopyCommitter.java:575)
        at 
org.apache.hadoop.tools.mapred.CopyCommitter.deleteMissing(CopyCommitter.java:402)
        at 
org.apache.hadoop.tools.mapred.CopyCommitter.commitJob(CopyCommitter.java:117)
        at 
org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:567)
2018-03-08 16:33:18,469 [main] INFO  mapreduce.Job 
(Job.java:monitorAndPrintJob(1660)) - Job job_local148600535_0001 failed with 
state FAILED due to: NA
2018-03-08 16:33:18,478 [main] INFO  mapreduce.Job 
(Job.java:monitorAndPrintJob(1665)) - Counters: 25
        File System Counters
                FILE: Number of bytes read=1621092
                FILE: Number of bytes written=1632776
                FILE: Number of read operations=0
                FILE: Number of large read operations=0
                FILE: Number of write operations=0
                S3A: Number of bytes read=0
                S3A: Number of bytes written=895927
                S3A: Number of read operations=1673
                S3A: Number of large read operations=0
                S3A: Number of write operations=904
        Map-Reduce Framework
                Map input records=96
{code}

I'm not going to fix that here

> DistCp to eliminate needless deletion of files under already-deleted 
> directories
> --------------------------------------------------------------------------------
>
>                 Key: HADOOP-15209
>                 URL: https://issues.apache.org/jira/browse/HADOOP-15209
>             Project: Hadoop Common
>          Issue Type: Improvement
>          Components: tools/distcp
>    Affects Versions: 2.9.0
>            Reporter: Steve Loughran
>            Assignee: Steve Loughran
>            Priority: Major
>         Attachments: HADOOP-15209-001.patch, HADOOP-15209-002.patch, 
> HADOOP-15209-003.patch, HADOOP-15209-004.patch, HADOOP-15209-005.patch, 
> HADOOP-15209-006.patch
>
>
> DistCP issues a delete(file) request even if is underneath an already deleted 
> directory. This generates needless load on filesystems/object stores, and, if 
> the store throttles delete, can dramatically slow down the delete operation.
> If the distcp delete operation can build a history of deleted directories, 
> then it will know when it does not need to issue those deletes.
> Care is needed here to make sure that whatever structure is created does not 
> overload the heap of the process.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org

Reply via email to