[ 
https://issues.apache.org/jira/browse/HADOOP-16382?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16867782#comment-16867782
 ] 

Steve Loughran commented on HADOOP-16382:
-----------------------------------------

Patch up. Issue observed while trying to debug a mismatch in DDB write counts 
in a HADOOP-15183 test under load; the debug level logs implied that a newly 
written file was still being updated in S3Guard on an existence check, because 
the timestamps were different.

It'd be nice if we actually got a timestamp off AWS on the completion of the 
PUT/POST call, so we'd have the actual value to put into the metadata. As it 
is, we don't, not until the first HEAD/LIST gives us the value.

> clock skew can cause S3Guard to think object metadata is out of date
> --------------------------------------------------------------------
>
>                 Key: HADOOP-16382
>                 URL: https://issues.apache.org/jira/browse/HADOOP-16382
>             Project: Hadoop Common
>          Issue Type: Sub-task
>          Components: fs/s3
>    Affects Versions: 3.3.0
>            Reporter: Steve Loughran
>            Priority: Minor
>
> When a S3Guard entry is added for an object, its last updated flag is taken 
> from the local clock: if a getFileStatus is made immediately afterwards, the 
> timestamp of the file from the HEAD may be > than the local time, so the DDB 
> entry updated.
> This is even if the clocks are *close*. When updating an entry from S3, the 
> actual timestamp of the file should be used to fix it, not local clocks



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to