[
https://issues.apache.org/jira/browse/HADOOP-12376?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Steve Loughran resolved HADOOP-12376.
-------------------------------------
Resolution: Won't Fix
Closing as a wontfix as there is a solution "move to S3a". That has a lot of
logic about when to skip forwards vs. close, seek optimisation for different io
policies, metrics on all of this, etc.
> S3NInputStream.close() downloads the remaining bytes of the object from S3
> --------------------------------------------------------------------------
>
> Key: HADOOP-12376
> URL: https://issues.apache.org/jira/browse/HADOOP-12376
> Project: Hadoop Common
> Issue Type: Bug
> Components: fs/s3
> Affects Versions: 2.6.0, 2.7.1
> Reporter: Steve Loughran
> Assignee: Ajith S
>
> This is the same as HADOOP-11570, possibly the swift code has the same
> problem.
> Apparently (as raised on ASF lists), when you close an s3n input stream, it
> reads through the remainder of the file. This kills performance on partial
> reads of large files.
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]