[ https://issues.apache.org/jira/browse/HADOOP-12376?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Steve Loughran resolved HADOOP-12376. ------------------------------------- Resolution: Won't Fix Closing as a wontfix as there is a solution "move to S3a". That has a lot of logic about when to skip forwards vs. close, seek optimisation for different io policies, metrics on all of this, etc. > S3NInputStream.close() downloads the remaining bytes of the object from S3 > -------------------------------------------------------------------------- > > Key: HADOOP-12376 > URL: https://issues.apache.org/jira/browse/HADOOP-12376 > Project: Hadoop Common > Issue Type: Bug > Components: fs/s3 > Affects Versions: 2.6.0, 2.7.1 > Reporter: Steve Loughran > Assignee: Ajith S > > This is the same as HADOOP-11570, possibly the swift code has the same > problem. > Apparently (as raised on ASF lists), when you close an s3n input stream, it > reads through the remainder of the file. This kills performance on partial > reads of large files. -- This message was sent by Atlassian JIRA (v6.3.15#6346) --------------------------------------------------------------------- To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-dev-h...@hadoop.apache.org