[
https://issues.apache.org/jira/browse/HADOOP-14531?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16068420#comment-16068420
]
Steve Loughran commented on HADOOP-14531:
-----------------------------------------
Implementing the core of this in HADOOP-13786, but not retrofitting it to
operations inside S3AFS/ block output stream.
# new exceptions for specific failures which merit special retry policy (bad
request => fail, throttle => retry with throttle policy, redirect => fail).
# new retry policy which sets that up, also implements failfast for
unrecoverable network errors
# AwsLambda adds a retry() call which retries around the executed operation
I'm using this as appropriate in commit/abort calls, all of which I'm
considering repeatable other than delete calls
> Improve S3A error handling & reporting
> --------------------------------------
>
> Key: HADOOP-14531
> URL: https://issues.apache.org/jira/browse/HADOOP-14531
> Project: Hadoop Common
> Issue Type: Improvement
> Components: fs/s3
> Affects Versions: 2.8.1
> Reporter: Steve Loughran
>
> Improve S3a error handling and reporting
> this includes
> # looking at error codes and translating to more specific exceptions
> # better retry logic where present
> # adding retry logic where not present
> # more diagnostics in exceptions
> # docs
> Overall goals
> * things that can be retried and will go away are retried for a bit
> * things that don't go away when retried failfast (302, no auth, unknown
> host, connection refused)
> * meaningful exceptions are built in translate exception
> * diagnostics are included, where possible
> * our troubleshooting docs are expanded with new failures we encounter
> AWS S3 error codes:
> http://docs.aws.amazon.com/AmazonS3/latest/API/ErrorResponses.html
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]