[ 
https://issues.apache.org/jira/browse/HADOOP-13130?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15291318#comment-15291318
 ] 

Steve Loughran commented on HADOOP-13130:
-----------------------------------------

Bad request here. The bucket exists, but isn't giving me access. This is 
frankfurt
{code}

testDeleteEmptyDirNonRecursive(org.apache.hadoop.fs.contract.s3a.TestS3AContractDelete)
  Time elapsed: 0.208 sec  <<< ERROR!
org.apache.hadoop.fs.InvalidRequestException: doesBucketExist on 
stevel-frankfurt-3: com.amazonaws.services.s3.model.AmazonS3Exception: Bad 
Request (Service: Amazon S3; Status Code: 400; Error Code: 400 Bad Request; 
Request ID: 49355E46D8DFCA6B), S3 Extended Request ID: 
KRFtrdEbdZdA4Z6ve2exgBmQArLniiq85f/yUf0NC+btW58ExNxoo3Omhe5Cup0QE7ub5lTes5U=
        at 
org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:96)
        at 
org.apache.hadoop.fs.s3a.S3AFileSystem.verifyBucketExists(S3AFileSystem.java:293)
        at 
org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:272)
        at 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2786)
        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:100)
        at 
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2823)
        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2805)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:382)
        at 
org.apache.hadoop.fs.contract.AbstractBondedFSContract.init(AbstractBondedFSContract.java:72)
        at 
org.apache.hadoop.fs.contract.AbstractFSContractTestBase.setup(AbstractFSContractTestBase.java:165)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
        at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
        at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
        at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
        at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
        at 
org.junit.internal.runners.statements.FailOnTimeout$StatementThread.run(FailOnTimeout.java:74)
{code}

> s3a failures can surface as RTEs, not IOEs
> ------------------------------------------
>
>                 Key: HADOOP-13130
>                 URL: https://issues.apache.org/jira/browse/HADOOP-13130
>             Project: Hadoop Common
>          Issue Type: Sub-task
>          Components: fs/s3
>    Affects Versions: 2.7.2
>            Reporter: Steve Loughran
>            Assignee: Steve Loughran
>         Attachments: HADOOP-13130-001.patch, HADOOP-13130-002.patch, 
> HADOOP-13130-002.patch, HADOOP-13130-003.patch, HADOOP-13130-004.patch, 
> HADOOP-13130-005.patch, HADOOP-13130-branch-2-006.patch, 
> HADOOP-13130-branch-2-007.patch, HADOOP-13130-branch-2-008.patch, 
> HADOOP-13130-branch-2-009.patch
>
>
> S3A failures happening in the AWS library surface as 
> {{AmazonClientException}} derivatives, rather than IOEs. As the amazon 
> exceptions are runtime exceptions, any code which catches IOEs for error 
> handling breaks.
> The fix will be to catch and wrap. The hard thing will be to wrap it with 
> meaningful exceptions rather than a generic IOE. Furthermore, if anyone has 
> been catching AWS exceptions, they are going to be disappointed. That means 
> that fixing this situation could be considered "incompatible" —but only for 
> code which contains assumptions about the underlying FS and the exceptions 
> they raise.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org

Reply via email to