[jira] [Commented] (HADOOP-17499) AbstractContractStreamIOStatisticsTest fails if read buffer not full

2024-01-18 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-17499?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17808222#comment-17808222
 ] 

Steve Loughran commented on HADOOP-17499:
-

seen this again; should be asserting the bytes read count >= 0 and <= requested 
len

> AbstractContractStreamIOStatisticsTest fails if read buffer not full
> 
>
> Key: HADOOP-17499
> URL: https://issues.apache.org/jira/browse/HADOOP-17499
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3, test
>Affects Versions: 3.3.1
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Minor
>
> {code}
> [ERROR] 
> testInputStreamStatisticRead(org.apache.hadoop.fs.s3a.statistics.ITestS3AContractStreamIOStatistics)
>   Time elapsed: 6.252 s  <<< FAILURE!
> org.junit.ComparisonFailure: [Counter named stream_read_bytes with expected 
> value 129] expected:<[129]L> but was:<[3]L>
>   at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> {code}
> Test should handle all cases where bytes read > 0



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19043) S3A: Regression: ITestS3AOpenCost fails on prefetch test runs

2024-01-17 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19043?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17807774#comment-17807774
 ] 

Steve Loughran commented on HADOOP-19043:
-


{code}
[ERROR] Tests run: 8, Failures: 2, Errors: 2, Skipped: 1, Time elapsed: 52.665 
s <<< FAILURE! - in org.apache.hadoop.fs.s3a.performance.ITestS3AOpenCost
[ERROR] 
testVectorReadPastEOF(org.apache.hadoop.fs.s3a.performance.ITestS3AOpenCost)  
Time elapsed: 5.406 s  <<< ERROR!
java.lang.ClassCastException: 
org.apache.hadoop.fs.s3a.prefetch.S3APrefetchingInputStream cannot be cast to 
org.apache.hadoop.fs.s3a.S3AInputStream
at 
org.apache.hadoop.fs.s3a.performance.ITestS3AOpenCost.assertS3StreamClosed(ITestS3AOpenCost.java:434)
at 
org.apache.hadoop.fs.s3a.performance.ITestS3AOpenCost.lambda$testVectorReadPastEOF$14(ITestS3AOpenCost.java:412)
at 
org.apache.hadoop.fs.s3a.performance.OperationCostValidator.exec(OperationCostValidator.java:167)
at 
org.apache.hadoop.fs.s3a.performance.AbstractS3ACostTest.verifyMetrics(AbstractS3ACostTest.java:318)
at 
org.apache.hadoop.fs.s3a.performance.ITestS3AOpenCost.testVectorReadPastEOF(ITestS3AOpenCost.java:409)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:750)

[ERROR] 
testPositionedReadableReadPastEOF(org.apache.hadoop.fs.s3a.performance.ITestS3AOpenCost)
  Time elapsed: 5.929 s  <<< FAILURE!
org.junit.ComparisonFailure: expected:<[16]> but was:<[-1]>
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at 
org.apache.hadoop.fs.s3a.performance.ITestS3AOpenCost.lambda$testPositionedReadableReadPastEOF$12(ITestS3AOpenCost.java:381)
at 
org.apache.hadoop.fs.s3a.performance.OperationCostValidator.exec(OperationCostValidator.java:167)
at 
org.apache.hadoop.fs.s3a.performance.AbstractS3ACostTest.verifyMetrics(AbstractS3ACostTest.java:318)
at 
org.apache.hadoop.fs.s3a.performance.ITestS3AOpenCost.testPositionedReadableReadPastEOF(ITestS3AOpenCost.java:374)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:750)

[ERROR] 
testPositionedReadableReadFullyPastEOF(org.apache.hadoop.fs.s3a.performance.ITestS3AOpenCost)
  Time elapsed: 7.195 s  <<< ERROR!
java.lang.ClassCastExcep

[jira] [Created] (HADOOP-19043) S3A: Regression: ITestS3AOpenCost fails on prefetch test runs

2024-01-17 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19043:
---

 Summary: S3A: Regression: ITestS3AOpenCost fails on prefetch test 
runs
 Key: HADOOP-19043
 URL: https://issues.apache.org/jira/browse/HADOOP-19043
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3, test
Affects Versions: 3.4.0
Reporter: Steve Loughran
Assignee: Steve Loughran


Getting test failures in the new ITestS3AOpenCost tests when run with 
{{-Dprefetch}}

Thought I'd tested this, but clearly not
* class cast failures on asserts (fix: skip)
* bytes read different in one test: (fix: identify and address)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19042) S3A: detect and recover from SSL ConnectionReset exceptions

2024-01-17 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17807685#comment-17807685
 ] 

Steve Loughran commented on HADOOP-19042:
-


{code}
Caused by: javax.net.ssl.SSLException: Connection reset
at sun.security.ssl.Alert.createSSLException(Alert.java:127)
at sun.security.ssl.TransportContext.fatal(TransportContext.java:331)
at sun.security.ssl.TransportContext.fatal(TransportContext.java:274)
at sun.security.ssl.TransportContext.fatal(TransportContext.java:269)
at sun.security.ssl.SSLTransport.decode(SSLTransport.java:138)
at sun.security.ssl.SSLSocketImpl.decode(SSLSocketImpl.java:1400)
at 
sun.security.ssl.SSLSocketImpl.readApplicationRecord(SSLSocketImpl.java:1368)
at sun.security.ssl.SSLSocketImpl.access$300(SSLSocketImpl.java:73)
at 
sun.security.ssl.SSLSocketImpl$AppInputStream.read(SSLSocketImpl.java:962)
at 
com.amazonaws.thirdparty.apache.http.impl.io.SessionInputBufferImpl.streamRead(SessionInputBufferImpl.java:137)
at 
com.amazonaws.thirdparty.apache.http.impl.io.SessionInputBufferImpl.read(SessionInputBufferImpl.java:197)
at 
com.amazonaws.thirdparty.apache.http.impl.io.ContentLengthInputStream.read(ContentLengthInputStream.java:176)
at 
com.amazonaws.thirdparty.apache.http.conn.EofSensorInputStream.read(EofSensorInputStream.java:135)
at java.io.InputStream.skip(InputStream.java:224)
at 
com.amazonaws.internal.SdkFilterInputStream.skip(SdkFilterInputStream.java:96)
at 
com.amazonaws.internal.SdkFilterInputStream.skip(SdkFilterInputStream.java:96)
at 
com.amazonaws.internal.SdkFilterInputStream.skip(SdkFilterInputStream.java:96)
at 
com.amazonaws.internal.SdkFilterInputStream.skip(SdkFilterInputStream.java:96)
at 
com.amazonaws.internal.SdkFilterInputStream.skip(SdkFilterInputStream.java:96)
at 
com.amazonaws.util.LengthCheckInputStream.skip(LengthCheckInputStream.java:182)
at 
com.amazonaws.internal.SdkFilterInputStream.skip(SdkFilterInputStream.java:96)
at 
com.amazonaws.services.s3.internal.S3AbortableInputStream.skip(S3AbortableInputStream.java:155)
at 
com.amazonaws.internal.SdkFilterInputStream.skip(SdkFilterInputStream.java:96)
at 
org.apache.hadoop.fs.s3a.S3AInputStream.seekInStream(S3AInputStream.java:368)
at 
org.apache.hadoop.fs.s3a.S3AInputStream.lambda$lazySeek$1(S3AInputStream.java:431)
at org.apache.hadoop.fs.s3a.Invoker.lambda$maybeRetry$3(Invoker.java:284)
at org.apache.hadoop.fs.s3a.Invoker.once(Invoker.java:122)
at org.apache.hadoop.fs.s3a.Invoker.maybeRetry(Invoker.java:410)
at org.apache.hadoop.fs.s3a.Invoker.maybeRetry(Invoker.java:282)
at org.apache.hadoop.fs.s3a.Invoker.maybeRetry(Invoker.java:326)
at org.apache.hadoop.fs.s3a.S3AInputStream.lazySeek(S3AInputStream.java:427)
at org.apache.hadoop.fs.s3a.S3AInputStream.read(S3AInputStream.java:545)
at java.io.DataInputStream.read(DataInputStream.java:149)
at 
org.apache.hadoop.hbase.io.FileLink$FileLinkInputStream.read(FileLink.java:132)
at java.io.DataInputStream.read(DataInputStream.java:149)
at 
org.apache.hadoop.hbase.io.util.BlockIOUtils.readWithExtraOnHeap(BlockIOUtils.java:130)
at 
org.apache.hadoop.hbase.io.util.BlockIOUtils.readWithExtra(BlockIOUtils.java:163)
at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readAtOffset(HFileBlock.java:1486)
at 
org.apache.hadoop.hbase.io.hfile.HFileBlock$FSReaderImpl.readBlockDataInternal(HFileBlock.java:1717)

{code}


> S3A: detect and recover from SSL ConnectionReset exceptions
> ---
>
> Key: HADOOP-19042
> URL: https://issues.apache.org/jira/browse/HADOOP-19042
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0, 3.3.6
>Reporter: Steve Loughran
>Priority: Major
>
> s3a input stream doesn't recover from SSL exceptions, specifically 
> ConnectionReset
> This is a variant of HADOOP-19027, except it's surfaced on an older release...
> # need to make sure the specific exception is handled by aborting stream and 
> retrying -so map to the new HttpChannelEOFException
> # all of thisd needs to be backported



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-19042) S3A: detect and recover from SSL ConnectionReset exceptions

2024-01-17 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19042:
---

 Summary: S3A: detect and recover from SSL ConnectionReset 
exceptions
 Key: HADOOP-19042
 URL: https://issues.apache.org/jira/browse/HADOOP-19042
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.3.6, 3.4.0
Reporter: Steve Loughran


s3a input stream doesn't recover from SSL exceptions, specifically 
ConnectionReset

This is a variant of HADOOP-19027, except it's surfaced on an older release...
# need to make sure the specific exception is handled by aborting stream and 
retrying -so map to the new HttpChannelEOFException
# all of thisd needs to be backported






--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-19027) S3A: S3AInputStream doesn't recover from HTTP/channel exceptions

2024-01-16 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-19027.
-
Fix Version/s: 3.5.0
   Resolution: Fixed

in 3.5 though I hope to backport to 3.4.1

> S3A: S3AInputStream doesn't recover from HTTP/channel exceptions
> 
>
> Key: HADOOP-19027
> URL: https://issues.apache.org/jira/browse/HADOOP-19027
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.5.0
>
>
> S3AInputStream doesn't seem to recover from Http exceptions raised through 
> HttpClient or through OpenSSL.
> * review the recovery code to make sure it is retrying enough, it looks 
> suspiciously like it doesn't
> * detect the relevant openssl, shaded httpclient and unshaded httpclient 
> exceptions, map to a standard one and treat as comms error in our retry policy
> This is not the same as the load balancer/proxy returning 443/444 which we 
> map to AWSNoResponseException. We can't reuse that as it expects to be 
> created from an 
> {{software.amazon.awssdk.awscore.exception.AwsServiceException}} exception 
> with the relevant fields...changing it could potentially be incompatible.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19039) Hadoop 3.4.0 Highlight big features and improvements.

2024-01-16 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19039?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17807181#comment-17807181
 ] 

Steve Loughran commented on HADOOP-19039:
-

+ java17 support
+ where the logging is right now. This is the first Log4j 2 release.
+ what have we cut?

Here is some markdown text regarding S3A changes.

--

This release of Hadoop moves the S3A connector to Amazon S3 to the V2 SDK. This 
is a significant change which offers a number of new features including the 
ability to work with Amazon S3 Express One Zone Storage -the new high 
performance, single AZ storage class. Inevitably, there will be some surprises 
-there is ongoing work to address some performance regressions and resilience 
to network problems.
* There is documentation on how to migrate to the new SDK -there is explicit 
support for moving V1 credential provider implementations; everything else will 
need rework.
* There is also documentation on the third-party S3 store support; the SDK's 
focus on having coolers specify the AWS region rather than the S3 Service and 
point complicates migration. We have successfully tested it against Dell ECS S3 
and Google GS Storage -validation with other "S3 compatible" stores would be 
very much appreciated.n
* If you encounter issues, please look for HADOOP JIRA entries on the same 
topic -and if there are none create a new issues under 
[HADOOP-18886](https://issues.apache.org/jira/browse/HADOOP-18886) _S3A: AWS 
SDK V2 Migration: stabilization and S3Express_. We will try to Address these as 
soon as they have been identified.

---


> Hadoop 3.4.0 Highlight big features and improvements.
> -
>
> Key: HADOOP-19039
> URL: https://issues.apache.org/jira/browse/HADOOP-19039
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Affects Versions: 3.4.0
>Reporter: Shilun Fan
>Assignee: Shilun Fan
>Priority: Major
>
> While preparing for the release of Hadoop-3.4.0, I've noticed the inclusion 
> of numerous commits in this version.  Therefore, highlighting significant 
> features and improvements becomes crucial.  I've completed the initial 
> version and now seek the review of more experienced partner to ensure the 
> finalization of the version's highlights.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19013) fs.getXattrs(path) for S3FS doesn't have x-amz-server-side-encryption-aws-kms-key-id header.

2024-01-16 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19013?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17807180#comment-17807180
 ] 

Steve Loughran commented on HADOOP-19013:
-

well, serve it up if there's no extra HEAD request for it: but do not copy it

> fs.getXattrs(path) for S3FS doesn't have 
> x-amz-server-side-encryption-aws-kms-key-id header.
> 
>
> Key: HADOOP-19013
> URL: https://issues.apache.org/jira/browse/HADOOP-19013
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.3.6
>Reporter: Mukund Thakur
>Assignee: Mukund Thakur
>Priority: Major
>
> Once a path while uploading has been encrypted with SSE-KMS with a key id and 
> then later when we try to read the attributes of the same file, it doesn't 
> contain the key id information as an attribute. should we add it?



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18717) Move CodecPool getCompressor/getDecompressor logs to DEBUG

2024-01-15 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18717?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18717:

Component/s: common

> Move CodecPool getCompressor/getDecompressor logs to DEBUG
> --
>
> Key: HADOOP-18717
> URL: https://issues.apache.org/jira/browse/HADOOP-18717
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Affects Versions: 3.3.6
>Reporter: Claire McGinty
>Priority: Trivial
>  Labels: pull-request-available
>
> The "Got brand new compressor|decompressor" logs in CodecPool[0] can be quite 
> noisy when reading thousands of blocks and aren't that illuminating for the 
> end user. I'd like to propose moving them from log.info to log.debug if 
> there's no objection.
>  
> [0] 
> https://github.com/apache/hadoop/blob/b737869e01fe3334b948a38fe3835e48873bf3a6/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/CodecPool.java#L149-L195



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18717) Move CodecPool getCompressor/getDecompressor logs to DEBUG

2024-01-15 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18717?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18717:

Affects Version/s: 3.3.6

> Move CodecPool getCompressor/getDecompressor logs to DEBUG
> --
>
> Key: HADOOP-18717
> URL: https://issues.apache.org/jira/browse/HADOOP-18717
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 3.3.6
>Reporter: Claire McGinty
>Priority: Trivial
>  Labels: pull-request-available
>
> The "Got brand new compressor|decompressor" logs in CodecPool[0] can be quite 
> noisy when reading thousands of blocks and aren't that illuminating for the 
> end user. I'd like to propose moving them from log.info to log.debug if 
> there's no objection.
>  
> [0] 
> https://github.com/apache/hadoop/blob/b737869e01fe3334b948a38fe3835e48873bf3a6/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/compress/CodecPool.java#L149-L195



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19037) S3A: S3A: ITestS3AConfiguration failing with region problems

2024-01-12 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19037?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17806094#comment-17806094
 ] 

Steve Loughran commented on HADOOP-19037:
-

stack trace comes from creating sts client with no region setting. either it 
needs to be a default or we use bucket location of test fs as driver

{code}
StsClient stsClient =
STSClientFactory.builder(config, bucket, new 
AnonymousAWSCredentialsProvider(), "",
"").build();

{code}


> S3A: S3A: ITestS3AConfiguration failing with region problems
> 
>
> Key: HADOOP-19037
> URL: https://issues.apache.org/jira/browse/HADOOP-19037
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3, test
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Priority: Major
>
> After commented out the default region in my ~/.aws/config [default} profile, 
> test ITestS3AConfiguration. testS3SpecificSignerOverride() fails
> {code}
> [ERROR] 
> testS3SpecificSignerOverride(org.apache.hadoop.fs.s3a.ITestS3AConfiguration)  
> Time elapsed: 0.054 s  <<< ERROR!
> software.amazon.awssdk.core.exception.SdkClientException: Unable to load 
> region from any of the providers in the chain 
> software.amazon.awssdk.regions.providers.DefaultAwsRegionProviderChain@12c626f8:
>  
> [software.amazon.awssdk.regions.providers.SystemSettingsRegionProvider@ae63559:
>  Unable to load region from system settings. Region must be specified either 
> via environment variable (AWS_REGION) or  system property (aws.region)., 
> software.amazon.awssdk.regions.providers.AwsProfileRegionProvider@6e6cfd4c: 
> No region provided in profile: default, 
> software.amazon.awssdk.regions.providers.InstanceProfileRegionProvider@139147de:
>  EC2 Metadata is disabled. Unable to retrieve region information from EC2 
> Metadata service.]
> {code}
> I'm worried the sdk update has rolled back to the 3.3.x region problems where 
> well-configured developer setups / ec2 deployments hid problems. certainly we 
> can see the code is checking these paths



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19037) S3A: S3A: ITestS3AConfiguration failing with region problems

2024-01-12 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19037?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17806093#comment-17806093
 ] 

Steve Loughran commented on HADOOP-19037:
-


{code}
[ERROR] 
testS3SpecificSignerOverride(org.apache.hadoop.fs.s3a.ITestS3AConfiguration)  
Time elapsed: 0.054 s  <<< ERROR!
software.amazon.awssdk.core.exception.SdkClientException: Unable to load region 
from any of the providers in the chain 
software.amazon.awssdk.regions.providers.DefaultAwsRegionProviderChain@12c626f8:
 
[software.amazon.awssdk.regions.providers.SystemSettingsRegionProvider@ae63559: 
Unable to load region from system settings. Region must be specified either via 
environment variable (AWS_REGION) or  system property (aws.region)., 
software.amazon.awssdk.regions.providers.AwsProfileRegionProvider@6e6cfd4c: No 
region provided in profile: default, 
software.amazon.awssdk.regions.providers.InstanceProfileRegionProvider@139147de:
 EC2 Metadata is disabled. Unable to retrieve region information from EC2 
Metadata service.]
at 
software.amazon.awssdk.core.exception.SdkClientException$BuilderImpl.build(SdkClientException.java:111)
at 
software.amazon.awssdk.regions.providers.AwsRegionProviderChain.getRegion(AwsRegionProviderChain.java:70)
at 
software.amazon.awssdk.awscore.client.builder.AwsDefaultClientBuilder.resolveRegion(AwsDefaultClientBuilder.java:293)
at 
software.amazon.awssdk.utils.AttributeMap$DerivedValue.primeCache(AttributeMap.java:600)
at 
software.amazon.awssdk.utils.AttributeMap$DerivedValue.get(AttributeMap.java:589)
at 
software.amazon.awssdk.utils.AttributeMap$Builder.resolveValue(AttributeMap.java:396)
at 
software.amazon.awssdk.utils.AttributeMap$Builder.internalGet(AttributeMap.java:389)
at 
software.amazon.awssdk.utils.AttributeMap$Builder.access$1300(AttributeMap.java:201)
at 
software.amazon.awssdk.utils.AttributeMap$Builder$1.get(AttributeMap.java:399)
at 
software.amazon.awssdk.awscore.client.builder.AwsDefaultClientBuilder.resolveSigningRegion(AwsDefaultClientBuilder.java:260)
at 
software.amazon.awssdk.utils.AttributeMap$DerivedValue.primeCache(AttributeMap.java:600)
at 
software.amazon.awssdk.utils.AttributeMap$DerivedValue.get(AttributeMap.java:589)
at 
software.amazon.awssdk.utils.AttributeMap$Builder.resolveValue(AttributeMap.java:396)
at java.util.ArrayList.forEach(ArrayList.java:1259)
at 
software.amazon.awssdk.utils.AttributeMap$Builder.build(AttributeMap.java:362)
at 
software.amazon.awssdk.core.client.config.SdkClientConfiguration$Builder.build(SdkClientConfiguration.java:232)
at 
software.amazon.awssdk.awscore.client.builder.AwsDefaultClientBuilder.finalizeAwsConfiguration(AwsDefaultClientBuilder.java:184)
at 
software.amazon.awssdk.awscore.client.builder.AwsDefaultClientBuilder.finalizeChildConfiguration(AwsDefaultClientBuilder.java:161)
at 
software.amazon.awssdk.core.client.builder.SdkDefaultClientBuilder.syncClientConfiguration(SdkDefaultClientBuilder.java:188)
at 
software.amazon.awssdk.services.sts.DefaultStsClientBuilder.buildClient(DefaultStsClientBuilder.java:36)
at 
software.amazon.awssdk.services.sts.DefaultStsClientBuilder.buildClient(DefaultStsClientBuilder.java:25)
at 
software.amazon.awssdk.core.client.builder.SdkDefaultClientBuilder.build(SdkDefaultClientBuilder.java:155)
at 
org.apache.hadoop.fs.s3a.ITestS3AConfiguration.testS3SpecificSignerOverride(ITestS3AConfiguration.java:569)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:750)

{code}


> S3A: S3A: ITestS3AConfiguration failing with region problems
> 
>
> Key: HADOOP-19037
> URL: https://issues.apache.org/jira/browse/HADOOP-19037
> Project: Hadoop Common
>  

[jira] [Created] (HADOOP-19037) S3A: S3A: ITestS3AConfiguration failing with region problems

2024-01-12 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19037:
---

 Summary: S3A: S3A: ITestS3AConfiguration failing with region 
problems
 Key: HADOOP-19037
 URL: https://issues.apache.org/jira/browse/HADOOP-19037
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3, test
Affects Versions: 3.4.0
Reporter: Steve Loughran


After commented out the default region in my ~/.aws/config [default} profile, 
test ITestS3AConfiguration. testS3SpecificSignerOverride() fails


{code}
[ERROR] 
testS3SpecificSignerOverride(org.apache.hadoop.fs.s3a.ITestS3AConfiguration)  
Time elapsed: 0.054 s  <<< ERROR!
software.amazon.awssdk.core.exception.SdkClientException: Unable to load region 
from any of the providers in the chain 
software.amazon.awssdk.regions.providers.DefaultAwsRegionProviderChain@12c626f8:
 
[software.amazon.awssdk.regions.providers.SystemSettingsRegionProvider@ae63559: 
Unable to load region from system settings. Region must be specified either via 
environment variable (AWS_REGION) or  system property (aws.region)., 
software.amazon.awssdk.regions.providers.AwsProfileRegionProvider@6e6cfd4c: No 
region provided in profile: default, 
software.amazon.awssdk.regions.providers.InstanceProfileRegionProvider@139147de:
 EC2 Metadata is disabled. Unable to retrieve region information from EC2 
Metadata service.]

{code}

I'm worried the sdk update has rolled back to the 3.3.x region problems where 
well-configured developer setups / ec2 deployments hid problems. certainly we 
can see the code is checking these paths



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-19033) S3A: disable checksum validation

2024-01-12 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19033?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19033:

Parent Issue: HADOOP-18886  (was: HADOOP-18477)

> S3A: disable checksum validation
> 
>
> Key: HADOOP-19033
> URL: https://issues.apache.org/jira/browse/HADOOP-19033
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
>
> AWS v2 sdk turns on client-side checksum validation; this kills performance
> Given we are using TLS to download from AWS s3, there's implicit channel 
> checksumming going on on, that's along with the IPv4 TCP checksumming.
> We don't need it, all it does is slow us down.
> proposed: disable in DefaultS3ClientFactory
> I don't want to add an option to enable it as it only complicates life (yet 
> another config option), but I am open to persuasion



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-19033) S3A: disable checksum validation

2024-01-12 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19033?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19033:

Affects Version/s: 3.4.0

> S3A: disable checksum validation
> 
>
> Key: HADOOP-19033
> URL: https://issues.apache.org/jira/browse/HADOOP-19033
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
>
> AWS v2 sdk turns on client-side checksum validation; this kills performance
> Given we are using TLS to download from AWS s3, there's implicit channel 
> checksumming going on on, that's along with the IPv4 TCP checksumming.
> We don't need it, all it does is slow us down.
> proposed: disable in DefaultS3ClientFactory
> I don't want to add an option to enable it as it only complicates life (yet 
> another config option), but I am open to persuasion



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18981) Move oncrpc/portmap from hadoop-nfs to hadoop-common

2024-01-11 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18981?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18981:

Fix Version/s: 3.5.0
   (was: 3.4.0)

> Move oncrpc/portmap from hadoop-nfs to hadoop-common
> 
>
> Key: HADOOP-18981
> URL: https://issues.apache.org/jira/browse/HADOOP-18981
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: net
>Affects Versions: 3.4.0
>Reporter: Xing Lin
>Assignee: Xing Lin
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.5.0
>
>
> We want to use udpserver/client for other use cases, rather than only for 
> NFS. One such use case is to export NameNodeHAState for NameNodes via a UDP 
> server. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-19004) S3A: Support Authentication through HttpSigner API

2024-01-11 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19004:

Fix Version/s: 3.5.0

> S3A: Support Authentication through HttpSigner API 
> ---
>
> Key: HADOOP-19004
> URL: https://issues.apache.org/jira/browse/HADOOP-19004
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Harshit Gupta
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.5.0
>
>
> The latest AWS SDK changes how signing works, and for signing S3Express 
> signatures the new {{software.amazon.awssdk.http.auth}} auth mechanism is 
> needed



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-19004) S3A: Support Authentication through HttpSigner API

2024-01-11 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19004:

Fix Version/s: (was: 3.4.0)

> S3A: Support Authentication through HttpSigner API 
> ---
>
> Key: HADOOP-19004
> URL: https://issues.apache.org/jira/browse/HADOOP-19004
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Harshit Gupta
>Priority: Major
>  Labels: pull-request-available
>
> The latest AWS SDK changes how signing works, and for signing S3Express 
> signatures the new {{software.amazon.awssdk.http.auth}} auth mechanism is 
> needed



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-19004) S3A: Support Authentication through HttpSigner API

2024-01-11 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-19004.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

> S3A: Support Authentication through HttpSigner API 
> ---
>
> Key: HADOOP-19004
> URL: https://issues.apache.org/jira/browse/HADOOP-19004
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Harshit Gupta
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> The latest AWS SDK changes how signing works, and for signing S3Express 
> signatures the new {{software.amazon.awssdk.http.auth}} auth mechanism is 
> needed



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18981) Move oncrpc/portmap from hadoop-nfs to hadoop-common

2024-01-11 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18981?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18981:

Component/s: net

> Move oncrpc/portmap from hadoop-nfs to hadoop-common
> 
>
> Key: HADOOP-18981
> URL: https://issues.apache.org/jira/browse/HADOOP-18981
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: net
>Affects Versions: 3.4.0
>Reporter: Xing Lin
>Assignee: Xing Lin
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> We want to use udpserver/client for other use cases, rather than only for 
> NFS. One such use case is to export NameNodeHAState for NameNodes via a UDP 
> server. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18981) Move oncrpc/portmap from hadoop-nfs to hadoop-common

2024-01-11 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18981?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-18981.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

> Move oncrpc/portmap from hadoop-nfs to hadoop-common
> 
>
> Key: HADOOP-18981
> URL: https://issues.apache.org/jira/browse/HADOOP-18981
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 3.4.0
>Reporter: Xing Lin
>Assignee: Xing Lin
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> We want to use udpserver/client for other use cases, rather than only for 
> NFS. One such use case is to export NameNodeHAState for NameNodes via a UDP 
> server. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-19033) S3A: disable checksum validation

2024-01-10 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19033:
---

 Summary: S3A: disable checksum validation
 Key: HADOOP-19033
 URL: https://issues.apache.org/jira/browse/HADOOP-19033
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Reporter: Steve Loughran
Assignee: Steve Loughran


AWS v2 sdk turns on client-side checksum validation; this kills performance

Given we are using TLS to download from AWS s3, there's implicit channel 
checksumming going on on, that's along with the IPv4 TCP checksumming.

We don't need it, all it does is slow us down.

proposed: disable in DefaultS3ClientFactory

I don't want to add an option to enable it as it only complicates life (yet 
another config option), but I am open to persuasion




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-19032) S3A: ITestS3AFileContextURI: MultiObjectDeleteException bulk delete of odd filenames

2024-01-10 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19032?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19032:

Summary: S3A: ITestS3AFileContextURI: MultiObjectDeleteException bulk 
delete of odd filenames  (was: S3A: ITestS3AFileContextURI: 
MultiObjectDeleteException bulk delete of odd filenames in )

> S3A: ITestS3AFileContextURI: MultiObjectDeleteException bulk delete of odd 
> filenames
> 
>
> Key: HADOOP-19032
> URL: https://issues.apache.org/jira/browse/HADOOP-19032
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Priority: Major
>
> Possibly transient. note bucket is versioned.
> {code}
> org.apache.hadoop.fs.s3a.AWSS3IOException: 
> Remove S3 Dir Markers on 
> s3a://stevel-london/Users/stevel/Projects/hadoop-trunk/hadoop-tools/hadoop-aws/target/test-dir/7/testContextURI/createTest:
>  org.apache.hadoop.fs.s3a.impl.MultiObjectDeleteException: 
> [S3Error(Key=Users/stevel/Projects/hadoop-trunk/hadoop-tools/hadoop-aws/target/test-dir/7/testContextURI/createTest/()&^%$#@!~_+}{>  Code=InternalError, Message=We encountered an internal error. Please try 
> again.)] (Service: Amazon S3, Status Code: 200, Request ID: 
> null):MultiObjectDeleteException: 
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-19032) S3A: ITestS3AFileContextURI: MultiObjectDeleteException bulk delete of odd filenames in

2024-01-10 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19032?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19032:

Summary: S3A: ITestS3AFileContextURI: MultiObjectDeleteException bulk 
delete of odd filenames in   (was: MultiObjectDeleteException bulk delete of 
odd filenames)

> S3A: ITestS3AFileContextURI: MultiObjectDeleteException bulk delete of odd 
> filenames in 
> 
>
> Key: HADOOP-19032
> URL: https://issues.apache.org/jira/browse/HADOOP-19032
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Priority: Major
>
> Possibly transient. note bucket is versioned.
> {code}
> org.apache.hadoop.fs.s3a.AWSS3IOException: 
> Remove S3 Dir Markers on 
> s3a://stevel-london/Users/stevel/Projects/hadoop-trunk/hadoop-tools/hadoop-aws/target/test-dir/7/testContextURI/createTest:
>  org.apache.hadoop.fs.s3a.impl.MultiObjectDeleteException: 
> [S3Error(Key=Users/stevel/Projects/hadoop-trunk/hadoop-tools/hadoop-aws/target/test-dir/7/testContextURI/createTest/()&^%$#@!~_+}{>  Code=InternalError, Message=We encountered an internal error. Please try 
> again.)] (Service: Amazon S3, Status Code: 200, Request ID: 
> null):MultiObjectDeleteException: 
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19032) MultiObjectDeleteException bulk delete of odd filenames

2024-01-10 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19032?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17805169#comment-17805169
 ] 

Steve Loughran commented on HADOOP-19032:
-


{code}
[ERROR] Tests run: 17, Failures: 0, Errors: 1, Skipped: 1, Time elapsed: 48.129 
s <<< FAILURE! - in org.apache.hadoop.fs.s3a.fileContext.ITestS3AFileContextURI
[ERROR] 
testCreateDirectory(org.apache.hadoop.fs.s3a.fileContext.ITestS3AFileContextURI)
  Time elapsed: 13.4 s  <<< ERROR!
org.apache.hadoop.fs.s3a.AWSS3IOException: 
Remove S3 Dir Markers on 
s3a://stevel-london/Users/stevel/Projects/hadoop-trunk/hadoop-tools/hadoop-aws/target/test-dir/7/testContextURI/createTest:
 org.apache.hadoop.fs.s3a.impl.MultiObjectDeleteException: 
[S3Error(Key=Users/stevel/Projects/hadoop-trunk/hadoop-tools/hadoop-aws/target/test-dir/7/testContextURI/createTest/()&^%$#@!~_+}{> MultiObjectDeleteException bulk delete of odd filenames
> ---
>
> Key: HADOOP-19032
> URL: https://issues.apache.org/jira/browse/HADOOP-19032
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Priority: Major
>
> Possibly transient. note bucket is versioned.
> {code}
> org.apache.hadoop.fs.s3a.AWSS3IOException: 
> Remove S3 Dir Markers on 
> s3a://stevel-london/Users/stevel/Projects/hadoop-trunk/hadoop-tools/hadoop-aws/target/test-dir/7/testContextURI/createTest:
>  org.apache.hadoop.fs.s3a.impl.MultiObjectDeleteException: 
> [S3Error(Key=Users/stevel/Projects/hadoop-trunk/hadoop-tools/hadoop-aws/target/test-dir/7/testContextURI/createTest/()&^%$#@!~_+}{>  Code=InternalError, Message=We encountered an internal error. Please try 
> again.)] (Service: Amazon S3, Status Code: 200, Request ID: 
> null):MultiObjectDeleteException: 
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-19032) MultiObjectDeleteException bulk delete of odd filenames

2024-01-10 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19032:
---

 Summary: MultiObjectDeleteException bulk delete of odd filenames
 Key: HADOOP-19032
 URL: https://issues.apache.org/jira/browse/HADOOP-19032
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.4.0
Reporter: Steve Loughran


Possibly transient. note bucket is versioned.

{code}
org.apache.hadoop.fs.s3a.AWSS3IOException: 
Remove S3 Dir Markers on 
s3a://stevel-london/Users/stevel/Projects/hadoop-trunk/hadoop-tools/hadoop-aws/target/test-dir/7/testContextURI/createTest:
 org.apache.hadoop.fs.s3a.impl.MultiObjectDeleteException: 
[S3Error(Key=Users/stevel/Projects/hadoop-trunk/hadoop-tools/hadoop-aws/target/test-dir/7/testContextURI/createTest/()&^%$#@!~_+}{>

[jira] [Commented] (HADOOP-19027) S3A: S3AInputStream doesn't recover from HTTP/channel exceptions

2024-01-09 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17804710#comment-17804710
 ] 

Steve Loughran commented on HADOOP-19027:
-

doing the mocking tests to simulate all the failures. complex

> S3A: S3AInputStream doesn't recover from HTTP/channel exceptions
> 
>
> Key: HADOOP-19027
> URL: https://issues.apache.org/jira/browse/HADOOP-19027
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
>
> S3AInputStream doesn't seem to recover from Http exceptions raised through 
> HttpClient or through OpenSSL.
> * review the recovery code to make sure it is retrying enough, it looks 
> suspiciously like it doesn't
> * detect the relevant openssl, shaded httpclient and unshaded httpclient 
> exceptions, map to a standard one and treat as comms error in our retry policy
> This is not the same as the load balancer/proxy returning 443/444 which we 
> map to AWSNoResponseException. We can't reuse that as it expects to be 
> created from an 
> {{software.amazon.awssdk.awscore.exception.AwsServiceException}} exception 
> with the relevant fields...changing it could potentially be incompatible.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-19027) S3A: S3AInputStream doesn't recover from HTTP/channel exceptions

2024-01-08 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19027:

Summary: S3A: S3AInputStream doesn't recover from HTTP/channel exceptions  
(was: S3A: S3AInputStream doesn't recover from HTTP exceptions)

> S3A: S3AInputStream doesn't recover from HTTP/channel exceptions
> 
>
> Key: HADOOP-19027
> URL: https://issues.apache.org/jira/browse/HADOOP-19027
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
>
> S3AInputStream doesn't seem to recover from Http exceptions raised through 
> HttpClient or through OpenSSL.
> * review the recovery code to make sure it is retrying enough, it looks 
> suspiciously like it doesn't
> * detect the relevant openssl, shaded httpclient and unshaded httpclient 
> exceptions, map to a standard one and treat as comms error in our retry policy
> This is not the same as the load balancer/proxy returning 443/444 which we 
> map to AWSNoResponseException. We can't reuse that as it expects to be 
> created from an 
> {{software.amazon.awssdk.awscore.exception.AwsServiceException}} exception 
> with the relevant fields...changing it could potentially be incompatible.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19027) S3A: S3AInputStream doesn't recover from HTTP exceptions

2024-01-05 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17803659#comment-17803659
 ] 

Steve Loughran commented on HADOOP-19027:
-

The retry logic in S3AInputStream only seems to retry on the first GET, later 
ones is treats as fatal and assumes this is a file version issue.

This is from https://github.com/apache/hadoop/pull/794
{code}
// With S3Guard, the metadatastore gave us metadata for the file in
// open(), so we use a slightly different retry policy, but only on initial
// open.  After that, an exception generally means the file has changed
// and there is no point retrying anymore.
Invoker invoker = context.getReadInvoker();
invoker.maybeRetry(streamStatistics.openOperations == 0,
"lazySeek", pathStr, true, ...
{code}

# We want to retry on failures, except for version change events.
# we need to map stream problems (closed, no response) to retryable


> S3A: S3AInputStream doesn't recover from HTTP exceptions
> 
>
> Key: HADOOP-19027
> URL: https://issues.apache.org/jira/browse/HADOOP-19027
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
>
> S3AInputStream doesn't seem to recover from Http exceptions raised through 
> HttpClient or through OpenSSL.
> * review the recovery code to make sure it is retrying enough, it looks 
> suspiciously like it doesn't
> * detect the relevant openssl, shaded httpclient and unshaded httpclient 
> exceptions, map to a standard one and treat as comms error in our retry policy
> This is not the same as the load balancer/proxy returning 443/444 which we 
> map to AWSNoResponseException. We can't reuse that as it expects to be 
> created from an 
> {{software.amazon.awssdk.awscore.exception.AwsServiceException}} exception 
> with the relevant fields...changing it could potentially be incompatible.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19027) S3A: S3AInputStream doesn't recover from HTTP exceptions

2024-01-05 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17803654#comment-17803654
 ] 

Steve Loughran commented on HADOOP-19027:
-

full stack trace 

{code}
java.lang.RuntimeException: 
software.amazon.awssdk.thirdparty.org.apache.http.NoHttpResponseException: 
software.amazon.awssdk.core.exception.SdkClientException: Unable to execute 
HTTP request: The target server failed to respond: The target server failed to 
respond
at 
org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:351)
at 
org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:280)
at 
org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:374)
at 
org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:84)
at 
org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:70)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1899)
at 
org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:70)
at 
org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:40)
at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36)
at 
com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
at 
com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:69)
at 
com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
java.io.IOException: java.lang.RuntimeException: 
software.amazon.awssdk.thirdparty.org.apache.http.NoHttpResponseException: 
software.amazon.awssdk.core.exception.SdkClientException: Unable to execute 
HTTP request: The target server failed to respond: The target server failed to 
respond
at 
org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.pushRecord(MapRecordSource.java:80)
at 
org.apache.hadoop.hive.ql.exec.tez.MapRecordProcessor.run(MapRecordProcessor.java:437)
at 
org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:297)
... 16 more
Caused by: java.io.IOException: java.lang.RuntimeException: 
software.amazon.awssdk.thirdparty.org.apache.http.NoHttpResponseException: 
software.amazon.awssdk.core.exception.SdkClientException: Unable to execute 
HTTP request: The target server failed to respond: The target server failed to 
respond
at 
org.apache.hadoop.hive.io.HiveIOExceptionHandlerChain.handleRecordReaderNextException(HiveIOExceptionHandlerChain.java:121)
at 
org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil.handleRecordReaderNextException(HiveIOExceptionHandlerUtil.java:77)
at 
org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:381)
at 
org.apache.hadoop.hive.ql.io.HiveRecordReader.doNext(HiveRecordReader.java:82)
at 
org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.next(HiveContextAwareRecordReader.java:119)
at 
org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.next(HiveContextAwareRecordReader.java:59)
at 
org.apache.hadoop.mapred.split.TezGroupedSplitsInputFormat$TezGroupedSplitsRecordReader.next(TezGroupedSplitsInputFormat.java:151)
at 
org.apache.tez.mapreduce.lib.MRReaderMapred.next(MRReaderMapred.java:116)
at 
org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.pushRecord(MapRecordSource.java:68)
... 18 more
Caused by: java.lang.RuntimeException: 
software.amazon.awssdk.thirdparty.org.apache.http.NoHttpResponseException: 
software.amazon.awssdk.core.exception.SdkClientException: Unable to execute 
HTTP request: The target server failed to respond: The target server failed to 
respond
at 
org.apache.iceberg.mr.hive.vector.HiveBatchIterator.advance(HiveBatchIterator.java:129)
at 
org.apache.iceberg.mr.hive.vector.HiveBatchIterator.hasNext(HiveBatchIterator.java:137)
at 
org.apache.iceberg.mr.hive.vector.HiveDeleteFilter$1$1.hasNext(HiveDeleteFilter.java:98)
at 
org.apache.iceberg.mr.mapreduce.IcebergInputFormat$IcebergRecordReader.nextKeyValue(IcebergInputFormat.java:299)
at 
org.apache.iceberg.mr.hive.vector.HiveIcebergVectorizedRecordReader.next(HiveI

[jira] [Commented] (HADOOP-19027) S3A: S3AInputStream doesn't recover from HTTP exceptions

2024-01-05 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19027?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17803649#comment-17803649
 ] 

Steve Loughran commented on HADOOP-19027:
-

openssl, "stream is closed"

{code}
Error while running task ( failure ) : 
attempt_1703842027450_0084_4_02_04_0:java.lang.RuntimeException: 
org.apache.hadoop.hive.ql.metadata.HiveException: java.io.IOException: 
java.lang.RuntimeException: java.io.IOException: 
software.amazon.awssdk.core.exception.SdkClientException: Unable to execute 
HTTP request: WFOPENSSL0035 Stream is closed: WFOPENSSL0035 Stream is closed
  at 
org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:351)
  at org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:280)
  at 
org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:374)
  at 
org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:84)
  at 
org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:70)
  at java.security.AccessController.doPrivileged(Native Method)
  at javax.security.auth.Subject.doAs(Subject.java:422)
  at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1899)
  at 
org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:70)
  at 
org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:40)
  at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36)
  at 
com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
  at 
com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:69)
  at 
com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
  at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
  at java.lang.Thread.run(Thread.java:750)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: 
java.io.IOException: java.lang.RuntimeException: java.io.IOException: 
software.amazon.awssdk.core.exception.SdkClientException: 
Unable to execute HTTP request: WFOPENSSL0035 Stream is closed: WFOPENSSL0035 
Stream is closed
  at 
org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.pushRecord(MapRecordSource.java:80)
  at 
org.apache.hadoop.hive.ql.exec.tez.MapRecordProcessor.run(MapRecordProcessor.java:437)
  at 
org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:297)
  ... 16 more
Caused by: java.io.IOException: java.lang.RuntimeException: 
java.io.IOException: software.amazon.awssdk.core.exception.SdkClientException: 
Unable to execute HTTP request: WFOPENSSL0035 Stream is closed: WFOPENSSL0035 
Stream is closed
  at 
org.apache.hadoop.hive.io.HiveIOExceptionHandlerChain.handleRecordReaderNextException(HiveIOExceptionHandlerChain.java:121)
  at 
org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil.handleRecordReaderNextException(HiveIOExceptionHandlerUtil.java:77)
  at 
org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.doNext(HiveContextAwareRecordReader.java:381)
  at 
org.apache.hadoop.hive.ql.io.HiveRecordReader.doNext(HiveRecordReader.java:82)
  at 
org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.next(HiveContextAwareRecordReader.java:119)
  at 
org.apache.hadoop.hive.ql.io.HiveContextAwareRecordReader.next(HiveContextAwareRecordReader.java:59)
  at 
org.apache.hadoop.mapred.split.TezGroupedSplitsInputFormat$TezGroupedSplitsRecordReader.next(TezGroupedSplitsInputFormat.java:151)
  at org.apache.tez.mapreduce.lib.MRReaderMapred.next(MRReaderMapred.java:116)
  at 
org.apache.hadoop.hive.ql.exec.tez.MapRecordSource.pushRecord(MapRecordSource.java:68)
  ... 18 more
Caused by: java.lang.RuntimeException: java.io.IOException: 
software.amazon.awssdk.core.exception.SdkClientException: Unable to execute 
HTTP request: WFOPENSSL0035 Stream is closed: WFOPENS
SL0035 Stream is closed
  at 
org.apache.iceberg.mr.hive.vector.HiveBatchIterator.advance(HiveBatchIterator.java:129){noformat}

{code}

and httpclient 
software.amazon.awssdk.thirdparty.org.apache.http.NoHttpResponseException with 
The target server failed to respond""

{code}
Error while running task ( failure ) : 
attempt_1704316855291_0218_13_02_04_0:java.lang.RuntimeException: 
org.apache.hadoop.hive.ql.metadata.HiveException: java.io.IOException: 
java.lang.RuntimeException: 
software.amazon.awssdk.thirdparty.org.apache.http.NoHttpResponseException: 
software.amazon.awssdk.core.exception.SdkClientException: Unable to execute 
HTTP request: The target server failed to respond: The target server failed to 
respond
at 
org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initi

[jira] [Created] (HADOOP-19027) S3A: S3AInputStream doesn't recover from HTTP exceptions

2024-01-05 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19027:
---

 Summary: S3A: S3AInputStream doesn't recover from HTTP exceptions
 Key: HADOOP-19027
 URL: https://issues.apache.org/jira/browse/HADOOP-19027
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.4.0
Reporter: Steve Loughran
Assignee: Steve Loughran




S3AInputStream doesn't seem to recover from Http exceptions raised through 
HttpClient or through OpenSSL.

* review the recovery code to make sure it is retrying enough, it looks 
suspiciously like it doesn't
* detect the relevant openssl, shaded httpclient and unshaded httpclient 
exceptions, map to a standard one and treat as comms error in our retry policy

This is not the same as the load balancer/proxy returning 443/444 which we map 
to AWSNoResponseException. We can't reuse that as it expects to be created from 
an {{software.amazon.awssdk.awscore.exception.AwsServiceException}} exception 
with the relevant fields...changing it could potentially be incompatible.




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18886) S3A: AWS SDK V2 Migration: stabilization and S3Express

2024-01-05 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18886?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18886:

Description: 
The final stabilisation changes to the V2 SDK MIgration; those moved off the 
HADOOP-18073 JIRA so we can close that.

also adds support to Amazon S3 Express One Zone storage

  was:The final stabilisation changes to the V2 SDK MIgration; those moved off 
the HADOOP-18073 JIRA so we can close that.


> S3A: AWS SDK V2 Migration: stabilization and S3Express
> --
>
> Key: HADOOP-18886
> URL: https://issues.apache.org/jira/browse/HADOOP-18886
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: ahmar#1
>Priority: Major
>
> The final stabilisation changes to the V2 SDK MIgration; those moved off the 
> HADOOP-18073 JIRA so we can close that.
> also adds support to Amazon S3 Express One Zone storage



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19023) ITestS3AConcurrentOps#testParallelRename intermittent timeout failure

2024-01-05 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19023?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17803521#comment-17803521
 ] 

Steve Loughran commented on HADOOP-19023:
-

* make sure you've not got a site config with an aggressive timeout
 * do set version/component in the issue fields...it's not picked up from the 
parent

> ITestS3AConcurrentOps#testParallelRename intermittent timeout failure
> -
>
> Key: HADOOP-19023
> URL: https://issues.apache.org/jira/browse/HADOOP-19023
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Viraj Jasani
>Priority: Major
>
> Need to configure higher timeout for the test.
>  
> {code:java}
> [ERROR] Tests run: 2, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 
> 256.281 s <<< FAILURE! - in 
> org.apache.hadoop.fs.s3a.scale.ITestS3AConcurrentOps
> [ERROR] 
> testParallelRename(org.apache.hadoop.fs.s3a.scale.ITestS3AConcurrentOps)  
> Time elapsed: 72.565 s  <<< ERROR!
> org.apache.hadoop.fs.s3a.AWSApiCallTimeoutException: Writing Object on 
> fork-0005/test/testParallelRename-source0: 
> software.amazon.awssdk.core.exception.ApiCallTimeoutException: Client 
> execution did not complete before the specified timeout configuration: 15000 
> millis
>   at 
> org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:215)
>   at org.apache.hadoop.fs.s3a.Invoker.once(Invoker.java:124)
>   at org.apache.hadoop.fs.s3a.Invoker.lambda$retry$4(Invoker.java:376)
>   at org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:468)
>   at org.apache.hadoop.fs.s3a.Invoker.retry(Invoker.java:372)
>   at org.apache.hadoop.fs.s3a.Invoker.retry(Invoker.java:347)
>   at 
> org.apache.hadoop.fs.s3a.WriteOperationHelper.retry(WriteOperationHelper.java:214)
>   at 
> org.apache.hadoop.fs.s3a.WriteOperationHelper.putObject(WriteOperationHelper.java:532)
>   at 
> org.apache.hadoop.fs.s3a.S3ABlockOutputStream.lambda$putObject$0(S3ABlockOutputStream.java:620)
>   at 
> org.apache.hadoop.thirdparty.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
>   at 
> org.apache.hadoop.thirdparty.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:69)
>   at 
> org.apache.hadoop.thirdparty.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
>   at 
> org.apache.hadoop.util.SemaphoredDelegatingExecutor$RunnableWithPermitRelease.run(SemaphoredDelegatingExecutor.java:225)
>   at 
> org.apache.hadoop.util.SemaphoredDelegatingExecutor$RunnableWithPermitRelease.run(SemaphoredDelegatingExecutor.java:225)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:750)
> Caused by: software.amazon.awssdk.core.exception.ApiCallTimeoutException: 
> Client execution did not complete before the specified timeout configuration: 
> 15000 millis
>   at 
> software.amazon.awssdk.core.exception.ApiCallTimeoutException$BuilderImpl.build(ApiCallTimeoutException.java:97)
>   at 
> software.amazon.awssdk.core.exception.ApiCallTimeoutException.create(ApiCallTimeoutException.java:38)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.generateApiCallTimeoutException(ApiCallTimeoutTrackingStage.java:151)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.handleInterruptedException(ApiCallTimeoutTrackingStage.java:139)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.translatePipelineException(ApiCallTimeoutTrackingStage.java:107)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.execute(ApiCallTimeoutTrackingStage.java:62)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.execute(ApiCallTimeoutTrackingStage.java:42)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallMetricCollectionStage.execute(ApiCallMetricCollectionStage.java:50)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallMetricCollectionStage.execute(ApiCallMetricCollectionStage.java:32)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.RequestPipelineBuilder$ComposingRequestPipelineStage.execute(RequestPipelineBuilder.java:206)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.RequestPipelineBuilder$ComposingRequestPipelineS

[jira] [Updated] (HADOOP-19023) ITestS3AConcurrentOps#testParallelRename intermittent timeout failure

2024-01-05 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19023?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19023:

Component/s: fs/s3

> ITestS3AConcurrentOps#testParallelRename intermittent timeout failure
> -
>
> Key: HADOOP-19023
> URL: https://issues.apache.org/jira/browse/HADOOP-19023
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Viraj Jasani
>Priority: Major
>
> Need to configure higher timeout for the test.
>  
> {code:java}
> [ERROR] Tests run: 2, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 
> 256.281 s <<< FAILURE! - in 
> org.apache.hadoop.fs.s3a.scale.ITestS3AConcurrentOps
> [ERROR] 
> testParallelRename(org.apache.hadoop.fs.s3a.scale.ITestS3AConcurrentOps)  
> Time elapsed: 72.565 s  <<< ERROR!
> org.apache.hadoop.fs.s3a.AWSApiCallTimeoutException: Writing Object on 
> fork-0005/test/testParallelRename-source0: 
> software.amazon.awssdk.core.exception.ApiCallTimeoutException: Client 
> execution did not complete before the specified timeout configuration: 15000 
> millis
>   at 
> org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:215)
>   at org.apache.hadoop.fs.s3a.Invoker.once(Invoker.java:124)
>   at org.apache.hadoop.fs.s3a.Invoker.lambda$retry$4(Invoker.java:376)
>   at org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:468)
>   at org.apache.hadoop.fs.s3a.Invoker.retry(Invoker.java:372)
>   at org.apache.hadoop.fs.s3a.Invoker.retry(Invoker.java:347)
>   at 
> org.apache.hadoop.fs.s3a.WriteOperationHelper.retry(WriteOperationHelper.java:214)
>   at 
> org.apache.hadoop.fs.s3a.WriteOperationHelper.putObject(WriteOperationHelper.java:532)
>   at 
> org.apache.hadoop.fs.s3a.S3ABlockOutputStream.lambda$putObject$0(S3ABlockOutputStream.java:620)
>   at 
> org.apache.hadoop.thirdparty.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
>   at 
> org.apache.hadoop.thirdparty.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:69)
>   at 
> org.apache.hadoop.thirdparty.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
>   at 
> org.apache.hadoop.util.SemaphoredDelegatingExecutor$RunnableWithPermitRelease.run(SemaphoredDelegatingExecutor.java:225)
>   at 
> org.apache.hadoop.util.SemaphoredDelegatingExecutor$RunnableWithPermitRelease.run(SemaphoredDelegatingExecutor.java:225)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:750)
> Caused by: software.amazon.awssdk.core.exception.ApiCallTimeoutException: 
> Client execution did not complete before the specified timeout configuration: 
> 15000 millis
>   at 
> software.amazon.awssdk.core.exception.ApiCallTimeoutException$BuilderImpl.build(ApiCallTimeoutException.java:97)
>   at 
> software.amazon.awssdk.core.exception.ApiCallTimeoutException.create(ApiCallTimeoutException.java:38)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.generateApiCallTimeoutException(ApiCallTimeoutTrackingStage.java:151)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.handleInterruptedException(ApiCallTimeoutTrackingStage.java:139)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.translatePipelineException(ApiCallTimeoutTrackingStage.java:107)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.execute(ApiCallTimeoutTrackingStage.java:62)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.execute(ApiCallTimeoutTrackingStage.java:42)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallMetricCollectionStage.execute(ApiCallMetricCollectionStage.java:50)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallMetricCollectionStage.execute(ApiCallMetricCollectionStage.java:32)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.RequestPipelineBuilder$ComposingRequestPipelineStage.execute(RequestPipelineBuilder.java:206)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.RequestPipelineBuilder$ComposingRequestPipelineStage.execute(RequestPipelineBuilder.java:206)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ExecutionFailureExceptionReportingStage.execute(ExecutionFailureExcepti

[jira] [Updated] (HADOOP-19023) ITestS3AConcurrentOps#testParallelRename intermittent timeout failure

2024-01-05 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19023?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19023:

Affects Version/s: 3.4.0

> ITestS3AConcurrentOps#testParallelRename intermittent timeout failure
> -
>
> Key: HADOOP-19023
> URL: https://issues.apache.org/jira/browse/HADOOP-19023
> Project: Hadoop Common
>  Issue Type: Sub-task
>Affects Versions: 3.4.0
>Reporter: Viraj Jasani
>Priority: Major
>
> Need to configure higher timeout for the test.
>  
> {code:java}
> [ERROR] Tests run: 2, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 
> 256.281 s <<< FAILURE! - in 
> org.apache.hadoop.fs.s3a.scale.ITestS3AConcurrentOps
> [ERROR] 
> testParallelRename(org.apache.hadoop.fs.s3a.scale.ITestS3AConcurrentOps)  
> Time elapsed: 72.565 s  <<< ERROR!
> org.apache.hadoop.fs.s3a.AWSApiCallTimeoutException: Writing Object on 
> fork-0005/test/testParallelRename-source0: 
> software.amazon.awssdk.core.exception.ApiCallTimeoutException: Client 
> execution did not complete before the specified timeout configuration: 15000 
> millis
>   at 
> org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:215)
>   at org.apache.hadoop.fs.s3a.Invoker.once(Invoker.java:124)
>   at org.apache.hadoop.fs.s3a.Invoker.lambda$retry$4(Invoker.java:376)
>   at org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:468)
>   at org.apache.hadoop.fs.s3a.Invoker.retry(Invoker.java:372)
>   at org.apache.hadoop.fs.s3a.Invoker.retry(Invoker.java:347)
>   at 
> org.apache.hadoop.fs.s3a.WriteOperationHelper.retry(WriteOperationHelper.java:214)
>   at 
> org.apache.hadoop.fs.s3a.WriteOperationHelper.putObject(WriteOperationHelper.java:532)
>   at 
> org.apache.hadoop.fs.s3a.S3ABlockOutputStream.lambda$putObject$0(S3ABlockOutputStream.java:620)
>   at 
> org.apache.hadoop.thirdparty.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
>   at 
> org.apache.hadoop.thirdparty.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:69)
>   at 
> org.apache.hadoop.thirdparty.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
>   at 
> org.apache.hadoop.util.SemaphoredDelegatingExecutor$RunnableWithPermitRelease.run(SemaphoredDelegatingExecutor.java:225)
>   at 
> org.apache.hadoop.util.SemaphoredDelegatingExecutor$RunnableWithPermitRelease.run(SemaphoredDelegatingExecutor.java:225)
>   at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>   at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>   at java.lang.Thread.run(Thread.java:750)
> Caused by: software.amazon.awssdk.core.exception.ApiCallTimeoutException: 
> Client execution did not complete before the specified timeout configuration: 
> 15000 millis
>   at 
> software.amazon.awssdk.core.exception.ApiCallTimeoutException$BuilderImpl.build(ApiCallTimeoutException.java:97)
>   at 
> software.amazon.awssdk.core.exception.ApiCallTimeoutException.create(ApiCallTimeoutException.java:38)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.generateApiCallTimeoutException(ApiCallTimeoutTrackingStage.java:151)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.handleInterruptedException(ApiCallTimeoutTrackingStage.java:139)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.translatePipelineException(ApiCallTimeoutTrackingStage.java:107)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.execute(ApiCallTimeoutTrackingStage.java:62)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.execute(ApiCallTimeoutTrackingStage.java:42)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallMetricCollectionStage.execute(ApiCallMetricCollectionStage.java:50)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallMetricCollectionStage.execute(ApiCallMetricCollectionStage.java:32)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.RequestPipelineBuilder$ComposingRequestPipelineStage.execute(RequestPipelineBuilder.java:206)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.RequestPipelineBuilder$ComposingRequestPipelineStage.execute(RequestPipelineBuilder.java:206)
>   at 
> software.amazon.awssdk.core.internal.http.pipeline.stages.ExecutionFailureExceptionReportingStage.execute(ExecutionFailureExceptionReportingStage.java:3

[jira] [Updated] (HADOOP-19022) ITestS3AConfiguration#testRequestTimeout failure

2024-01-05 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19022?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19022:

Component/s: fs/s3
 test

> ITestS3AConfiguration#testRequestTimeout failure
> 
>
> Key: HADOOP-19022
> URL: https://issues.apache.org/jira/browse/HADOOP-19022
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3, test
>Affects Versions: 3.4.0
>Reporter: Viraj Jasani
>Priority: Minor
>
> "fs.s3a.connection.request.timeout" should be specified in milliseconds as per
> {code:java}
> Duration apiCallTimeout = getDuration(conf, REQUEST_TIMEOUT,
> DEFAULT_REQUEST_TIMEOUT_DURATION, TimeUnit.MILLISECONDS, Duration.ZERO); 
> {code}
> The test fails consistently because it sets 120 ms timeout which is less than 
> 15s (min network operation duration), and hence gets reset to 15000 ms based 
> on the enforcement.
>  
> {code:java}
> [ERROR] testRequestTimeout(org.apache.hadoop.fs.s3a.ITestS3AConfiguration)  
> Time elapsed: 0.016 s  <<< FAILURE!
> java.lang.AssertionError: Configured fs.s3a.connection.request.timeout is 
> different than what AWS sdk configuration uses internally expected:<12> 
> but was:<15000>
>   at org.junit.Assert.fail(Assert.java:89)
>   at org.junit.Assert.failNotEquals(Assert.java:835)
>   at org.junit.Assert.assertEquals(Assert.java:647)
>   at 
> org.apache.hadoop.fs.s3a.ITestS3AConfiguration.testRequestTimeout(ITestS3AConfiguration.java:444)
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-19022) ITestS3AConfiguration#testRequestTimeout failure

2024-01-05 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19022?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17803520#comment-17803520
 ] 

Steve Loughran edited comment on HADOOP-19022 at 1/5/24 11:51 AM:
--

should be a string now, e.g "20s".

have you explicitly set it in your site config?


was (Author: ste...@apache.org):
should be a string now, e.g "20s"

> ITestS3AConfiguration#testRequestTimeout failure
> 
>
> Key: HADOOP-19022
> URL: https://issues.apache.org/jira/browse/HADOOP-19022
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3, test
>Affects Versions: 3.4.0
>Reporter: Viraj Jasani
>Priority: Minor
>
> "fs.s3a.connection.request.timeout" should be specified in milliseconds as per
> {code:java}
> Duration apiCallTimeout = getDuration(conf, REQUEST_TIMEOUT,
> DEFAULT_REQUEST_TIMEOUT_DURATION, TimeUnit.MILLISECONDS, Duration.ZERO); 
> {code}
> The test fails consistently because it sets 120 ms timeout which is less than 
> 15s (min network operation duration), and hence gets reset to 15000 ms based 
> on the enforcement.
>  
> {code:java}
> [ERROR] testRequestTimeout(org.apache.hadoop.fs.s3a.ITestS3AConfiguration)  
> Time elapsed: 0.016 s  <<< FAILURE!
> java.lang.AssertionError: Configured fs.s3a.connection.request.timeout is 
> different than what AWS sdk configuration uses internally expected:<12> 
> but was:<15000>
>   at org.junit.Assert.fail(Assert.java:89)
>   at org.junit.Assert.failNotEquals(Assert.java:835)
>   at org.junit.Assert.assertEquals(Assert.java:647)
>   at 
> org.apache.hadoop.fs.s3a.ITestS3AConfiguration.testRequestTimeout(ITestS3AConfiguration.java:444)
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-19022) ITestS3AConfiguration#testRequestTimeout failure

2024-01-05 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19022?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19022:

Affects Version/s: 3.4.0

> ITestS3AConfiguration#testRequestTimeout failure
> 
>
> Key: HADOOP-19022
> URL: https://issues.apache.org/jira/browse/HADOOP-19022
> Project: Hadoop Common
>  Issue Type: Sub-task
>Affects Versions: 3.4.0
>Reporter: Viraj Jasani
>Priority: Minor
>
> "fs.s3a.connection.request.timeout" should be specified in milliseconds as per
> {code:java}
> Duration apiCallTimeout = getDuration(conf, REQUEST_TIMEOUT,
> DEFAULT_REQUEST_TIMEOUT_DURATION, TimeUnit.MILLISECONDS, Duration.ZERO); 
> {code}
> The test fails consistently because it sets 120 ms timeout which is less than 
> 15s (min network operation duration), and hence gets reset to 15000 ms based 
> on the enforcement.
>  
> {code:java}
> [ERROR] testRequestTimeout(org.apache.hadoop.fs.s3a.ITestS3AConfiguration)  
> Time elapsed: 0.016 s  <<< FAILURE!
> java.lang.AssertionError: Configured fs.s3a.connection.request.timeout is 
> different than what AWS sdk configuration uses internally expected:<12> 
> but was:<15000>
>   at org.junit.Assert.fail(Assert.java:89)
>   at org.junit.Assert.failNotEquals(Assert.java:835)
>   at org.junit.Assert.assertEquals(Assert.java:647)
>   at 
> org.apache.hadoop.fs.s3a.ITestS3AConfiguration.testRequestTimeout(ITestS3AConfiguration.java:444)
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19022) ITestS3AConfiguration#testRequestTimeout failure

2024-01-05 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19022?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17803520#comment-17803520
 ] 

Steve Loughran commented on HADOOP-19022:
-

should be a string now, e.g "20s"

> ITestS3AConfiguration#testRequestTimeout failure
> 
>
> Key: HADOOP-19022
> URL: https://issues.apache.org/jira/browse/HADOOP-19022
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Viraj Jasani
>Priority: Minor
>
> "fs.s3a.connection.request.timeout" should be specified in milliseconds as per
> {code:java}
> Duration apiCallTimeout = getDuration(conf, REQUEST_TIMEOUT,
> DEFAULT_REQUEST_TIMEOUT_DURATION, TimeUnit.MILLISECONDS, Duration.ZERO); 
> {code}
> The test fails consistently because it sets 120 ms timeout which is less than 
> 15s (min network operation duration), and hence gets reset to 15000 ms based 
> on the enforcement.
>  
> {code:java}
> [ERROR] testRequestTimeout(org.apache.hadoop.fs.s3a.ITestS3AConfiguration)  
> Time elapsed: 0.016 s  <<< FAILURE!
> java.lang.AssertionError: Configured fs.s3a.connection.request.timeout is 
> different than what AWS sdk configuration uses internally expected:<12> 
> but was:<15000>
>   at org.junit.Assert.fail(Assert.java:89)
>   at org.junit.Assert.failNotEquals(Assert.java:835)
>   at org.junit.Assert.assertEquals(Assert.java:647)
>   at 
> org.apache.hadoop.fs.s3a.ITestS3AConfiguration.testRequestTimeout(ITestS3AConfiguration.java:444)
>  {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19025) Migrate ContractTestUtils to AssertJ

2024-01-05 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19025?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17803519#comment-17803519
 ] 

Steve Loughran commented on HADOOP-19025:
-

looking forward to this!

> Migrate ContractTestUtils to AssertJ
> 
>
> Key: HADOOP-19025
> URL: https://issues.apache.org/jira/browse/HADOOP-19025
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: test
>Reporter: Attila Doroszlai
>Assignee: Attila Doroszlai
>Priority: Major
>
> Replace assertions from JUnit4 with equivalent functionality from AssertJ, to 
> make {{ContractTestUtils}} independent of JUnit version.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19026) S3A: TestIAMInstanceCredentialsProvider.testIAMInstanceCredentialsInstantiate failure

2024-01-05 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19026?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17803518#comment-17803518
 ] 

Steve Loughran commented on HADOOP-19026:
-


{code}
java.lang.AssertionError: Cause not a IOException
at 
org.apache.hadoop.fs.s3a.auth.TestIAMInstanceCredentialsProvider.__CLR4_4_1ey8o37g7b(TestIAMInstanceCredentialsProvider.java:100)
at 
org.apache.hadoop.fs.s3a.auth.TestIAMInstanceCredentialsProvider.testIAMInstanceCredentialsInstantiate(TestIAMInstanceCredentialsProvider.java:72)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:750)
Caused by: software.amazon.awssdk.core.exception.SdkClientException: Failed to 
load credentials from IMDS.
at 
software.amazon.awssdk.core.exception.SdkClientException$BuilderImpl.build(SdkClientException.java:111)
at 
software.amazon.awssdk.core.exception.SdkClientException.create(SdkClientException.java:47)
at 
software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider.refreshCredentials(InstanceProfileCredentialsProvider.java:159)
at 
software.amazon.awssdk.utils.cache.CachedSupplier.lambda$jitteredPrefetchValueSupplier$8(CachedSupplier.java:300)
at 
software.amazon.awssdk.utils.cache.NonBlocking.fetch(NonBlocking.java:141)
at 
software.amazon.awssdk.utils.cache.CachedSupplier.refreshCache(CachedSupplier.java:208)
at 
software.amazon.awssdk.utils.cache.CachedSupplier.get(CachedSupplier.java:135)
at 
software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider.resolveCredentials(InstanceProfileCredentialsProvider.java:141)
at 
org.apache.hadoop.fs.s3a.auth.IAMInstanceCredentialsProvider.getCredentials(IAMInstanceCredentialsProvider.java:135)
at 
org.apache.hadoop.fs.s3a.auth.IAMInstanceCredentialsProvider.resolveCredentials(IAMInstanceCredentialsProvider.java:98)
at 
org.apache.hadoop.fs.s3a.auth.TestIAMInstanceCredentialsProvider.__CLR4_4_1ey8o37g7b(TestIAMInstanceCredentialsProvider.java:75)
... 14 more
Caused by: software.amazon.awssdk.core.exception.SdkClientException: The 
requested metadata is not found
at http://169.254.169.254/latest/meta-data/iam/security-credentials/
at 
software.amazon.awssdk.core.exception.SdkClientException$BuilderImpl.build(SdkClientException.java:111)
at 
software.amazon.awssdk.regions.util.HttpResourcesUtils.readResource(HttpResourcesUtils.java:125)
at 
software.amazon.awssdk.regions.util.HttpResourcesUtils.readResource(HttpResourcesUtils.java:91)
at 
software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider.lambda$getSecurityCredentials$3(InstanceProfileCredentialsProvider.java:256)
at 
software.amazon.awssdk.utils.FunctionalUtils.lambda$safeSupplier$4(FunctionalUtils.java:108)
at 
software.amazon.awssdk.utils.FunctionalUtils.invokeSafely(FunctionalUtils.java:136)
at 
software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider.getSecurityCredentials(InstanceProfileCredentialsProvider.java:256)
at 
software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider.createEndpointProvider(InstanceProfileCredentialsProvider.java:204)
at 
software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider.refreshCredentials(InstanceProfileCredentialsProvider.java:150)
... 22 more

{code}


> S3A: TestIAMInstanceCredentialsProvider.testIAMInstanceCredentialsInstantiate 
> failure
> -
>
> Key: HADOOP-19026
> URL: https://issues.apache.org/jira/browse/HADOOP-19026
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3, test
>Affects Versions: 3.4.0
>  

[jira] [Created] (HADOOP-19026) S3A: TestIAMInstanceCredentialsProvider.testIAMInstanceCredentialsInstantiate failure

2024-01-05 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19026:
---

 Summary: S3A: 
TestIAMInstanceCredentialsProvider.testIAMInstanceCredentialsInstantiate failure
 Key: HADOOP-19026
 URL: https://issues.apache.org/jira/browse/HADOOP-19026
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3, test
Affects Versions: 3.4.0
Reporter: Steve Loughran


test failure in TestIAMInstanceCredentialsProvider; looks like the test is 
running in an EC2 VM whose IAM service isn't providing credentials -and the 
test isn't set up to ignore that.


{code}
Caused by: software.amazon.awssdk.core.exception.SdkClientException: The 
requested metadata is not found
at http://169.254.169.254/latest/meta-data/iam/security-credentials/
at 
software.amazon.awssdk.core.exception.SdkClientException$BuilderImpl.build(SdkClientException.java:111)
at 
software.amazon.awssdk.regions.util.HttpResourcesUtils.readResource(HttpResourcesUtils.java:125)
at 
software.amazon.awssdk.regions.util.HttpResourcesUtils.readResource(HttpResourcesUtils.java:91)
at 
software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider.lambda$getSecurityCredentials$3(InstanceProfileCredentialsProvider.java:256)
at 
software.amazon.awssdk.utils.FunctionalUtils.lambda$safeSupplier$4(FunctionalUtils.java:108)
at 
software.amazon.awssdk.utils.FunctionalUtils.invokeSafely(FunctionalUtils.java:136)
at 
software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider.getSecurityCredentials(InstanceProfileCredentialsProvider.java:256)
at 
software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider.createEndpointProvider(InstanceProfileCredentialsProvider.java:204)
at 
software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider.refreshCredentials(InstanceProfileCredentialsProvider.java:150)

{code}




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-17912) ABFS: Support for Encryption Context

2024-01-03 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-17912?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-17912:

Fix Version/s: 3.3.9

> ABFS: Support for Encryption Context
> 
>
> Key: HADOOP-17912
> URL: https://issues.apache.org/jira/browse/HADOOP-17912
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Affects Versions: 3.3.1
>Reporter: Sumangala Patki
>Assignee: Pranav Saxena
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0, 3.3.9
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> Support for customer-provided encryption keys at the file level, superceding 
> the global (account-level) key use in HADOOP-17536.
> ABFS driver will support an "EncryptionContext" plugin for retrieving 
> encryption information, the implementation for which should be provided by 
> the client. The keys/context retrieved will be sent via request headers to 
> the server, which will store the encryption context. Subsequent REST calls to 
> server that access data/user metadata of the file will require fetching the 
> encryption context through a GetFileProperties call and retrieving the key 
> from the custom provider, before sending the request.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-19021) [ABFS] move to jdk11 HttpClient for http2 and connection keep alive

2024-01-02 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19021?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19021:

Summary: [ABFS] move to jdk11 HttpClient for http2 and connection keep 
alive  (was: in hadoop-azure, use jdk11 HttpClient instead of legacy 
java.net.HttpURLConnection, for supporting http2 and connection keep alive)

> [ABFS] move to jdk11 HttpClient for http2 and connection keep alive
> ---
>
> Key: HADOOP-19021
> URL: https://issues.apache.org/jira/browse/HADOOP-19021
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 3.4.0
>Reporter: Arnaud Nauwynck
>Priority: Critical
>
> As described in Jira Title: "in hadoop-azure, use jdk11 HttpClient instead of 
> legacy java.net.HttpURLConnection, for supporting http2 and connection keep 
> alive"
> Few remarks:
> 1/ The official Azure SDK supports either OkHttp or Netty for the Http 
> transport.
> 2/ the actual hadoop-azure use the class java.net.HttpURLConnection, which is 
> slow.
>   It does not use Http2, does not optimize SSL Hand-shake very well, and does 
> not keep TCP connection alive for re-use.
> 3/ JDK since version >=11 have a new class HttpClient which should be a 
> better replacement 
> 4/ it might be possible to introduce a configuration property (with defaut to 
> use legacy class) , and an abstract factory to create connection via either 
> HttpURLConnection or any other pluggeable implementation (jdk 11 HttpClient, 
> OkHttp, Netty, ...)
> 5/ the official Azure SDK is maintained by Microsoft, so should better follow 
> bug fixes and improvements than custom hadoop implementation?
> [https://learn.microsoft.com/en-us/java/api/overview/azure/storage-file-datalake-readme?view=azure-java-stable
> |https://learn.microsoft.com/en-us/java/api/overview/azure/storage-file-datalake-readme?view=azure-java-stable]
> 6/ when we use code with the official Azure SDK and Hadoop(in Spark), it is 
> chocking to have 2 different implementations within the same JVM... 
> 7/ The official Azure SDK has more features that what allows the legacy 
> hadoop class FileSystem to do... In particular, we can append (=upload) file 
> by multiple threads (upload by fragments at different offsets), then flush 
> when every fragments are sent.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18991) Remove commons-beanutils dependency from Hadoop 3

2024-01-02 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17801756#comment-17801756
 ] 

Steve Loughran commented on HADOOP-18991:
-

so leave in branch-3.3 and noted as optional; cut from trunk?

> Remove commons-beanutils dependency from Hadoop 3
> -
>
> Key: HADOOP-18991
> URL: https://issues.apache.org/jira/browse/HADOOP-18991
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Istvan Toth
>Priority: Major
>
> Hadoop doesn't acually use it, and it pollutes the classpath of dependent 
> projects.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-19017) Setup pre-commit CI for Windows 10

2024-01-02 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19017?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-19017.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

> Setup pre-commit CI for Windows 10
> --
>
> Key: HADOOP-19017
> URL: https://issues.apache.org/jira/browse/HADOOP-19017
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: build
>Affects Versions: 3.4.0
> Environment: Windows 10
>Reporter: Gautham Banasandra
>Assignee: Gautham Banasandra
>Priority: Critical
>  Labels: Jenkins, pull-request-available
> Fix For: 3.4.0
>
>
> We need to setup a pre-commit CI for validating the Hadoop PRs against 
> Windows 10.
> On a sidenote, we've got the nightly Jenkins CI running for Hadoop on Windows 
> 10 - 
> https://ci-hadoop.apache.org/view/Hadoop/job/hadoop-qbt-trunk-java8-win10-x86_64/.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18991) Remove commons-beanutils dependency from Hadoop 3

2024-01-01 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18991?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17801599#comment-17801599
 ] 

Steve Loughran commented on HADOOP-18991:
-

hive should add explicit dependencies here

> Remove commons-beanutils dependency from Hadoop 3
> -
>
> Key: HADOOP-18991
> URL: https://issues.apache.org/jira/browse/HADOOP-18991
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Istvan Toth
>Priority: Major
>
> Hadoop doesn't acually use it, and it pollutes the classpath of dependent 
> projects.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-17347) ABFS: read/cache footer with fs.azure.footer.read.request.size

2024-01-01 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-17347?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-17347:

Summary: ABFS: read/cache footer with fs.azure.footer.read.request.size  
(was: ABFS: Optimise read for small files/tails of files)

> ABFS: read/cache footer with fs.azure.footer.read.request.size
> --
>
> Key: HADOOP-17347
> URL: https://issues.apache.org/jira/browse/HADOOP-17347
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Affects Versions: 3.4.0
>Reporter: Bilahari T H
>Assignee: Bilahari T H
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.3.1
>
>  Time Spent: 12h 50m
>  Remaining Estimate: 0h
>
> Optimize read performance for the following scenarios
>  # Read small files completely
>  Files that are of size smaller than the read buffer size can be considered 
> as small files. In case of such files it would be better to read the full 
> file into the AbfsInputStream buffer.
>  # Read last block if the read is for footer
>  If the read is for the last 8 bytes, read the full file.
>  This will optimize reads for parquet files. [Parquet file 
> format|https://www.ellicium.com/parquet-file-format-structure/]
> Both these optimizations will be present under configs as follows
>  # fs.azure.read.smallfilescompletely
>  # fs.azure.read.optimizefooterread



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-17912) ABFS: Support for Encryption Context

2024-01-01 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-17912?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-17912.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

> ABFS: Support for Encryption Context
> 
>
> Key: HADOOP-17912
> URL: https://issues.apache.org/jira/browse/HADOOP-17912
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Affects Versions: 3.3.1
>Reporter: Sumangala Patki
>Assignee: Pranav Saxena
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> Support for customer-provided encryption keys at the file level, superceding 
> the global (account-level) key use in HADOOP-17536.
> ABFS driver will support an "EncryptionContext" plugin for retrieving 
> encryption information, the implementation for which should be provided by 
> the client. The keys/context retrieved will be sent via request headers to 
> the server, which will store the encryption context. Subsequent REST calls to 
> server that access data/user metadata of the file will require fetching the 
> encryption context through a GetFileProperties call and retrieving the key 
> from the custom provider, before sending the request.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18540) Upgrade Bouncy Castle to 1.70

2024-01-01 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18540?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-18540.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

> Upgrade Bouncy Castle to 1.70
> -
>
> Key: HADOOP-18540
> URL: https://issues.apache.org/jira/browse/HADOOP-18540
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: build
>Affects Versions: 3.4.0
>Reporter: D M Murali Krishna Reddy
>Assignee: D M Murali Krishna Reddy
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> Upgrade Bouncycastle to 1.70 to resolve
>  
> |[[sonatype-2021-4916] CWE-327: Use of a Broken or Risky Cryptographic 
> Algorithm|https://ossindex.sonatype.org/vulnerability/sonatype-2021-4916?component-type=maven&component-name=org.bouncycastle/bcprov-jdk15on]|
> |[[sonatype-2019-0673] CWE-400: Uncontrolled Resource Consumption ('Resource 
> Exhaustion')|https://ossindex.sonatype.org/vulnerability/sonatype-2019-0673?component-type=maven&component-name=org.bouncycastle/bcprov-jdk15on]|



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19013) fs.getXattrs(path) for S3FS doesn't have x-amz-server-side-encryption-aws-kms-key-id header.

2024-01-01 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19013?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17801542#comment-17801542
 ] 

Steve Loughran commented on HADOOP-19013:
-

this a standard header? or do you just think it'd be useful?

Either way, I support it -just make sure COPY does the right thing and sets the 
new header value

> fs.getXattrs(path) for S3FS doesn't have 
> x-amz-server-side-encryption-aws-kms-key-id header.
> 
>
> Key: HADOOP-19013
> URL: https://issues.apache.org/jira/browse/HADOOP-19013
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.3.6
>Reporter: Mukund Thakur
>Assignee: Mukund Thakur
>Priority: Major
>
> Once a path while uploading has been encrypted with SSE-KMS with a key id and 
> then later when we try to read the attributes of the same file, it doesn't 
> contain the key id information as an attribute. should we add it?



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-19008) S3A: Upgrade AWS SDK to 2.21.41

2023-12-12 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19008?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-19008.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

> S3A: Upgrade AWS SDK to 2.21.41
> ---
>
> Key: HADOOP-19008
> URL: https://issues.apache.org/jira/browse/HADOOP-19008
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0, 3.3.7-aws
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> sdk 2.21.41 is out and logging now picks up the log4j.properties options. 
> move to this ASAP



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-19008) S3A: Upgrade AWS SDK to 2.21.41

2023-12-08 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19008:
---

 Summary: S3A: Upgrade AWS SDK to 2.21.41
 Key: HADOOP-19008
 URL: https://issues.apache.org/jira/browse/HADOOP-19008
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.4.0, 3.3.7-aws
Reporter: Steve Loughran
Assignee: Steve Loughran


sdk 2.21.41 is out and logging now picks up the log4j.properties options. 

move to this ASAP



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18999) S3A: debug logging for http traffic to S3 stores

2023-12-08 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18999?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-18999.
-
Resolution: Not A Problem

no longer needed as 2.21.41 logs properly!

> S3A: debug logging for http traffic to S3 stores
> 
>
> Key: HADOOP-18999
> URL: https://issues.apache.org/jira/browse/HADOOP-18999
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
>
> AWS SDK bundle.jar logging doesn't set up right.
> {code}
> SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
> SLF4J: Defaulting to no-operation (NOP) logger implementation
> SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further 
> details.
> {code}
> Cloudstore commands have a -debug option to force set this through log4j 
> APIs; this does work. 
> Proposed:
> * add reflection-based ability to set/query log4j log levels (+tests, 
> obviously)
> * add a new log `org.apache.hadoop.fs.s3a.logging.sdk`
> * if set to DEBUG, DefaultS3ClientFactory will enable logging on the aws 
> internal/shaded classes
> this allows log4j.properties to turn on logging; reflection ensures all is 
> well on other log back-ends and when unshaded aws sdk jars are in use



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19007) S3A: transfer manager not wired up to s3a executor pool

2023-12-08 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19007?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17794656#comment-17794656
 ] 

Steve Loughran commented on HADOOP-19007:
-

[~ahmar] any specific reason why this isn't done? and what happens without it?

> S3A: transfer manager not wired up to s3a executor pool
> ---
>
> Key: HADOOP-19007
> URL: https://issues.apache.org/jira/browse/HADOOP-19007
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Priority: Major
>
> S3ClientFactory.createS3TransferManager() doesn't use the executor declared 
> in S3ClientCreationParameters.transferManagerExecutor
> * method needs to take S3ClientCreationParameters
> * and set the transfer manager executor



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-19007) S3A: transfer manager not wired up to s3a executor pool

2023-12-08 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19007:
---

 Summary: S3A: transfer manager not wired up to s3a executor pool
 Key: HADOOP-19007
 URL: https://issues.apache.org/jira/browse/HADOOP-19007
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.4.0
Reporter: Steve Loughran


S3ClientFactory.createS3TransferManager() doesn't use the executor declared in 
S3ClientCreationParameters.transferManagerExecutor

* method needs to take S3ClientCreationParameters
* and set the transfer manager executor



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18915) Tune/extend S3A http connection and thread pool settings

2023-12-07 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17794302#comment-17794302
 ] 

Steve Loughran commented on HADOOP-18915:
-

looking at the AWS SDK source, it looks like the s3express createsession code 
doesn't pick up the expanded settings as the timeout is fixed at 10s. 
 
What does that mean? It means if you are working with S3Express buckets *you 
had better have a large http connection pool and no connectivity issues*

we may want to think about to test/handle this best in our own code. We should 
be retrying.

> Tune/extend S3A http connection and thread pool settings
> 
>
> Key: HADOOP-18915
> URL: https://issues.apache.org/jira/browse/HADOOP-18915
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0, 3.3.7-aws
>
>
> Increases existing pool sizes, as with server scale and vector
> IO, larger pools are needed
>   fs.s3a.connection.maximum 200
>   fs.s3a.threads.max 96
> Adds new configuration options for v2 sdk internal timeouts,
> both with default of 60s:
>   fs.s3a.connection.acquisition.timeout
>   fs.s3a.connection.idle.time
> All the pool/timoeut options are covered in performance.md
> Moves all timeout/duration options in the s3a FS to taking
> temporal units (h, m, s, ms,...); retaining the previous default
> unit (normally millisecond)
> Adds a minimum duration for most of these, in order to recover from
> deployments where a timeout has been set on the assumption the unit
> was seconds, not millis.
> Uses java.time.Duration throughout the codebase;
> retaining the older numeric constants in
> org.apache.hadoop.fs.s3a.Constants for backwards compatibility;
> these are now deprecated.
> Adds new class AWSApiCallTimeoutException to be raised on
> sdk-related methods and also gateway timeouts. This is a subclass
> of org.apache.hadoop.net.ConnectTimeoutException to support
> existing retry logic.
> + reverted default value of fs.s3a.create.performance to false; 
> inadvertently set to true during testing.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18540) Upgrade Bouncy Castle to 1.70

2023-12-07 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18540?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17794283#comment-17794283
 ] 

Steve Loughran commented on HADOOP-18540:
-

[~BilwaST] ok, lets get that PR rebased for a retest and then merged. in 27h go 
offline until 2024 so other people will have to merge

> Upgrade Bouncy Castle to 1.70
> -
>
> Key: HADOOP-18540
> URL: https://issues.apache.org/jira/browse/HADOOP-18540
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: build
>Affects Versions: 3.4.0
>Reporter: D M Murali Krishna Reddy
>Assignee: D M Murali Krishna Reddy
>Priority: Major
>  Labels: pull-request-available
>
> Upgrade Bouncycastle to 1.70 to resolve
>  
> |[[sonatype-2021-4916] CWE-327: Use of a Broken or Risky Cryptographic 
> Algorithm|https://ossindex.sonatype.org/vulnerability/sonatype-2021-4916?component-type=maven&component-name=org.bouncycastle/bcprov-jdk15on]|
> |[[sonatype-2019-0673] CWE-400: Uncontrolled Resource Consumption ('Resource 
> Exhaustion')|https://ossindex.sonatype.org/vulnerability/sonatype-2019-0673?component-type=maven&component-name=org.bouncycastle/bcprov-jdk15on]|



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18925) S3A: add option "fs.s3a.copy.from.local.enabled" to enable/disable CopyFromLocalOperation

2023-12-07 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18925?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-18925.
-
Fix Version/s: 3.4.0
   3.3.9
   Resolution: Fixed

> S3A: add option "fs.s3a.copy.from.local.enabled" to enable/disable 
> CopyFromLocalOperation
> -
>
> Key: HADOOP-18925
> URL: https://issues.apache.org/jira/browse/HADOOP-18925
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Affects Versions: 3.3.6
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0, 3.3.9
>
>
> reported failure of CopyFromLocalOperation.getFinalPath() during job 
> submission with s3a declared as cluster fs.
> add an emergency option to disable this optimised uploader and revert to the 
> superclass implementation



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18997) S3A: Add option fs.s3a.s3express.create.session to enable/disable CreateSession

2023-12-07 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18997?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-18997.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

> S3A: Add option fs.s3a.s3express.create.session to enable/disable 
> CreateSession
> ---
>
> Key: HADOOP-18997
> URL: https://issues.apache.org/jira/browse/HADOOP-18997
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> add a way to disable the need to use the createsession call, so as to allow 
> for
> * simplifying our role test runs
> * benchmarking the performance hit
> * troubleshooting IAM permissions
> this can also be disabled from the sysprop "aws.disableS3ExpressAuth"



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-19004) S3A: Support Authentication through HttpSigner API

2023-12-06 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19004?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19004:

Summary: S3A: Support Authentication through HttpSigner API   (was: S3A: 
Move to a new HttpSigner for S3Express)

> S3A: Support Authentication through HttpSigner API 
> ---
>
> Key: HADOOP-19004
> URL: https://issues.apache.org/jira/browse/HADOOP-19004
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Harshit Gupta
>Priority: Major
>  Labels: pull-request-available
>
> The latest AWS SDK changes how signing works, and for signing S3Express 
> signatures the new {{software.amazon.awssdk.http.auth}} auth mechanism is 
> needed



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-19004) S3A: Move to a new HttpSigner for S3Express

2023-12-05 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19004:
---

 Summary: S3A: Move to a new HttpSigner for S3Express
 Key: HADOOP-19004
 URL: https://issues.apache.org/jira/browse/HADOOP-19004
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.4.0
Reporter: Steve Loughran
Assignee: Harshit Gupta


The latest AWS SDK changes how signing works, and for signing S3Express 
signatures the new {{software.amazon.awssdk.http.auth}} auth mechanism is needed



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18955) AWS SDK v2: add path capability probe "fs.s3a.capability.aws.v2"

2023-12-05 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-18955.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

Fixed in HADOOP-18996

> AWS SDK v2: add path capability probe "fs.s3a.capability.aws.v2"
> 
>
> Key: HADOOP-18955
> URL: https://issues.apache.org/jira/browse/HADOOP-18955
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Minor
> Fix For: 3.4.0
>
>
> Add a "hasPathCapability()" probe for s3a v2 builds to aid diagnostics 
> -avoids needing to look for specific s3a files on classpath. 
> plus bucket-info to enum all capabilities which may be present



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18961) S3A: add s3guard command "bucket" to create buckets

2023-12-05 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18961?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-18961.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

Fixed in HADOOP-18996

> S3A: add s3guard command "bucket" to create buckets
> ---
>
> Key: HADOOP-18961
> URL: https://issues.apache.org/jira/browse/HADOOP-18961
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Fix For: 3.4.0
>
>
> cloudstore has an mkbucket command
> https://github.com/steveloughran/cloudstore/blob/main/src/main/site/mkbucket.md
> however, its v1 api, has problems with spans and needs rework for v2. Oh, and 
> it has not tests.
> * move the command into hadoop-aws
> * add a test or two, 
> * add a mapper of  409 to bad request, as "BucketAlreadyOwnedByYouException" 
> comes in as a 409.
> Testing will be tricky...as well as not actually wanting to create new 
> buckets, my test a/c doesn't even have the permission to do so.
> Proposed tests
> # invalid args must be rejected
> # trying to create the current bucket must be rejected. This happens even if 
> you lack the permission to create
> *No actual attempts to create a new bucket*



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18997) S3A: Add option fs.s3a.s3express.create.session to enable/disable CreateSession

2023-12-05 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18997?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18997:

Summary: S3A: Add option fs.s3a.s3express.create.session to enable/disable 
CreateSession  (was: S3A: add way to turn off CreateSession API)

> S3A: Add option fs.s3a.s3express.create.session to enable/disable 
> CreateSession
> ---
>
> Key: HADOOP-18997
> URL: https://issues.apache.org/jira/browse/HADOOP-18997
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Minor
>  Labels: pull-request-available
>
> add a way to disable the need to use the createsession call, so as to allow 
> for
> * simplifying our role test runs
> * benchmarking the performance hit
> * troubleshooting IAM permissions
> this can also be disabled from the sysprop "aws.disableS3ExpressAuth"



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-18708) AWS SDK V2 - Implement CSE

2023-12-05 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18708?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran reassigned HADOOP-18708:
---

Assignee: Ahmar Suhail

> AWS SDK V2 - Implement CSE
> --
>
> Key: HADOOP-18708
> URL: https://issues.apache.org/jira/browse/HADOOP-18708
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Ahmar Suhail
>Priority: Major
>  Labels: pull-request-available
>
> S3 Encryption client for SDK V2 is now available, so add client side 
> encryption back in. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19003) S3A Assume role tests failing against S3Express stores

2023-12-05 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19003?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17793310#comment-17793310
 ] 

Steve Loughran commented on HADOOP-19003:
-


{code}
[ERROR]   ITestAssumeRole.testPartialDelete:696->executePartialDelete:732 » 
AccessDenied
[ERROR]   
ITestAssumeRole.testPartialDeleteSingleDelete:702->executePartialDelete:732 » 
AccessDenied
[ERROR]   ITestAssumeRole.testReadOnlyOperations:463 » AccessDenied 
s3a://stevel--usw2-a...
[ERROR]   ITestAssumeRole.testRestrictedCommitActions:626 » AccessDenied 
s3a://stevel--u...
[ERROR]   ITestAssumeRole.testRestrictedWriteSubdir:513 » AccessDenied 
s3a://stevel--usw...
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testAbortNonexistentDir:234
 » AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testBaseRelativePath:282 
» AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testBulkCommitFiles:602 
» AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testCommitEmptyFile:217->ITestCommitOperations.createCommitAndVerify:346
 » AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testCommitSmallFile:223->ITestCommitOperations.createCommitAndVerify:346
 » AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testCreateAbortEmptyFile:172
 » AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testFailuresInAbortListing:566
 » AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testMarkerFileRename:304 
» AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testRevertCommit:544 » 
AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testRevertMissingCommit:556
 » AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testUploadEmptyFile:472 
» AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testUploadSmallFile:503 
» AccessDenied
[ERROR]   
ITestAssumedRoleCommitOperations>ITestCommitOperations.testWriteNormalStream:581
 » AccessDenied
[ERROR]   
ITestRestrictedReadAccess.testNoReadAccess:211->checkBasicFileOperations:283 » 
AccessDenied
[ERROR]   
ITestRoleDelegationInFilesystem>ITestSessionDelegationInFilesystem.testDelegatedFileSystem:406->ITestSessionDelegationInFilesystem.executeDelegatedFSOperations:454
 » AccessDenied
[ERROR]   ITestPartialRenamesDeletes.testCopyDirFailsToReadOnlyDir:519 » 
AccessDenied s3...
[ERROR]   ITestPartialRenamesDeletes.testCopyDirFailsToReadOnlyDir:519 » 
AccessDenied s3...
[ERROR]   ITestPartialRenamesDeletes.testPartialDirDelete:607 » AccessDenied 
s3a://steve...
[ERROR]   ITestPartialRenamesDeletes.testPartialDirDelete:607 » AccessDenied 
s3a://steve...
[ERROR]   ITestPartialRenamesDeletes.testPartialEmptyDirDelete:572 » 
AccessDenied s3a://...
[ERROR]   ITestPartialRenamesDeletes.testPartialEmptyDirDelete:572 » 
AccessDenied s3a://...
[ERROR]   ITestPartialRenamesDeletes.testRenameDirFailsInDelete:457 » 
AccessDenied s3a:/...
[ERROR]   ITestPartialRenamesDeletes.testRenameDirFailsInDelete:457 » 
AccessDenied s3a:/...
[ERROR]   ITestPartialRenamesDeletes.testRenameFileFailsNoWrite:503 » 
AccessDenied s3a:/...
[ERROR]   ITestPartialRenamesDeletes.testRenameFileFailsNoWrite:503 » 
AccessDenied s3a:/...
[ERROR]   ITestPartialRenamesDeletes.testRenameParentPathNotWriteable:393 » 
AccessDenied
[ERROR]   ITestPartialRenamesDeletes.testRenameParentPathNotWriteable:393 » 
AccessDenied
[ERROR]   ITestPartialRenamesDeletes.testRenamePermissionRequirements:756 » 
AccessDenied
[ERROR]   ITestPartialRenamesDeletes.testRenamePermissionRequirements:756 » 
AccessDenied
[ERROR]   ITestPartialRenamesDeletes.testRenameSingleFileFailsInDelete:416 » 
AccessDenied
[ERROR]   ITestPartialRenamesDeletes.testRenameSingleFileFailsInDelete:416 » 
AccessDenied
[INFO] 

{code}


> S3A Assume role tests failing against S3Express stores
> --
>
> Key: HADOOP-19003
> URL: https://issues.apache.org/jira/browse/HADOOP-19003
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3, test
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Priority: Minor
>
> The test suits which assume roles with restricted permissions down paths 
> still fail on S3Express, even after disabling createSession.
> This is with a role which *should* work.
> Either the role setup is wrong, or there's something special about role 
> configuration for S3Express buckets



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h

[jira] [Created] (HADOOP-19003) S3A Assume role tests failing against S3Express stores

2023-12-05 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19003:
---

 Summary: S3A Assume role tests failing against S3Express stores
 Key: HADOOP-19003
 URL: https://issues.apache.org/jira/browse/HADOOP-19003
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3, test
Affects Versions: 3.4.0
Reporter: Steve Loughran


The test suits which assume roles with restricted permissions down paths still 
fail on S3Express, even after disabling createSession.

This is with a role which *should* work.

Either the role setup is wrong, or there's something special about role 
configuration for S3Express buckets



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18997) S3A: add way to turn off CreateSession API

2023-12-05 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18997?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18997:

Description: 
add a way to disable the need to use the createsession call, so as to allow for

* simplifying our role test runs
* benchmarking the performance hit
* troubleshooting IAM permissions



this can also be disabled from the sysprop "aws.disableS3ExpressAuth"

  was:
add a way to disable the need to use the createsession call, so as to allow for

benchmarking the performance hit
troubleshooting IAM permissions

for our role tests we can turn this off


> S3A: add way to turn off CreateSession API
> --
>
> Key: HADOOP-18997
> URL: https://issues.apache.org/jira/browse/HADOOP-18997
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Minor
>  Labels: pull-request-available
>
> add a way to disable the need to use the createsession call, so as to allow 
> for
> * simplifying our role test runs
> * benchmarking the performance hit
> * troubleshooting IAM permissions
> this can also be disabled from the sysprop "aws.disableS3ExpressAuth"



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18968) S3A FileSystem does not correctly list empty directories after HADOOP-17199

2023-12-05 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18968?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18968:

Component/s: fs/s3

> S3A FileSystem does not correctly list empty directories after HADOOP-17199
> ---
>
> Key: HADOOP-18968
> URL: https://issues.apache.org/jira/browse/HADOOP-18968
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs/s3
>Affects Versions: 3.2.4
>Reporter: Max  Xie
>Priority: Major
> Attachments: image-2023-11-10-16-36-53-337.png
>
>
> After HADOOP-17199, s3a filesystem dose not  correctly ls empty directories.  
>  For example. 
> Before HADOOP-17199
> ```
> hdfs dfs -ls s3a://testbucket/dir/empty
>  
> // work right
> ```
>  
> After HADOOP-17199
> ```
> hdfs dfs -ls s3a://testbucket/dir/empty
> ls: `s3a://testbucket/dir/empty': No such file or directory
>  
> // wrong ouput
> ```
>  
> After Check the code, s3a empty folder don't have prefixes Or objects, it 
> whil throw FIleNotFouldException.
> !image-2023-11-10-16-36-53-337.png!  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19000) Unknown S3Express bucket raises UnknownHostException rather than NoSuchBucketException; will block for retries

2023-12-04 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19000?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17792835#comment-17792835
 ] 

Steve Loughran commented on HADOOP-19000:
-

no, not related to HADOOP-18915, though if ApiCallTimeoutException does start 
containing exceptions the placement there is wrong as we should look inside for 
any IOE. moving it down will mean that if it is fixed in the SDK then anyone 
upgrading their SDK will get the better handling.

> Unknown S3Express bucket raises UnknownHostException rather than 
> NoSuchBucketException; will block for retries
> --
>
> Key: HADOOP-19000
> URL: https://issues.apache.org/jira/browse/HADOOP-19000
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Priority: Minor
>
> When an attempt is made to work with an s3 express bucket which isn't there. 
> the createSession API fails with UnknownHostException
> This is actually retried within the sdk, so we are lucky that HADOOP-18889 
> cut the retry count down. Even so, failures are slow and not very informative.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-19001) LD_LIBRARY_PATH is missing HADOOP_COMMON_LIB_NATIVE_DIR

2023-12-02 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19001?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17792400#comment-17792400
 ] 

Steve Loughran commented on HADOOP-19001:
-

we've tried to lock this down for security, if you can find any explicit 
discussion on what decisions were made with the new scripts, that may show the 
reasoning. Or its just an accidental change. See what you can find out before 
we rush to change this. 

> LD_LIBRARY_PATH is missing HADOOP_COMMON_LIB_NATIVE_DIR
> ---
>
> Key: HADOOP-19001
> URL: https://issues.apache.org/jira/browse/HADOOP-19001
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.2.4
>Reporter: Zilong Zhu
>Priority: Major
>
> When we run a spark job, we find that it cannot load the native library 
> successfully.
> We found a difference between hadoop2 and hadoop3.
> hadoop2-Spark-System Properties:
> |java.library.path|:/hadoop/lib/native:/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib|
> hadoop3-Spark-System Properties:
> |java.library.path|:/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib|
> The key point is:
> hadoop2-hadoop-config.sh:
> HADOOP_OPTS="$HADOOP_OPTS -Djava.library.path=$JAVA_LIBRARY_PATH"   <--267
> export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$JAVA_LIBRARY_PATH     <--268
>  
> hadoop3-hadoop-functions.sh:
> hadoop_add_param HADOOP_OPTS java.library.path \
> "-Djava.library.path=${JAVA_LIBRARY_PATH}"
> export LD_LIBRARY_PATH        <--1484
>  
> At the same time, the hadoop3 will clear all non-whitelisted environment 
> variables.
> I'm not sure if it was intentional. But it makes our spark job unable to find 
> the native library on hadoop3. 
> Maybe we should modify hadoop-functions.sh(1484) and add LD_LIBRARY_PATH to 
> the default configuration item yarn.nodemanager.env-whitelist.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-19000) Unknown S3Express bucket raises UnknownHostException rather than NoSuchBucketException; will block for retries

2023-12-01 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-19000?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-19000:

Summary: Unknown S3Express bucket raises UnknownHostException rather than 
NoSuchBucketException; will block for retries  (was: Unknown S3Express bucket 
raises UnknownHostException rather than NoSuckBucketException)

> Unknown S3Express bucket raises UnknownHostException rather than 
> NoSuchBucketException; will block for retries
> --
>
> Key: HADOOP-19000
> URL: https://issues.apache.org/jira/browse/HADOOP-19000
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Priority: Minor
>
> When an attempt is made to work with an s3 express bucket which isn't there. 
> the createSession API fails with UnknownHostException
> This is actually retried within the sdk, so we are lucky that HADOOP-18889 
> cut the retry count down. Even so, failures are slow and not very informative.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-19000) Unknown S3Express bucket raises UnknownHostException rather than NoSuckBucketException

2023-12-01 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19000?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17792110#comment-17792110
 ] 

Steve Loughran edited comment on HADOOP-19000 at 12/1/23 3:34 PM:
--

replicate by trying to list a bucket which isn't present. note that my test 
setup has fs.s3a.attempts.maximum = 1, so I don't get to wait for failures. the 
default of 5 means a wait for about 5s.


{code}


> bin/hadoop fs -ls s3a://stevel2--usw2-az1--x-s3/

2023-12-01 15:15:52,385 
[s3a-transfer-stevel2--usw2-az1--x-s3-unbounded-pool2-t1] DEBUG s3a.Invoker 
(Invoker.java:retryUntranslated(474)) - List stevel2--usw2-az1--x-s3:/ 
delimiter=/ keys=5000 requester pays=null ; 
software.amazon.awssdk.core.exception.SdkClientException: Received an 
UnknownHostException when attempting to interact with a service. See cause for 
the exact endpoint that is failing to resolve. If this is happening on an 
endpoint that previously worked, there may be a network connectivity issue or 
your DNS cache could be storing endpoints for too long., 
2023-12-01 15:15:52,386 
[s3a-transfer-stevel2--usw2-az1--x-s3-unbounded-pool2-t1] DEBUG 
s3a.S3ARetryPolicy (S3ARetryPolicy.java:shouldRetry(305)) - Retry probe for 
UnknownHostException with 0 retries and 0 failovers, idempotent=true, due to 
java.net.UnknownHostException: 
software.amazon.awssdk.core.exception.SdkClientException: Received an 
UnknownHostException when attempting to interact with a service. See cause for 
the exact endpoint that is failing to resolve. If this is happening on an 
endpoint that previously worked, there may be a network connectivity issue or 
your DNS cache could be storing endpoints for too long.: 
stevel2--usw2-az1--x-s3.s3express-usw2-az1.us-east-2.amazonaws.com
java.net.UnknownHostException: 
software.amazon.awssdk.core.exception.SdkClientException: Received an 
UnknownHostException when attempting to interact with a service. See cause for 
the exact endpoint that is failing to resolve. If this is happening on an 
endpoint that previously worked, there may be a network connectivity issue or 
your DNS cache could be storing endpoints for too long.: 
stevel2--usw2-az1--x-s3.s3express-usw2-az1.us-east-2.amazonaws.com
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at 
org.apache.hadoop.fs.s3a.impl.ErrorTranslation.wrapWithInnerIOE(ErrorTranslation.java:132)
at 
org.apache.hadoop.fs.s3a.impl.ErrorTranslation.maybeExtractIOException(ErrorTranslation.java:105)
at 
org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:213)
at org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:481)
at org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:431)
at 
org.apache.hadoop.fs.s3a.S3AFileSystem.listObjects(S3AFileSystem.java:2954)
at 
org.apache.hadoop.fs.s3a.S3AFileSystem$ListingOperationCallbacksImpl.lambda$listObjectsAsync$0(S3AFileSystem.java:2573)
at 
org.apache.hadoop.fs.s3a.impl.CallableSupplier.get(CallableSupplier.java:88)
at 
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1604)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: software.amazon.awssdk.core.exception.SdkClientException: Received 
an UnknownHostException when attempting to interact with a service. See cause 
for the exact endpoint that is failing to resolve. If this is happening on an 
endpoint that previously worked, there may be a network connectivity issue or 
your DNS cache could be storing endpoints for too long.
at 
software.amazon.awssdk.core.exception.SdkClientException$BuilderImpl.build(SdkClientException.java:111)
at 
software.amazon.awssdk.awscore.interceptor.HelpfulUnknownHostExceptionInterceptor.modifyException(HelpfulUnknownHostExceptionInterceptor.java:59)
at 
software.amazon.awssdk.core.interceptor.ExecutionInterceptorChain.modifyException(ExecutionInterceptorChain.java:202)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.utils.ExceptionReportingUtils.runModifyException(ExceptionReportingUtils.java:54)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.utils.ExceptionReportingUtils.reportFailureToInterceptors(ExceptionReportingUtils.java:38)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.ExecutionFailureExceptionReportingStage.execute

[jira] [Commented] (HADOOP-19000) Unknown S3Express bucket raises UnknownHostException rather than NoSuckBucketException

2023-12-01 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19000?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17792112#comment-17792112
 ] 

Steve Loughran commented on HADOOP-19000:
-

now, set the retries to 10 and instead of any UHE being raised, an 
ApiCallTImeoutException is raised, which we then retry ourselves, *and the 
inner exception is lost*. 

API retry code should include last exception raised in its stack, so we can 
extract and react to. Or have I just broken this with HADOOP-18915 placement of 
retry mapping. will test that locally.

{code}
ed-b026-4dc8dfc37f68-0007 Executing op_list_status with 
{software.amazon.awssdk.services.s3.model.CreateSessionRequest size=0, 
mutating=true}; 
https://audit.example.org/hadoop/1/op_list_status/b8db02c1-bbfc-4bed-b026-4dc8dfc37f68-0007/?op=op_list_status&pr=stevel&ps=3d28ab00-f260-4fd4-bd72-f2209126ee96&cm=FsShell&id=b8db02c1-bbfc-4bed-b026-4dc8dfc37f68-0007&t0=1&fs=b8db02c1-bbfc-4bed-b026-4dc8dfc37f68&t1=16&ts=170151871
2023-12-01 15:27:53,025 
[s3a-transfer-stevel2--usw2-az1--x-s3-unbounded-pool2-t1] DEBUG s3a.Invoker 
(Invoker.java:retryUntranslated(474)) - List stevel2--usw2-az1--x-s3:/ 
delimiter=/ keys=5000 requester pays=null ; 
software.amazon.awssdk.core.exception.ApiCallTimeoutException: Client execution 
did not complete before the specified timeout configuration: 1 millis, 
2023-12-01 15:27:53,025 
[s3a-transfer-stevel2--usw2-az1--x-s3-unbounded-pool2-t1] DEBUG 
s3a.S3ARetryPolicy (S3ARetryPolicy.java:shouldRetry(305)) - Retry probe for 
AWSApiCallTimeoutException with 1 retries and 0 failovers, idempotent=true, due 
to org.apache.hadoop.fs.s3a.AWSApiCallTimeoutException: List 
stevel2--usw2-az1--x-s3:/ delimiter=/ keys=5000 requester pays=null: 
software.amazon.awssdk.core.exception.ApiCallTimeoutException: Client execution 
did not complete before the specified timeout configuration: 1 millis
org.apache.hadoop.fs.s3a.AWSApiCallTimeoutException: List 
stevel2--usw2-az1--x-s3:/ delimiter=/ keys=5000 requester pays=null: 
software.amazon.awssdk.core.exception.ApiCallTimeoutException: Client execution 
did not complete before the specified timeout configuration: 1 millis
at 
org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:189)
at org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:481)
at org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:431)
at 
org.apache.hadoop.fs.s3a.S3AFileSystem.listObjects(S3AFileSystem.java:2954)
at 
org.apache.hadoop.fs.s3a.S3AFileSystem$ListingOperationCallbacksImpl.lambda$listObjectsAsync$0(S3AFileSystem.java:2573)
at 
org.apache.hadoop.fs.s3a.impl.CallableSupplier.get(CallableSupplier.java:88)
at 
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1604)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: software.amazon.awssdk.core.exception.ApiCallTimeoutException: 
Client execution did not complete before the specified timeout configuration: 
1 millis
at 
software.amazon.awssdk.core.exception.ApiCallTimeoutException$BuilderImpl.build(ApiCallTimeoutException.java:97)
at 
software.amazon.awssdk.core.exception.ApiCallTimeoutException.create(ApiCallTimeoutException.java:38)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.generateApiCallTimeoutException(ApiCallTimeoutTrackingStage.java:151)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.handleInterruptedException(ApiCallTimeoutTrackingStage.java:139)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.translatePipelineException(ApiCallTimeoutTrackingStage.java:107)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.execute(ApiCallTimeoutTrackingStage.java:62)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallTimeoutTrackingStage.execute(ApiCallTimeoutTrackingStage.java:42)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallMetricCollectionStage.execute(ApiCallMetricCollectionStage.java:50)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.ApiCallMetricCollectionStage.execute(ApiCallMetricCollectionStage.java:32)
at 
software.amazon.awssdk.core.internal.http.pipeline.RequestPipelineBuilder$ComposingRequestPipelineStage.execute(RequestPipelineBuilder.java:206)
at 
software.amazon.awssdk.core.internal.http.pipeline.RequestPipelineBuilder$ComposingRequestPipelineStage.execute(RequestPipelineBuilder.java:206)
at 
software.amazon.awssdk.core.internal.

[jira] [Commented] (HADOOP-19000) Unknown S3Express bucket raises UnknownHostException rather than NoSuckBucketException

2023-12-01 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-19000?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17792110#comment-17792110
 ] 

Steve Loughran commented on HADOOP-19000:
-

replicate by trying to list a bucket which isn't present. note that my test 
setup has fs.s3a.attempts.maximum = 1, so I don't get to wait for failures. the 
default of 5 means a wait for about 5s.

> bin/hadoop fs -ls s3a://stevel2--usw2-az1--x-s3/

2023-12-01 15:15:52,385 
[s3a-transfer-stevel2--usw2-az1--x-s3-unbounded-pool2-t1] DEBUG s3a.Invoker 
(Invoker.java:retryUntranslated(474)) - List stevel2--usw2-az1--x-s3:/ 
delimiter=/ keys=5000 requester pays=null ; 
software.amazon.awssdk.core.exception.SdkClientException: Received an 
UnknownHostException when attempting to interact with a service. See cause for 
the exact endpoint that is failing to resolve. If this is happening on an 
endpoint that previously worked, there may be a network connectivity issue or 
your DNS cache could be storing endpoints for too long., 
2023-12-01 15:15:52,386 
[s3a-transfer-stevel2--usw2-az1--x-s3-unbounded-pool2-t1] DEBUG 
s3a.S3ARetryPolicy (S3ARetryPolicy.java:shouldRetry(305)) - Retry probe for 
UnknownHostException with 0 retries and 0 failovers, idempotent=true, due to 
java.net.UnknownHostException: 
software.amazon.awssdk.core.exception.SdkClientException: Received an 
UnknownHostException when attempting to interact with a service. See cause for 
the exact endpoint that is failing to resolve. If this is happening on an 
endpoint that previously worked, there may be a network connectivity issue or 
your DNS cache could be storing endpoints for too long.: 
stevel2--usw2-az1--x-s3.s3express-usw2-az1.us-east-2.amazonaws.com
java.net.UnknownHostException: 
software.amazon.awssdk.core.exception.SdkClientException: Received an 
UnknownHostException when attempting to interact with a service. See cause for 
the exact endpoint that is failing to resolve. If this is happening on an 
endpoint that previously worked, there may be a network connectivity issue or 
your DNS cache could be storing endpoints for too long.: 
stevel2--usw2-az1--x-s3.s3express-usw2-az1.us-east-2.amazonaws.com
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at 
org.apache.hadoop.fs.s3a.impl.ErrorTranslation.wrapWithInnerIOE(ErrorTranslation.java:132)
at 
org.apache.hadoop.fs.s3a.impl.ErrorTranslation.maybeExtractIOException(ErrorTranslation.java:105)
at 
org.apache.hadoop.fs.s3a.S3AUtils.translateException(S3AUtils.java:213)
at org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:481)
at org.apache.hadoop.fs.s3a.Invoker.retryUntranslated(Invoker.java:431)
at 
org.apache.hadoop.fs.s3a.S3AFileSystem.listObjects(S3AFileSystem.java:2954)
at 
org.apache.hadoop.fs.s3a.S3AFileSystem$ListingOperationCallbacksImpl.lambda$listObjectsAsync$0(S3AFileSystem.java:2573)
at 
org.apache.hadoop.fs.s3a.impl.CallableSupplier.get(CallableSupplier.java:88)
at 
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1604)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: software.amazon.awssdk.core.exception.SdkClientException: Received 
an UnknownHostException when attempting to interact with a service. See cause 
for the exact endpoint that is failing to resolve. If this is happening on an 
endpoint that previously worked, there may be a network connectivity issue or 
your DNS cache could be storing endpoints for too long.
at 
software.amazon.awssdk.core.exception.SdkClientException$BuilderImpl.build(SdkClientException.java:111)
at 
software.amazon.awssdk.awscore.interceptor.HelpfulUnknownHostExceptionInterceptor.modifyException(HelpfulUnknownHostExceptionInterceptor.java:59)
at 
software.amazon.awssdk.core.interceptor.ExecutionInterceptorChain.modifyException(ExecutionInterceptorChain.java:202)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.utils.ExceptionReportingUtils.runModifyException(ExceptionReportingUtils.java:54)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.utils.ExceptionReportingUtils.reportFailureToInterceptors(ExceptionReportingUtils.java:38)
at 
software.amazon.awssdk.core.internal.http.pipeline.stages.ExecutionFailureExceptionReportingStage.execute(ExecutionFailureExceptionReportingStage.java:39)
a

[jira] [Created] (HADOOP-19000) Unknown S3Express bucket raises UnknownHostException rather than NoSuckBucketException

2023-12-01 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-19000:
---

 Summary: Unknown S3Express bucket raises UnknownHostException 
rather than NoSuckBucketException
 Key: HADOOP-19000
 URL: https://issues.apache.org/jira/browse/HADOOP-19000
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.4.0
Reporter: Steve Loughran


When an attempt is made to work with an s3 express bucket which isn't there. 
the createSession API fails with UnknownHostException

This is actually retried within the sdk, so we are lucky that HADOOP-18889 cut 
the retry count down. Even so, failures are slow and not very informative.





--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18996) S3A to provide full support for S3 Express One Zone

2023-12-01 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18996:

Parent Issue: HADOOP-18886  (was: HADOOP-18477)

> S3A to provide full support for S3 Express One Zone
> ---
>
> Key: HADOOP-18996
> URL: https://issues.apache.org/jira/browse/HADOOP-18996
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> HADOOP-18995 upgrades the SDK version which allows connecting to a s3 express 
> one zone support. 
> Complete support needs to be added to address tests that fail with s3 express 
> one zone, additional tests, documentation etc. 
> * hadoop-common path capability to indicate that treewalking may encounter 
> missing dirs
> * use this in treewalking code in shell, mapreduce FileInputFormat etc to not 
> fail during treewalks
> * extra path capability for s3express too.
> * tests for this
> * anything else



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18996) S3A to provide full support for S3 Express One Zone

2023-12-01 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18996:

Release Note: Hadoop S3A connector has explicit awareness of and support 
for S3Express storage. Hadoop-common and hadoop-mapreduce treewalking code 
(shell commands, FileInputFormat,...) can cope with paths with incomplete 
uploads being visible. 

> S3A to provide full support for S3 Express One Zone
> ---
>
> Key: HADOOP-18996
> URL: https://issues.apache.org/jira/browse/HADOOP-18996
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> HADOOP-18995 upgrades the SDK version which allows connecting to a s3 express 
> one zone support. 
> Complete support needs to be added to address tests that fail with s3 express 
> one zone, additional tests, documentation etc. 
> * hadoop-common path capability to indicate that treewalking may encounter 
> missing dirs
> * use this in treewalking code in shell, mapreduce FileInputFormat etc to not 
> fail during treewalks
> * extra path capability for s3express too.
> * tests for this
> * anything else



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-18999) S3A: debug logging for http traffic to S3 stores

2023-12-01 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-18999:
---

 Summary: S3A: debug logging for http traffic to S3 stores
 Key: HADOOP-18999
 URL: https://issues.apache.org/jira/browse/HADOOP-18999
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.4.0
Reporter: Steve Loughran
Assignee: Steve Loughran


AWS SDK bundle.jar logging doesn't set up right.

{code}
SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
SLF4J: Defaulting to no-operation (NOP) logger implementation
SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further 
details.
{code}

Cloudstore commands have a -debug option to force set this through log4j APIs; 
this does work. 

Proposed:

* add reflection-based ability to set/query log4j log levels (+tests, obviously)
* add a new log `org.apache.hadoop.fs.s3a.logging.sdk`
* if set to DEBUG, DefaultS3ClientFactory will enable logging on the aws 
internal/shaded classes

this allows log4j.properties to turn on logging; reflection ensures all is well 
on other log back-ends and when unshaded aws sdk jars are in use




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18886) S3A: AWS SDK V2 Migration: stabilization and S3Express

2023-12-01 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18886?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18886:

Summary: S3A: AWS SDK V2 Migration: stabilization and S3Express  (was: S3A: 
AWS SDK V2 Migration: stabilization)

> S3A: AWS SDK V2 Migration: stabilization and S3Express
> --
>
> Key: HADOOP-18886
> URL: https://issues.apache.org/jira/browse/HADOOP-18886
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Ahmar Suhail
>Priority: Major
>
> The final stabilisation changes to the V2 SDK MIgration; those moved off the 
> HADOOP-18073 JIRA so we can close that.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-18997) S3A: add way to turn off CreateSession API

2023-12-01 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18997?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran reassigned HADOOP-18997:
---

Assignee: Steve Loughran

> S3A: add way to turn off CreateSession API
> --
>
> Key: HADOOP-18997
> URL: https://issues.apache.org/jira/browse/HADOOP-18997
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Minor
>  Labels: pull-request-available
>
> add a way to disable the need to use the createsession call, so as to allow 
> for
> benchmarking the performance hit
> troubleshooting IAM permissions
> for our role tests we can turn this off



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18996) S3A to provide full support for S3 Express One Zone

2023-12-01 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-18996.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

> S3A to provide full support for S3 Express One Zone
> ---
>
> Key: HADOOP-18996
> URL: https://issues.apache.org/jira/browse/HADOOP-18996
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> HADOOP-18995 upgrades the SDK version which allows connecting to a s3 express 
> one zone support. 
> Complete support needs to be added to address tests that fail with s3 express 
> one zone, additional tests, documentation etc. 
> * hadoop-common path capability to indicate that treewalking may encounter 
> missing dirs
> * use this in treewalking code in shell, mapreduce FileInputFormat etc to not 
> fail during treewalks
> * extra path capability for s3express too.
> * tests for this
> * anything else



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-18997) S3A: add way to turn off CreateSession API

2023-11-30 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-18997:
---

 Summary: S3A: add way to turn off CreateSession API
 Key: HADOOP-18997
 URL: https://issues.apache.org/jira/browse/HADOOP-18997
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.4.0
Reporter: Steve Loughran


add a way to disable the need to use the createsession call, so as to allow for

benchmarking the performance hit
troubleshooting IAM permissions

for our role tests we can turn this off



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18996) S3A to provide full support for S3 Express One Zone

2023-11-29 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17791267#comment-17791267
 ] 

Steve Loughran commented on HADOOP-18996:
-

PR is up; as I say in the PR, if people review it and for any change other than 
deleting stuff/disabling tests, I'd like the issues to be added to the v2 sdk 
stabilisation. Why so? people need to play with the new store before they can 
really review the changes

> S3A to provide full support for S3 Express One Zone
> ---
>
> Key: HADOOP-18996
> URL: https://issues.apache.org/jira/browse/HADOOP-18996
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
>
> HADOOP-18995 upgrades the SDK version which allows connecting to a s3 express 
> one zone support. 
> Complete support needs to be added to address tests that fail with s3 express 
> one zone, additional tests, documentation etc. 
> * hadoop-common path capability to indicate that treewalking may encounter 
> missing dirs
> * use this in treewalking code in shell, mapreduce FileInputFormat etc to not 
> fail during treewalks
> * extra path capability for s3express too.
> * tests for this
> * anything else



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18996) S3A to provide full support for S3 Express One Zone

2023-11-29 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18996:

Description: 
HADOOP-18995 upgrades the SDK version which allows connecting to a s3 express 
one zone support. 

Complete support needs to be added to address tests that fail with s3 express 
one zone, additional tests, documentation etc. 

* hadoop-common path capability to indicate that treewalking may encounter 
missing dirs
* use this in treewalking code in shell, mapreduce FileInputFormat etc to not 
fail during treewalks
* extra path capability for s3express too.
* tests for this
* anything else

  was:
HADOOP-18995 upgrades the SDK version which allows connecting to a s3 express 
one zone support. 

Complete support needs to be added to address tests that fail with s3 express 
one zone, additional tests, documentation etc. 


> S3A to provide full support for S3 Express One Zone
> ---
>
> Key: HADOOP-18996
> URL: https://issues.apache.org/jira/browse/HADOOP-18996
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>
> HADOOP-18995 upgrades the SDK version which allows connecting to a s3 express 
> one zone support. 
> Complete support needs to be added to address tests that fail with s3 express 
> one zone, additional tests, documentation etc. 
> * hadoop-common path capability to indicate that treewalking may encounter 
> missing dirs
> * use this in treewalking code in shell, mapreduce FileInputFormat etc to not 
> fail during treewalks
> * extra path capability for s3express too.
> * tests for this
> * anything else



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18996) S3A to provide full support for S3 Express One Zone

2023-11-29 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18996:

Summary: S3A to provide full support for S3 Express One Zone  (was: Add 
necessary software support for S3 Express One Zone)

> S3A to provide full support for S3 Express One Zone
> ---
>
> Key: HADOOP-18996
> URL: https://issues.apache.org/jira/browse/HADOOP-18996
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>
> HADOOP-18995 upgrades the SDK version which allows connecting to a s3 express 
> one zone support. 
> Complete support needs to be added to address tests that fail with s3 express 
> one zone, additional tests, documentation etc. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18915) Tune/extend S3A http connection and thread pool settings

2023-11-29 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18915?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-18915.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

> Tune/extend S3A http connection and thread pool settings
> 
>
> Key: HADOOP-18915
> URL: https://issues.apache.org/jira/browse/HADOOP-18915
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> Increases existing pool sizes, as with server scale and vector
> IO, larger pools are needed
>   fs.s3a.connection.maximum 200
>   fs.s3a.threads.max 96
> Adds new configuration options for v2 sdk internal timeouts,
> both with default of 60s:
>   fs.s3a.connection.acquisition.timeout
>   fs.s3a.connection.idle.time
> All the pool/timoeut options are covered in performance.md
> Moves all timeout/duration options in the s3a FS to taking
> temporal units (h, m, s, ms,...); retaining the previous default
> unit (normally millisecond)
> Adds a minimum duration for most of these, in order to recover from
> deployments where a timeout has been set on the assumption the unit
> was seconds, not millis.
> Uses java.time.Duration throughout the codebase;
> retaining the older numeric constants in
> org.apache.hadoop.fs.s3a.Constants for backwards compatibility;
> these are now deprecated.
> Adds new class AWSApiCallTimeoutException to be raised on
> sdk-related methods and also gateway timeouts. This is a subclass
> of org.apache.hadoop.net.ConnectTimeoutException to support
> existing retry logic.
> + reverted default value of fs.s3a.create.performance to false; 
> inadvertently set to true during testing.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18915) Tune/extend S3A http connection and thread pool settings

2023-11-29 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18915?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18915:

Summary: Tune/extend S3A http connection and thread pool settings  (was: 
Extend S3A http client connection timeouts)

> Tune/extend S3A http connection and thread pool settings
> 
>
> Key: HADOOP-18915
> URL: https://issues.apache.org/jira/browse/HADOOP-18915
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
>
> Increases existing pool sizes, as with server scale and vector
> IO, larger pools are needed
>   fs.s3a.connection.maximum 200
>   fs.s3a.threads.max 96
> Adds new configuration options for v2 sdk internal timeouts,
> both with default of 60s:
>   fs.s3a.connection.acquisition.timeout
>   fs.s3a.connection.idle.time
> All the pool/timoeut options are covered in performance.md
> Moves all timeout/duration options in the s3a FS to taking
> temporal units (h, m, s, ms,...); retaining the previous default
> unit (normally millisecond)
> Adds a minimum duration for most of these, in order to recover from
> deployments where a timeout has been set on the assumption the unit
> was seconds, not millis.
> Uses java.time.Duration throughout the codebase;
> retaining the older numeric constants in
> org.apache.hadoop.fs.s3a.Constants for backwards compatibility;
> these are now deprecated.
> Adds new class AWSApiCallTimeoutException to be raised on
> sdk-related methods and also gateway timeouts. This is a subclass
> of org.apache.hadoop.net.ConnectTimeoutException to support
> existing retry logic.
> + reverted default value of fs.s3a.create.performance to false; 
> inadvertently set to true during testing.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18915) Extend S3A http client connection timeouts

2023-11-29 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18915?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18915:

Description: 


Increases existing pool sizes, as with server scale and vector
IO, larger pools are needed

  fs.s3a.connection.maximum 200
  fs.s3a.threads.max 96

Adds new configuration options for v2 sdk internal timeouts,
both with default of 60s:

  fs.s3a.connection.acquisition.timeout
  fs.s3a.connection.idle.time

All the pool/timoeut options are covered in performance.md

Moves all timeout/duration options in the s3a FS to taking
temporal units (h, m, s, ms,...); retaining the previous default
unit (normally millisecond)

Adds a minimum duration for most of these, in order to recover from
deployments where a timeout has been set on the assumption the unit
was seconds, not millis.

Uses java.time.Duration throughout the codebase;
retaining the older numeric constants in
org.apache.hadoop.fs.s3a.Constants for backwards compatibility;
these are now deprecated.

Adds new class AWSApiCallTimeoutException to be raised on
sdk-related methods and also gateway timeouts. This is a subclass
of org.apache.hadoop.net.ConnectTimeoutException to support
existing retry logic.

+ reverted default value of fs.s3a.create.performance to false; 
inadvertently set to true during testing.


  was:
* Add ability to configure *all* timeouts, especially acquisition time
* recognise ApiCallTimeout and map tp a retryable exception
* use getDuration so suffixes can be used -so remove all ambiguity about time 
unit
* use units in core-default.xml so warnings aren't printed


> Extend S3A http client connection timeouts
> --
>
> Key: HADOOP-18915
> URL: https://issues.apache.org/jira/browse/HADOOP-18915
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
>
> Increases existing pool sizes, as with server scale and vector
> IO, larger pools are needed
>   fs.s3a.connection.maximum 200
>   fs.s3a.threads.max 96
> Adds new configuration options for v2 sdk internal timeouts,
> both with default of 60s:
>   fs.s3a.connection.acquisition.timeout
>   fs.s3a.connection.idle.time
> All the pool/timoeut options are covered in performance.md
> Moves all timeout/duration options in the s3a FS to taking
> temporal units (h, m, s, ms,...); retaining the previous default
> unit (normally millisecond)
> Adds a minimum duration for most of these, in order to recover from
> deployments where a timeout has been set on the assumption the unit
> was seconds, not millis.
> Uses java.time.Duration throughout the codebase;
> retaining the older numeric constants in
> org.apache.hadoop.fs.s3a.Constants for backwards compatibility;
> these are now deprecated.
> Adds new class AWSApiCallTimeoutException to be raised on
> sdk-related methods and also gateway timeouts. This is a subclass
> of org.apache.hadoop.net.ConnectTimeoutException to support
> existing retry logic.
> + reverted default value of fs.s3a.create.performance to false; 
> inadvertently set to true during testing.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Resolved] (HADOOP-18995) S3A: Upgrade AWS SDK version to 2.21.33 for Amazon S3 Express One Zone support

2023-11-29 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18995?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran resolved HADOOP-18995.
-
Fix Version/s: 3.4.0
   Resolution: Fixed

> S3A: Upgrade AWS SDK version to 2.21.33 for Amazon S3 Express One Zone support
> --
>
> Key: HADOOP-18995
> URL: https://issues.apache.org/jira/browse/HADOOP-18995
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Ahmar Suhail
>Priority: Major
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>
> Upgrade SDK version to 2.21.33, which adds S3 Express One Zone support.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18995) S3A: Upgrade AWS SDK version to 2.21.33 for Amazon S3 Express One Zone support

2023-11-29 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18995?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18995:

Summary: S3A: Upgrade AWS SDK version to 2.21.33 for Amazon S3 Express One 
Zone support  (was: Add support for Amazon S3 Express One Zone Storage - SDK 
version upgrade)

> S3A: Upgrade AWS SDK version to 2.21.33 for Amazon S3 Express One Zone support
> --
>
> Key: HADOOP-18995
> URL: https://issues.apache.org/jira/browse/HADOOP-18995
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Ahmar Suhail
>Priority: Major
>  Labels: pull-request-available
>
> Upgrade SDK version to 2.21.33, which adds S3 Express One Zone support.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-18996) Add necessary software support for S3 Express One Zone

2023-11-29 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-18996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17791037#comment-17791037
 ] 

Steve Loughran commented on HADOOP-18996:
-

i can probably help here

> Add necessary software support for S3 Express One Zone
> --
>
> Key: HADOOP-18996
> URL: https://issues.apache.org/jira/browse/HADOOP-18996
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>
> HADOOP-18995 upgrades the SDK version which allows connecting to a s3 express 
> one zone support. 
> Complete support needs to be added to address tests that fail with s3 express 
> one zone, additional tests, documentation etc. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-18996) Add necessary software support for S3 Express One Zone

2023-11-29 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran reassigned HADOOP-18996:
---

Assignee: Steve Loughran

> Add necessary software support for S3 Express One Zone
> --
>
> Key: HADOOP-18996
> URL: https://issues.apache.org/jira/browse/HADOOP-18996
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>
> HADOOP-18995 upgrades the SDK version which allows connecting to a s3 express 
> one zone support. 
> Complete support needs to be added to address tests that fail with s3 express 
> one zone, additional tests, documentation etc. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18994) Support OpenSSL 3

2023-11-28 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18994?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18994:

Summary: Support OpenSSL 3  (was: Support OpneSSL 3)

> Support OpenSSL 3
> -
>
> Key: HADOOP-18994
> URL: https://issues.apache.org/jira/browse/HADOOP-18994
> Project: Hadoop Common
>  Issue Type: Task
>Affects Versions: 3.3.6
>Reporter: Thippana Vamsi Kalyan
>Priority: Major
>
> Hadoop uses native OpenSSL library provided by linux distribution.
> OpenSSL 1.1.1 is deprecated: 
> https://www.openssl.org/blog/blog/2023/03/28/1.1.1-EOL/ 
> We need to ensure that Hadoop works with OpenSSL 3.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18991) Remove commons-benautils dependency from Hadoop 3

2023-11-28 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18991?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18991:

Issue Type: Improvement  (was: Bug)

> Remove commons-benautils dependency from Hadoop 3
> -
>
> Key: HADOOP-18991
> URL: https://issues.apache.org/jira/browse/HADOOP-18991
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Reporter: Istvan Toth
>Priority: Major
>
> Hadoop doesn't acually use it, and it pollutes the classpath of dependent 
> projects.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-18990) S3A: retry on credential expiry

2023-11-25 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-18990:
---

 Summary: S3A: retry on credential expiry
 Key: HADOOP-18990
 URL: https://issues.apache.org/jira/browse/HADOOP-18990
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.4.0
Reporter: Steve Loughran


Reported in AWS SDK https://github.com/aws/aws-sdk-java-v2/issues/3408

bq. In RetryableStage execute method, the "AwsCredentails" does not attempt to 
renew if it has expired. Therefore, if a method called with the existing 
credential is expiring soon, the number of retry is less than intended due to 
the expiration of the credential.

The stack from this report doesn't show any error detail we can use to identify 
the 400 exception as something we should be retrying on. This could be due to 
the logging, or it could actually hold. we've have to generate some socket 
credentials, let them expire and then see how hadoop fs commands failed. 
Something to do by hand as an STS test to do this is probably slow. *unless we 
expire all session credentials of a given role?*. Could be good, would be 
traumatic for other test runs though.

{code}
software.amazon.awssdk.services.s3.model.S3Exception: The provided token has 
expired. (Service: S3, Status Code: 400, Request ID: 3YWKVBNJPNTXPJX2, Extended 
Request ID: 
GkR56xA0r/Ek7zqQdB2ZdP3wqMMhf49HH7hc5N2TAIu47J3HEk6yvSgVNbX7ADuHDy/Irhr2rPQ=)

{code}




--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18915) Extend S3A http client connection timeouts

2023-11-25 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18915?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18915:

Description: 
* Add ability to configure *all* timeouts, especially acquisition time
* recognise ApiCallTimeout and map tp a retryable exception
* use getDuration so suffixes can be used -so remove all ambiguity about time 
unit
* use units in core-default.xml so warnings aren't printed

  was:
In the client config builders, when [setting 
timeouts|https://github.com/apache/hadoop/blob/trunk/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/impl/AWSClientConfig.java#L120],
 it uses Duration.ofSeconds(), configs all use milliseconds so this needs to be 
updated to Duration.ofMillis().

 


> Extend S3A http client connection timeouts
> --
>
> Key: HADOOP-18915
> URL: https://issues.apache.org/jira/browse/HADOOP-18915
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
>
> * Add ability to configure *all* timeouts, especially acquisition time
> * recognise ApiCallTimeout and map tp a retryable exception
> * use getDuration so suffixes can be used -so remove all ambiguity about time 
> unit
> * use units in core-default.xml so warnings aren't printed



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-18915) Extend S3A http client connection timeouts

2023-11-25 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-18915?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-18915:

Summary: Extend S3A http client connection timeouts  (was: S3A client 
builder to use getTImeDuration() for HTTP timeouts)

> Extend S3A http client connection timeouts
> --
>
> Key: HADOOP-18915
> URL: https://issues.apache.org/jira/browse/HADOOP-18915
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.4.0
>Reporter: Ahmar Suhail
>Assignee: Steve Loughran
>Priority: Major
>  Labels: pull-request-available
>
> In the client config builders, when [setting 
> timeouts|https://github.com/apache/hadoop/blob/trunk/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/impl/AWSClientConfig.java#L120],
>  it uses Duration.ofSeconds(), configs all use milliseconds so this needs to 
> be updated to Duration.ofMillis().
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



<    2   3   4   5   6   7   8   9   10   11   >