danielcweeks commented on code in PR #11052:
URL: https://github.com/apache/iceberg/pull/11052#discussion_r1777363011
##########
aws/src/main/java/org/apache/iceberg/aws/s3/S3FileIOProperties.java:
##########
@@ -393,6 +403,21 @@ public class S3FileIOProperties implements Serializable {
*/
private static final String S3_FILE_IO_USER_AGENT = "s3fileio/" +
EnvironmentContext.get();
+ /** Number of times to retry S3 operations. */
+ public static final String S3_RETRY_NUM_RETRIES = "s3.retry.num-retries";
+
+ public static final int S3_RETRY_NUM_RETRIES_DEFAULT = 32;
Review Comment:
I would agree with the concern here. If we can narrow this retry behavior
to just the 503 error code, I would be more amenable (though 32 is still very
high), but as a default retry this causes really bad behaviors for other
errors. The other issue is that there are retries in the surrounding execution
path (task retries and stage retries in spark for example). The compounding
effect of high retry values has a multiplicative effect for the overall job.
I would also say that trying to hide this is problematic for those
diagnosing slowness with their workloads, so we should log messages when
slowdowns are occurring.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]