VitoMakarevich commented on code in PR #10460:
URL: https://github.com/apache/hudi/pull/10460#discussion_r1524878795
##########
hudi-aws/src/main/java/org/apache/hudi/config/GlueCatalogSyncClientConfig.java:
##########
@@ -40,6 +42,28 @@ public class GlueCatalogSyncClientConfig extends
HoodieConfig {
.sinceVersion("0.14.0")
.withDocumentation("Glue catalog sync based client will skip archiving
the table version if this config is set to true");
+ public static final ConfigProperty<Integer> ALL_PARTITIONS_READ_PARALLELISM
= ConfigProperty
+ .key(GLUE_CLIENT_PROPERTY_PREFIX + "all_partitions_read_parallelism")
+ .defaultValue(1)
+ .markAdvanced()
+ .withValidValues(IntStream.rangeClosed(1,
10).mapToObj(Integer::toString).toArray(String[]::new))
+ .sinceVersion("1.0.0")
+ .withDocumentation("Parallelism for listing all partitions(first time
sync). Should be in interval [1, 10].");
+
+ public static final ConfigProperty<Integer>
CHANGED_PARTITIONS_READ_PARALLELISM = ConfigProperty
+ .key(GLUE_CLIENT_PROPERTY_PREFIX + "changed_partitions_read_parallelism")
+ .defaultValue(1)
+ .markAdvanced()
+ .sinceVersion("1.0.0")
+ .withDocumentation("Parallelism for listing changed partitions(second
and subsequent syncs).");
+
+ public static final ConfigProperty<Integer> CHANGE_PARALLELISM =
ConfigProperty
+ .key(GLUE_CLIENT_PROPERTY_PREFIX + "change_parallelism")
+ .defaultValue(1)
+ .markAdvanced()
+ .sinceVersion("1.0.0")
+ .withDocumentation("Parallelism for change operations - such as
create/update/delete.");
Review Comment:
There is no range as such - only rationale limits that users see, to prevent
throttling exceptions. I made it 1 since expect only some users to face this
problem - like us. I can make it higher, anyway, it includes code for cap -
`number of threads = min(partitions/1000, user_specified_parallelism)`
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]