This is an automated email from the ASF dual-hosted git repository. zhangyue19921010 pushed a commit to branch HUDI-8990 in repository https://gitbox.apache.org/repos/asf/hudi.git
commit 60de489d9a925d4f681894fc310081ffc3f27555 Author: YueZhang <[email protected]> AuthorDate: Mon Mar 17 20:04:40 2025 +0800 finish hashing_config initial && related UT --- .../java/org/apache/hudi/config/HoodieIndexConfig.java | 6 +++--- .../apache/hudi/index/bucket/SimpleBucketIndexUtils.java | 2 +- .../common/model/PartitionBucketIndexHashingConfig.java | 2 -- .../org/apache/hudi/table/catalog/TestHoodieCatalog.java | 14 +++++++------- 4 files changed, 11 insertions(+), 13 deletions(-) diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/config/HoodieIndexConfig.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/config/HoodieIndexConfig.java index 626e894fd1a..bfad563a9e7 100644 --- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/config/HoodieIndexConfig.java +++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/config/HoodieIndexConfig.java @@ -297,9 +297,9 @@ public class HoodieIndexConfig extends HoodieConfig { .key("hoodie.bucket.index.partition.expressions") .noDefaultValue() .markAdvanced() - .withDocumentation("Users can use this parameter to specify expression and the corresponding bucket " + - "numbers (separated by commas).Multiple rules are separated by semicolons like " + - "hoodie.bucket.index.partition.expressions=expression1,bucket-number1;expression2,bucket-number2"); + .withDocumentation("Users can use this parameter to specify expression and the corresponding bucket " + + "numbers (separated by commas).Multiple rules are separated by semicolons like " + + "hoodie.bucket.index.partition.expressions=expression1,bucket-number1;expression2,bucket-number2"); public static final ConfigProperty<String> BUCKET_INDEX_MAX_NUM_BUCKETS = ConfigProperty .key("hoodie.bucket.index.max.num.buckets") diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/bucket/SimpleBucketIndexUtils.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/bucket/SimpleBucketIndexUtils.java index 74c8022c9aa..f3a20c6a002 100644 --- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/bucket/SimpleBucketIndexUtils.java +++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/bucket/SimpleBucketIndexUtils.java @@ -46,7 +46,7 @@ public class SimpleBucketIndexUtils { String expressions, String rule, int defaultBucketNumber, - String instant){ + String instant) { if (StringUtils.isNullOrEmpty(expressions)) { return false; } diff --git a/hudi-common/src/main/java/org/apache/hudi/common/model/PartitionBucketIndexHashingConfig.java b/hudi-common/src/main/java/org/apache/hudi/common/model/PartitionBucketIndexHashingConfig.java index 225629b52ff..46b38896925 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/model/PartitionBucketIndexHashingConfig.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/model/PartitionBucketIndexHashingConfig.java @@ -38,7 +38,6 @@ public class PartitionBucketIndexHashingConfig implements Serializable { private final String expressions; private final int defaultBucketNumber; private final String rule; - private final int version; private final String instant; @@ -71,7 +70,6 @@ public class PartitionBucketIndexHashingConfig implements Serializable { return JsonUtils.getObjectMapper().readValue(jsonStr, clazz); } - public static PartitionBucketIndexHashingConfig fromBytes(byte[] bytes) throws IOException { try { return fromJsonString(new String(bytes, StandardCharsets.UTF_8), PartitionBucketIndexHashingConfig.class); diff --git a/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/table/catalog/TestHoodieCatalog.java b/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/table/catalog/TestHoodieCatalog.java index 1e518c5ef45..84e982df8f3 100644 --- a/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/table/catalog/TestHoodieCatalog.java +++ b/hudi-flink-datasource/hudi-flink/src/test/java/org/apache/hudi/table/catalog/TestHoodieCatalog.java @@ -360,14 +360,14 @@ public class TestHoodieCatalog { .createMetaClient( new HadoopStorageConfiguration(HadoopConfigurations.getHadoopConf(flinkConf)), tablePathStr); HoodieStorage storage = metaClient.getStorage(); - StoragePath initialHashing_config = + StoragePath initialHashingConfig = new StoragePath(metaClient.getHashingMetadataConfigPath(), SimpleBucketIndexUtils.INITIAL_HASHING_CONFIG_INSTANT + PartitionBucketIndexHashingConfig.HASHING_CONFIG_FILE_SUFFIX); - StoragePathInfo info = storage.getPathInfo(initialHashing_config); - Option<PartitionBucketIndexHashingConfig> hashing_config = SimpleBucketIndexUtils.loadHashingConfig(storage, info); - assertTrue(hashing_config.isPresent()); - assertEquals(hashing_config.get().getDefaultBucketNumber(), Integer.parseInt(defaultBucketNumber)); - assertEquals(hashing_config.get().getRule(), rule); - assertEquals(hashing_config.get().getExpressions(), expressions); + StoragePathInfo info = storage.getPathInfo(initialHashingConfig); + Option<PartitionBucketIndexHashingConfig> hashingConfig = SimpleBucketIndexUtils.loadHashingConfig(storage, info); + assertTrue(hashingConfig.isPresent()); + assertEquals(hashingConfig.get().getDefaultBucketNumber(), Integer.parseInt(defaultBucketNumber)); + assertEquals(hashingConfig.get().getRule(), rule); + assertEquals(hashingConfig.get().getExpressions(), expressions); } @Test
