This is an automated email from the ASF dual-hosted git repository. yihua pushed a commit to branch branch-0.x in repository https://gitbox.apache.org/repos/asf/hudi.git
commit 09d311360475bf7498f4a675743c23d79342022b Author: Y Ethan Guo <[email protected]> AuthorDate: Sun Feb 25 10:57:39 2024 -0800 [MINOR] Fix typos in hudi-common module (#10748) --- hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java | 2 +- hudi-common/src/main/java/org/apache/hudi/common/bloom/Key.java | 2 +- .../org/apache/hudi/common/util/queue/BaseHoodieQueueBasedExecutor.java | 2 +- .../src/main/java/org/apache/hudi/metadata/HoodieTableMetadataUtil.java | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) diff --git a/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java b/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java index 3c5486c47c7..5ec466cca3d 100644 --- a/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java +++ b/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java @@ -57,7 +57,7 @@ public class AvroSchemaUtils { /** * Establishes whether {@code newSchema} is compatible w/ {@code prevSchema}, as * defined by Avro's {@link AvroSchemaCompatibility}. - * From avro's compatability standpoint, prevSchema is writer schema and new schema is reader schema. + * From avro's compatibility standpoint, prevSchema is writer schema and new schema is reader schema. * {@code newSchema} is considered compatible to {@code prevSchema}, iff data written using {@code prevSchema} * could be read by {@code newSchema} * diff --git a/hudi-common/src/main/java/org/apache/hudi/common/bloom/Key.java b/hudi-common/src/main/java/org/apache/hudi/common/bloom/Key.java index 37ae6e68f73..f14d301ae3b 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/bloom/Key.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/bloom/Key.java @@ -136,7 +136,7 @@ public class Key implements Comparable<Key> { /** * Serialize the fields of this object to <code>out</code>. * - * @param out <code>DataOuput</code> to serialize this object into. + * @param out <code>DataOutput</code> to serialize this object into. * @throws IOException */ public void write(DataOutput out) throws IOException { diff --git a/hudi-common/src/main/java/org/apache/hudi/common/util/queue/BaseHoodieQueueBasedExecutor.java b/hudi-common/src/main/java/org/apache/hudi/common/util/queue/BaseHoodieQueueBasedExecutor.java index 20b9c802f60..f2843c56b03 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/util/queue/BaseHoodieQueueBasedExecutor.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/util/queue/BaseHoodieQueueBasedExecutor.java @@ -215,7 +215,7 @@ public abstract class BaseHoodieQueueBasedExecutor<I, O, E> implements HoodieExe // to be interrupted as well Thread.currentThread().interrupt(); } - // throw if we have any other exception seen already. There is a chance that cancellation/closing of producers with CompeletableFuture wins before the actual exception + // throw if we have any other exception seen already. There is a chance that cancellation/closing of producers with CompletableFuture wins before the actual exception // is thrown. if (this.queue.getThrowable() != null) { throw new HoodieException(queue.getThrowable()); diff --git a/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieTableMetadataUtil.java b/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieTableMetadataUtil.java index 7167a785f9f..0aa11042ab9 100644 --- a/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieTableMetadataUtil.java +++ b/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieTableMetadataUtil.java @@ -1615,7 +1615,7 @@ public class HoodieTableMetadataUtil { * * @param partitionType Type of the partition for which the file group count is to be estimated. * @param recordCount The number of records expected to be written. - * @param averageRecordSize Average size of each record to be writen. + * @param averageRecordSize Average size of each record to be written. * @param minFileGroupCount Minimum number of file groups to use. * @param maxFileGroupCount Maximum number of file groups to use. * @param growthFactor By what factor are the records (recordCount) expected to grow?
