This is an automated email from the ASF dual-hosted git repository.

yihua pushed a commit to branch branch-0.x
in repository https://gitbox.apache.org/repos/asf/hudi.git

commit 09d311360475bf7498f4a675743c23d79342022b
Author: Y Ethan Guo <[email protected]>
AuthorDate: Sun Feb 25 10:57:39 2024 -0800

    [MINOR] Fix typos in hudi-common module (#10748)
---
 hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java     | 2 +-
 hudi-common/src/main/java/org/apache/hudi/common/bloom/Key.java         | 2 +-
 .../org/apache/hudi/common/util/queue/BaseHoodieQueueBasedExecutor.java | 2 +-
 .../src/main/java/org/apache/hudi/metadata/HoodieTableMetadataUtil.java | 2 +-
 4 files changed, 4 insertions(+), 4 deletions(-)

diff --git 
a/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java 
b/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java
index 3c5486c47c7..5ec466cca3d 100644
--- a/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java
+++ b/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java
@@ -57,7 +57,7 @@ public class AvroSchemaUtils {
   /**
    * Establishes whether {@code newSchema} is compatible w/ {@code 
prevSchema}, as
    * defined by Avro's {@link AvroSchemaCompatibility}.
-   * From avro's compatability standpoint, prevSchema is writer schema and new 
schema is reader schema.
+   * From avro's compatibility standpoint, prevSchema is writer schema and new 
schema is reader schema.
    * {@code newSchema} is considered compatible to {@code prevSchema}, iff 
data written using {@code prevSchema}
    * could be read by {@code newSchema}
    *
diff --git a/hudi-common/src/main/java/org/apache/hudi/common/bloom/Key.java 
b/hudi-common/src/main/java/org/apache/hudi/common/bloom/Key.java
index 37ae6e68f73..f14d301ae3b 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/bloom/Key.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/bloom/Key.java
@@ -136,7 +136,7 @@ public class Key implements Comparable<Key> {
   /**
    * Serialize the fields of this object to <code>out</code>.
    *
-   * @param out <code>DataOuput</code> to serialize this object into.
+   * @param out <code>DataOutput</code> to serialize this object into.
    * @throws IOException
    */
   public void write(DataOutput out) throws IOException {
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/common/util/queue/BaseHoodieQueueBasedExecutor.java
 
b/hudi-common/src/main/java/org/apache/hudi/common/util/queue/BaseHoodieQueueBasedExecutor.java
index 20b9c802f60..f2843c56b03 100644
--- 
a/hudi-common/src/main/java/org/apache/hudi/common/util/queue/BaseHoodieQueueBasedExecutor.java
+++ 
b/hudi-common/src/main/java/org/apache/hudi/common/util/queue/BaseHoodieQueueBasedExecutor.java
@@ -215,7 +215,7 @@ public abstract class BaseHoodieQueueBasedExecutor<I, O, E> 
implements HoodieExe
         // to be interrupted as well
         Thread.currentThread().interrupt();
       }
-      // throw if we have any other exception seen already. There is a chance 
that cancellation/closing of producers with CompeletableFuture wins before the 
actual exception
+      // throw if we have any other exception seen already. There is a chance 
that cancellation/closing of producers with CompletableFuture wins before the 
actual exception
       // is thrown.
       if (this.queue.getThrowable() != null) {
         throw new HoodieException(queue.getThrowable());
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieTableMetadataUtil.java
 
b/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieTableMetadataUtil.java
index 7167a785f9f..0aa11042ab9 100644
--- 
a/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieTableMetadataUtil.java
+++ 
b/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieTableMetadataUtil.java
@@ -1615,7 +1615,7 @@ public class HoodieTableMetadataUtil {
    *
    * @param partitionType         Type of the partition for which the file 
group count is to be estimated.
    * @param recordCount           The number of records expected to be written.
-   * @param averageRecordSize     Average size of each record to be writen.
+   * @param averageRecordSize     Average size of each record to be written.
    * @param minFileGroupCount     Minimum number of file groups to use.
    * @param maxFileGroupCount     Maximum number of file groups to use.
    * @param growthFactor          By what factor are the records (recordCount) 
expected to grow?

Reply via email to