This is an automated email from the ASF dual-hosted git repository.

danny0405 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/master by this push:
     new 817b3ad7de92 fix(common): fix typos commited -> committed, commiting 
-> committing (#18363)
817b3ad7de92 is described below

commit 817b3ad7de92602ae301c0b1d1931e016f1261a3
Author: Xinli Shang <[email protected]>
AuthorDate: Sun Mar 22 20:55:01 2026 -0700

    fix(common): fix typos commited -> committed, commiting -> committing 
(#18363)
---
 .../hudi/client/transaction/SchemaConflictResolutionStrategy.java | 2 +-
 .../transaction/SimpleSchemaConflictResolutionStrategy.java       | 4 ++--
 .../java/org/apache/hudi/metadata/HoodieMetadataWriteUtils.java   | 8 ++++----
 .../hudi/table/action/commit/FlinkAutoCommitActionExecutor.java   | 2 +-
 .../apache/hudi/table/action/commit/SparkAutoCommitExecutor.java  | 2 +-
 5 files changed, 9 insertions(+), 9 deletions(-)

diff --git 
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SchemaConflictResolutionStrategy.java
 
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SchemaConflictResolutionStrategy.java
index 2eb09bd27d6b..f56f31dd3841 100644
--- 
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SchemaConflictResolutionStrategy.java
+++ 
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SchemaConflictResolutionStrategy.java
@@ -64,7 +64,7 @@ public interface SchemaConflictResolutionStrategy {
             + "schema when transaction enters validation phase: %s 
tableSchemaAtTxnValidation, "
             + "schema the transaction tries to commit with: %s. 
lastCompletedTxnOwnerInstant is %s "
             + " and currTxnOwnerInstant is %s.",
-        tableSchemaAtTxnStart.isPresent() ? tableSchemaAtTxnStart : "Not 
exists as no commited txn at that time",
+        tableSchemaAtTxnStart.isPresent() ? tableSchemaAtTxnStart : "Not 
exists as no committed txn at that time",
         tableSchemaAtTxnValidation.isPresent() ? tableSchemaAtTxnValidation : 
"Not exists",
         writerSchemaOfTxn,
         lastCompletedTxnOwnerInstant.isPresent() ? 
lastCompletedTxnOwnerInstant : "Not exists",
diff --git 
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SimpleSchemaConflictResolutionStrategy.java
 
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SimpleSchemaConflictResolutionStrategy.java
index 0ae986cd0f78..cfcd26362552 100644
--- 
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SimpleSchemaConflictResolutionStrategy.java
+++ 
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SimpleSchemaConflictResolutionStrategy.java
@@ -81,7 +81,7 @@ public class SimpleSchemaConflictResolutionStrategy 
implements SchemaConflictRes
         : null;
     // If lastCompletedInstantAtTxnValidation is null there are 2 
possibilities:
     // - No committed txn at validation starts
-    // - [Almost impossible, so we ignore it] there is a commited txn, yet it 
is archived which cannot be found
+    // - [Almost impossible, so we ignore it] there is a committed txn, yet it 
is archived which cannot be found
     // in the active timeline.
     HoodieInstant lastCompletedInstantAtTxnValidation = 
schemaResolver.computeSchemaEvolutionTimelineInReverseOrder().findFirst().orElse(null);
     // Please refer to RFC 82 for details of the case numbers.
@@ -143,7 +143,7 @@ public class SimpleSchemaConflictResolutionStrategy 
implements SchemaConflictRes
 
     // Incompatible case 8: Initial table schema is S1, there is a concurrent 
txn evolves schema to S2,
     // current writer schema is S3.
-    // Before the curr txn started, there are commited txn, with optional txn 
that commited during the
+    // Before the curr txn started, there are committed txn, with optional txn 
that committed during the
     // read-write phase of the curr txn (they can lead to concurrently schema 
evolution along with the curr txn).
     // table schema: 
----------------S1----------------------------S2-------------------------
     // txn 1(S1):     |validate & commit|
diff --git 
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieMetadataWriteUtils.java
 
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieMetadataWriteUtils.java
index 571d19b2ea22..bcabac2c516d 100644
--- 
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieMetadataWriteUtils.java
+++ 
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieMetadataWriteUtils.java
@@ -504,7 +504,7 @@ public class HoodieMetadataWriteUtils {
                 .filter(stats -> 
filesToFetchColumnStats.contains(stats.getFileName()))
                 
.map(HoodieColumnRangeMetadata::fromColumnStats).collectAsList();
             // fileColumnMetadata already contains stats for the files from 
the current inflight commit.
-            // Here it adds the stats for the commited files part of the 
latest merged file slices
+            // Here it adds the stats for the committed files part of the 
latest merged file slices
             fileColumnMetadata.addAll(partitionColumnMetadata);
             return Pair.of(partitionName, fileColumnMetadata);
           });
@@ -552,16 +552,16 @@ public class HoodieMetadataWriteUtils {
                                                        Map<String, 
Set<String>> fileGroupIdsToReplaceMap,
                                                        List<String> 
colsToIndex,
                                                        HoodieIndexVersion 
partitionStatsIndexVersion) {
-    // Get the latest merged file slices based on the commited files part of 
the latest snapshot and the new files of the current commit metadata
+    // Get the latest merged file slices based on the committed files part of 
the latest snapshot and the new files of the current commit metadata
     List<StoragePathInfo> consolidatedPathInfos = new ArrayList<>();
     partitionedWriteStat.forEach(
         stat -> consolidatedPathInfos.add(
             new StoragePathInfo(new StoragePath(dataMetaClient.getBasePath(), 
stat.getPath()), stat.getFileSizeInBytes(), false, (short) 0, 0, 0)));
-    SyncableFileSystemView fileSystemViewForCommitedFiles =
+    SyncableFileSystemView fileSystemViewForCommittedFiles =
         FileSystemViewManager.createViewManager(new 
HoodieLocalEngineContext(dataMetaClient.getStorageConf()),
             dataWriteConfig.getMetadataConfig(), 
dataWriteConfig.getViewStorageConfig(), dataWriteConfig.getCommonConfig(),
             unused -> tableMetadata).getFileSystemView(dataMetaClient);
-    
fileSystemViewForCommitedFiles.getLatestMergedFileSlicesBeforeOrOn(partitionName,
 maxInstantTime)
+    
fileSystemViewForCommittedFiles.getLatestMergedFileSlicesBeforeOrOn(partitionName,
 maxInstantTime)
         .forEach(fileSlice -> {
           if (fileSlice.getBaseFile().isPresent()) {
             
consolidatedPathInfos.add(getBaseFileStoragePathInfo(fileSlice.getBaseFile().get()));
diff --git 
a/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/table/action/commit/FlinkAutoCommitActionExecutor.java
 
b/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/table/action/commit/FlinkAutoCommitActionExecutor.java
index 899d80968f7f..76f87a6f8d4e 100644
--- 
a/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/table/action/commit/FlinkAutoCommitActionExecutor.java
+++ 
b/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/table/action/commit/FlinkAutoCommitActionExecutor.java
@@ -24,7 +24,7 @@ import org.apache.hudi.table.action.HoodieWriteMetadata;
 import java.util.List;
 
 /**
- * Simpler Wrapper to assist with auto commiting with action executor layer.
+ * Simpler Wrapper to assist with auto committing with action executor layer.
  * Only PartitionTTL and DeletePartition action executors are leveraging this 
at the moment.
  */
 public class FlinkAutoCommitActionExecutor {
diff --git 
a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkAutoCommitExecutor.java
 
b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkAutoCommitExecutor.java
index 58bd50eb9793..ec71b836f5d2 100644
--- 
a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkAutoCommitExecutor.java
+++ 
b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkAutoCommitExecutor.java
@@ -23,7 +23,7 @@ import org.apache.hudi.common.data.HoodieData;
 import org.apache.hudi.table.action.HoodieWriteMetadata;
 
 /**
- * Simpler Wrapper to assist with auto commiting with action executor layer.
+ * Simpler Wrapper to assist with auto committing with action executor layer.
  * Only bootstrap, PartitionTTL action executors are leveraging this at the 
moment.
  */
 public class SparkAutoCommitExecutor {

Reply via email to