This is an automated email from the ASF dual-hosted git repository.
danny0405 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/master by this push:
new 817b3ad7de92 fix(common): fix typos commited -> committed, commiting
-> committing (#18363)
817b3ad7de92 is described below
commit 817b3ad7de92602ae301c0b1d1931e016f1261a3
Author: Xinli Shang <[email protected]>
AuthorDate: Sun Mar 22 20:55:01 2026 -0700
fix(common): fix typos commited -> committed, commiting -> committing
(#18363)
---
.../hudi/client/transaction/SchemaConflictResolutionStrategy.java | 2 +-
.../transaction/SimpleSchemaConflictResolutionStrategy.java | 4 ++--
.../java/org/apache/hudi/metadata/HoodieMetadataWriteUtils.java | 8 ++++----
.../hudi/table/action/commit/FlinkAutoCommitActionExecutor.java | 2 +-
.../apache/hudi/table/action/commit/SparkAutoCommitExecutor.java | 2 +-
5 files changed, 9 insertions(+), 9 deletions(-)
diff --git
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SchemaConflictResolutionStrategy.java
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SchemaConflictResolutionStrategy.java
index 2eb09bd27d6b..f56f31dd3841 100644
---
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SchemaConflictResolutionStrategy.java
+++
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SchemaConflictResolutionStrategy.java
@@ -64,7 +64,7 @@ public interface SchemaConflictResolutionStrategy {
+ "schema when transaction enters validation phase: %s
tableSchemaAtTxnValidation, "
+ "schema the transaction tries to commit with: %s.
lastCompletedTxnOwnerInstant is %s "
+ " and currTxnOwnerInstant is %s.",
- tableSchemaAtTxnStart.isPresent() ? tableSchemaAtTxnStart : "Not
exists as no commited txn at that time",
+ tableSchemaAtTxnStart.isPresent() ? tableSchemaAtTxnStart : "Not
exists as no committed txn at that time",
tableSchemaAtTxnValidation.isPresent() ? tableSchemaAtTxnValidation :
"Not exists",
writerSchemaOfTxn,
lastCompletedTxnOwnerInstant.isPresent() ?
lastCompletedTxnOwnerInstant : "Not exists",
diff --git
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SimpleSchemaConflictResolutionStrategy.java
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SimpleSchemaConflictResolutionStrategy.java
index 0ae986cd0f78..cfcd26362552 100644
---
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SimpleSchemaConflictResolutionStrategy.java
+++
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/client/transaction/SimpleSchemaConflictResolutionStrategy.java
@@ -81,7 +81,7 @@ public class SimpleSchemaConflictResolutionStrategy
implements SchemaConflictRes
: null;
// If lastCompletedInstantAtTxnValidation is null there are 2
possibilities:
// - No committed txn at validation starts
- // - [Almost impossible, so we ignore it] there is a commited txn, yet it
is archived which cannot be found
+ // - [Almost impossible, so we ignore it] there is a committed txn, yet it
is archived which cannot be found
// in the active timeline.
HoodieInstant lastCompletedInstantAtTxnValidation =
schemaResolver.computeSchemaEvolutionTimelineInReverseOrder().findFirst().orElse(null);
// Please refer to RFC 82 for details of the case numbers.
@@ -143,7 +143,7 @@ public class SimpleSchemaConflictResolutionStrategy
implements SchemaConflictRes
// Incompatible case 8: Initial table schema is S1, there is a concurrent
txn evolves schema to S2,
// current writer schema is S3.
- // Before the curr txn started, there are commited txn, with optional txn
that commited during the
+ // Before the curr txn started, there are committed txn, with optional txn
that committed during the
// read-write phase of the curr txn (they can lead to concurrently schema
evolution along with the curr txn).
// table schema:
----------------S1----------------------------S2-------------------------
// txn 1(S1): |validate & commit|
diff --git
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieMetadataWriteUtils.java
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieMetadataWriteUtils.java
index 571d19b2ea22..bcabac2c516d 100644
---
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieMetadataWriteUtils.java
+++
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieMetadataWriteUtils.java
@@ -504,7 +504,7 @@ public class HoodieMetadataWriteUtils {
.filter(stats ->
filesToFetchColumnStats.contains(stats.getFileName()))
.map(HoodieColumnRangeMetadata::fromColumnStats).collectAsList();
// fileColumnMetadata already contains stats for the files from
the current inflight commit.
- // Here it adds the stats for the commited files part of the
latest merged file slices
+ // Here it adds the stats for the committed files part of the
latest merged file slices
fileColumnMetadata.addAll(partitionColumnMetadata);
return Pair.of(partitionName, fileColumnMetadata);
});
@@ -552,16 +552,16 @@ public class HoodieMetadataWriteUtils {
Map<String,
Set<String>> fileGroupIdsToReplaceMap,
List<String>
colsToIndex,
HoodieIndexVersion
partitionStatsIndexVersion) {
- // Get the latest merged file slices based on the commited files part of
the latest snapshot and the new files of the current commit metadata
+ // Get the latest merged file slices based on the committed files part of
the latest snapshot and the new files of the current commit metadata
List<StoragePathInfo> consolidatedPathInfos = new ArrayList<>();
partitionedWriteStat.forEach(
stat -> consolidatedPathInfos.add(
new StoragePathInfo(new StoragePath(dataMetaClient.getBasePath(),
stat.getPath()), stat.getFileSizeInBytes(), false, (short) 0, 0, 0)));
- SyncableFileSystemView fileSystemViewForCommitedFiles =
+ SyncableFileSystemView fileSystemViewForCommittedFiles =
FileSystemViewManager.createViewManager(new
HoodieLocalEngineContext(dataMetaClient.getStorageConf()),
dataWriteConfig.getMetadataConfig(),
dataWriteConfig.getViewStorageConfig(), dataWriteConfig.getCommonConfig(),
unused -> tableMetadata).getFileSystemView(dataMetaClient);
-
fileSystemViewForCommitedFiles.getLatestMergedFileSlicesBeforeOrOn(partitionName,
maxInstantTime)
+
fileSystemViewForCommittedFiles.getLatestMergedFileSlicesBeforeOrOn(partitionName,
maxInstantTime)
.forEach(fileSlice -> {
if (fileSlice.getBaseFile().isPresent()) {
consolidatedPathInfos.add(getBaseFileStoragePathInfo(fileSlice.getBaseFile().get()));
diff --git
a/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/table/action/commit/FlinkAutoCommitActionExecutor.java
b/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/table/action/commit/FlinkAutoCommitActionExecutor.java
index 899d80968f7f..76f87a6f8d4e 100644
---
a/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/table/action/commit/FlinkAutoCommitActionExecutor.java
+++
b/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/table/action/commit/FlinkAutoCommitActionExecutor.java
@@ -24,7 +24,7 @@ import org.apache.hudi.table.action.HoodieWriteMetadata;
import java.util.List;
/**
- * Simpler Wrapper to assist with auto commiting with action executor layer.
+ * Simpler Wrapper to assist with auto committing with action executor layer.
* Only PartitionTTL and DeletePartition action executors are leveraging this
at the moment.
*/
public class FlinkAutoCommitActionExecutor {
diff --git
a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkAutoCommitExecutor.java
b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkAutoCommitExecutor.java
index 58bd50eb9793..ec71b836f5d2 100644
---
a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkAutoCommitExecutor.java
+++
b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkAutoCommitExecutor.java
@@ -23,7 +23,7 @@ import org.apache.hudi.common.data.HoodieData;
import org.apache.hudi.table.action.HoodieWriteMetadata;
/**
- * Simpler Wrapper to assist with auto commiting with action executor layer.
+ * Simpler Wrapper to assist with auto committing with action executor layer.
* Only bootstrap, PartitionTTL action executors are leveraging this at the
moment.
*/
public class SparkAutoCommitExecutor {