This is an automated email from the ASF dual-hosted git repository.

yuxia pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/fluss.git


The following commit(s) were added to refs/heads/main by this push:
     new c5e651313 [hotfix] Fix retrieve typos (#2727)
c5e651313 is described below

commit c5e651313286c26217fd8629eb344a0fc629e721
Author: xiaozhou <[email protected]>
AuthorDate: Thu Feb 26 19:54:35 2026 +0800

    [hotfix] Fix retrieve typos (#2727)
---
 .../lake/paimon/tiering/PaimonLakeCommitter.java   |  2 +-
 .../DvTableReadableSnapshotRetrieverTest.java      | 36 +++++++++++-----------
 .../metadata/CoordinatorMetadataProvider.java      |  6 ++--
 .../fluss/server/utils/TableAssignmentUtils.java   | 16 +++++-----
 .../coordinator/CoordinatorEventProcessorTest.java |  4 +--
 5 files changed, 32 insertions(+), 32 deletions(-)

diff --git 
a/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeCommitter.java
 
b/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeCommitter.java
index 5a98dcd73..03ced0db2 100644
--- 
a/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeCommitter.java
+++ 
b/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeCommitter.java
@@ -116,7 +116,7 @@ public class PaimonLakeCommitter implements 
LakeCommitter<PaimonWriteResult, Pai
             if (!fileStoreTable.coreOptions().deletionVectorsEnabled()) {
                 return 
LakeCommitResult.committedIsReadable(committedSnapshotId);
             } else {
-                // retrive the readable snapshot during commit
+                // retrieve the readable snapshot during commit
                 try (DvTableReadableSnapshotRetriever retriever =
                         new DvTableReadableSnapshotRetriever(
                                 tablePath, tableId, fileStoreTable, 
flussClientConfig)) {
diff --git 
a/fluss-lake/fluss-lake-paimon/src/test/java/org/apache/fluss/lake/paimon/utils/DvTableReadableSnapshotRetrieverTest.java
 
b/fluss-lake/fluss-lake-paimon/src/test/java/org/apache/fluss/lake/paimon/utils/DvTableReadableSnapshotRetrieverTest.java
index a5a12e214..0ce3d5b8e 100644
--- 
a/fluss-lake/fluss-lake-paimon/src/test/java/org/apache/fluss/lake/paimon/utils/DvTableReadableSnapshotRetrieverTest.java
+++ 
b/fluss-lake/fluss-lake-paimon/src/test/java/org/apache/fluss/lake/paimon/utils/DvTableReadableSnapshotRetrieverTest.java
@@ -168,7 +168,7 @@ class DvTableReadableSnapshotRetrieverTest {
 
         // No readable_snapshot yet (all buckets have L0 files)
         DvTableReadableSnapshotRetriever.ReadableSnapshotResult 
readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot1);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot1);
         assertThat(readableSnapshotAndOffsets).isNull();
         // commit tiered snapshot and readable snapshot to fluss
         // (simulate TieringCommitOperator) behavior
@@ -221,9 +221,9 @@ class DvTableReadableSnapshotRetrieverTest {
                         fileStoreTable,
                         Collections.singletonMap(bucket0, 
generateRows(bucket0, 3, 8)));
         tieredLakeSnapshotEndOffset.put(tb0, 8L);
-        // retrive readable snapshot and offsets
+        // retrieve readable snapshot and offsets
         readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot4);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot4);
         // (simulate TieringCommitOperator) behavior
         commitSnapshot(
                 tableId,
@@ -272,7 +272,7 @@ class DvTableReadableSnapshotRetrieverTest {
                         Collections.singletonMap(bucket1, 
generateRows(bucket1, 3, 10)));
         tieredLakeSnapshotEndOffset.put(tb1, 10L);
         readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot6);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot6);
         // readable_snapshot = snapshot5 (latest compacted snapshot)
         // readable_offsets: all buckets use snapshot1's offsets (base 
snapshot for flushed L0)
         
assertThat(readableSnapshotAndOffsets.getReadableSnapshotId()).isEqualTo(snapshot5);
@@ -309,7 +309,7 @@ class DvTableReadableSnapshotRetrieverTest {
 
         // retrieve the readable snapshot and offsets
         readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot8);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot8);
         // readable_snapshot = snapshot7
         // readable_offsets: bucket0 uses snapshot4's offset (8L), others use 
snapshot1's offset
         
assertThat(readableSnapshotAndOffsets.getReadableSnapshotId()).isEqualTo(snapshot7);
@@ -362,7 +362,7 @@ class DvTableReadableSnapshotRetrieverTest {
         appendRows12.put(bucket2, generateRows(bucket2, 3, 11));
         long snapshot12 = writeAndCommitData(fileStoreTable, appendRows12);
         readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot12);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot12);
         
assertThat(readableSnapshotAndOffsets.getReadableSnapshotId()).isEqualTo(snapshot11);
         tieredLakeSnapshotEndOffset.put(tb0, 13L);
         tieredLakeSnapshotEndOffset.put(tb1, 20L);
@@ -398,7 +398,7 @@ class DvTableReadableSnapshotRetrieverTest {
         // Create an empty tiered snapshot (snapshot14) to simulate tiered 
snapshot commit
         long snapshot14 = writeAndCommitData(fileStoreTable, 
Collections.emptyMap());
         readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot14);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot14);
         // readable_snapshot = snapshot13
         // readable_offsets: bucket0 uses snapshot4's offset (8L), bucket1 
uses snapshot6's offset
         // (10L), bucket2 uses snapshot12's offset (11L)
@@ -418,13 +418,13 @@ class DvTableReadableSnapshotRetrieverTest {
                 tablePath,
                 snapshot14,
                 tieredLakeSnapshotEndOffset,
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot14));
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot14));
 
         // when the compacted snapshot is already registered in ZK,
         // getReadableSnapshotAndOffsets skips recomputation and returns null.
         long snapshot15 = writeAndCommitData(fileStoreTable, 
Collections.emptyMap());
         DvTableReadableSnapshotRetriever.ReadableSnapshotResult result15 =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot15);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot15);
         assertThat(result15)
                 .as(
                         "Compacted snapshot 13 is already in ZK, should skip 
recomputation and return null")
@@ -474,7 +474,7 @@ class DvTableReadableSnapshotRetrieverTest {
 
         // No readable_snapshot yet (all buckets have L0 files)
         DvTableReadableSnapshotRetriever.ReadableSnapshotResult 
readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot1);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot1);
         assertThat(readableSnapshotAndOffsets).isNull();
         commitSnapshot(
                 tableId,
@@ -518,7 +518,7 @@ class DvTableReadableSnapshotRetrieverTest {
         // files flushed from the first tiered snapshot to latest compacted 
snapshot
         assertThat(readableSnapshotAndOffsets).isNull();
         readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot3);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot3);
         assertThat(readableSnapshotAndOffsets).isNull();
         commitSnapshot(
                 tableId,
@@ -559,7 +559,7 @@ class DvTableReadableSnapshotRetrieverTest {
         long snapshot6 = writeAndCommitData(fileStoreTable, appendRowsP0More);
         tieredLakeSnapshotEndOffset.put(tbP0B0, 6L);
         readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot6);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot6);
         // readable_snapshot = snapshot5
         // readable_offsets: partition0/bucket0 uses snapshot1's offset (3L),
         //                   partition0/bucket1 uses snapshot1's offset (3L),
@@ -596,7 +596,7 @@ class DvTableReadableSnapshotRetrieverTest {
         long snapshot8 = writeAndCommitData(fileStoreTable, 
Collections.emptyMap());
 
         readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot8);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot8);
         // readable_snapshot = snapshot7
         // readable_offsets: partition0/bucket0 uses snapshot6's offset (6L),
         //                   partition0/bucket1 uses snapshot1's offset (3L),
@@ -634,7 +634,7 @@ class DvTableReadableSnapshotRetrieverTest {
         // Snapshot 7 is already registered in Fluss cluster, so the retrieve 
result is null
         // (no update needed).
         readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot9);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot9);
         assertThat(readableSnapshotAndOffsets).isNull();
         commitSnapshot(
                 tableId,
@@ -660,7 +660,7 @@ class DvTableReadableSnapshotRetrieverTest {
         long snapshot11 = writeAndCommitData(fileStoreTable, 
Collections.emptyMap());
 
         readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot11);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot11);
         // readable_snapshot = snapshot10
         // readable_offsets: partition0/bucket0 uses snapshot6's offset (6L),
         //                   partition0/bucket1 uses snapshot1's offset (3L),
@@ -699,7 +699,7 @@ class DvTableReadableSnapshotRetrieverTest {
                 tablePath,
                 snapshot12,
                 tieredLakeSnapshotEndOffset,
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot12));
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot12));
 
         // Step 10: COMPACT snapshot 13 - compact partition0, bucket1 again 
(flushes snapshot12's
         // L0)
@@ -719,7 +719,7 @@ class DvTableReadableSnapshotRetrieverTest {
         long snapshot14 = writeAndCommitData(fileStoreTable, 
Collections.emptyMap());
 
         readableSnapshotAndOffsets =
-                retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot14);
+                retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable, 
snapshot14);
         // readable_snapshot = snapshot13
         // readable_offsets: partition0/bucket0 uses snapshot6's offset (6L),
         //                   partition0/bucket1 uses snapshot12's offset (6L) 
- flushed in s13,
@@ -862,7 +862,7 @@ class DvTableReadableSnapshotRetrieverTest {
     }
 
     private DvTableReadableSnapshotRetriever.ReadableSnapshotResult
-            retriveReadableSnapshotAndOffsets(
+            retrieveReadableSnapshotAndOffsets(
                     TablePath tablePath, FileStoreTable fileStoreTable, long 
tieredSnapshot)
                     throws Exception {
         try (DvTableReadableSnapshotRetriever retriever =
diff --git 
a/fluss-server/src/main/java/org/apache/fluss/server/metadata/CoordinatorMetadataProvider.java
 
b/fluss-server/src/main/java/org/apache/fluss/server/metadata/CoordinatorMetadataProvider.java
index f119c49db..27511d054 100644
--- 
a/fluss-server/src/main/java/org/apache/fluss/server/metadata/CoordinatorMetadataProvider.java
+++ 
b/fluss-server/src/main/java/org/apache/fluss/server/metadata/CoordinatorMetadataProvider.java
@@ -122,16 +122,16 @@ public class CoordinatorMetadataProvider extends 
ZkBasedMetadataProvider {
      * @param ctx the coordinator context containing leader and epoch 
information
      * @param tableId the table identifier
      * @param partitionId the partition identifier, null for non-partitioned 
tables
-     * @param tableAssigment the assignment map from bucket ID to list of 
replica server IDs
+     * @param tableAssignment the assignment map from bucket ID to list of 
replica server IDs
      * @return a list of bucket metadata objects containing complete bucket 
information
      */
     private static List<BucketMetadata> getBucketMetadataFromContext(
             CoordinatorContext ctx,
             long tableId,
             @Nullable Long partitionId,
-            Map<Integer, List<Integer>> tableAssigment) {
+            Map<Integer, List<Integer>> tableAssignment) {
         List<BucketMetadata> bucketMetadataList = new ArrayList<>();
-        tableAssigment.forEach(
+        tableAssignment.forEach(
                 (bucketId, serverIds) -> {
                     TableBucket tableBucket = new TableBucket(tableId, 
partitionId, bucketId);
                     Optional<LeaderAndIsr> optLeaderAndIsr = 
ctx.getBucketLeaderAndIsr(tableBucket);
diff --git 
a/fluss-server/src/main/java/org/apache/fluss/server/utils/TableAssignmentUtils.java
 
b/fluss-server/src/main/java/org/apache/fluss/server/utils/TableAssignmentUtils.java
index eca96ddf5..e69b295f3 100644
--- 
a/fluss-server/src/main/java/org/apache/fluss/server/utils/TableAssignmentUtils.java
+++ 
b/fluss-server/src/main/java/org/apache/fluss/server/utils/TableAssignmentUtils.java
@@ -63,7 +63,7 @@ public class TableAssignmentUtils {
         }
 
         if (Arrays.stream(servers).noneMatch(tsInfo -> tsInfo.getRack() != 
null)) {
-            return generateRackUnawareAssigment(
+            return generateRackUnawareAssignment(
                     nBuckets,
                     replicationFactor,
                     
Arrays.stream(servers).mapToInt(TabletServerInfo::getId).toArray(),
@@ -74,7 +74,7 @@ public class TableAssignmentUtils {
                 throw new InvalidServerRackInfoException(
                         "Not all tabletServers have rack information for 
replica rack aware assignment.");
             } else {
-                return generateRackAwareAssigment(
+                return generateRackAwareAssignment(
                         nBuckets, replicationFactor, servers, startIndex, 
nextReplicaShift);
             }
         }
@@ -107,12 +107,12 @@ public class TableAssignmentUtils {
      * <tr><td>bucket5      </td><td>bucket6      </td><td>bucket7      
</td><td>bucket8      </td><td>bucket9      </td><td>(1st replica)</td></tr>
      * <tr><td>bucket4      </td><td>bucket0      </td><td>bucket1      
</td><td>bucket2      </td><td>bucket3      </td><td>(2nd replica)</td></tr>
      * <tr><td>bucket8      </td><td>bucket9      </td><td>bucket5      
</td><td>bucket6      </td><td>bucket7      </td><td>(2nd replica)</td></tr>
-     * <tr><td>bucket3      </td><td>bucket4      </td><td>bucket0      
</td><td>bucket1      </td><td>bucket2      </td><td>(3nd replica)</td></tr>
-     * <tr><td>bucket7      </td><td>bucket8      </td><td>bucket9      
</td><td>bucket5      </td><td>bucket6      </td><td>(3nd replica)</td></tr>
+     * <tr><td>bucket3      </td><td>bucket4      </td><td>bucket0      
</td><td>bucket1      </td><td>bucket2      </td><td>(3rd replica)</td></tr>
+     * <tr><td>bucket7      </td><td>bucket8      </td><td>bucket9      
</td><td>bucket5      </td><td>bucket6      </td><td>(3rd replica)</td></tr>
      * </table>
      *
-     * <p>To create rack aware assigment, this API will first create a rack 
alternated tabletServers
-     * list. For example, from this tabletServerId -> rack mapping:
+     * <p>To create rack aware assignment, this API will first create a rack 
alternated
+     * tabletServers list. For example, from this tabletServerId -> rack 
mapping:
      *
      * <pre>
      *     0 -> "rack1"
@@ -172,7 +172,7 @@ public class TableAssignmentUtils {
                 randomInt(servers.length));
     }
 
-    private static TableAssignment generateRackUnawareAssigment(
+    private static TableAssignment generateRackUnawareAssignment(
             int nBuckets,
             int replicationFactor,
             int[] serverIds,
@@ -198,7 +198,7 @@ public class TableAssignmentUtils {
         return new TableAssignment(assignments);
     }
 
-    private static TableAssignment generateRackAwareAssigment(
+    private static TableAssignment generateRackAwareAssignment(
             int nBuckets,
             int replicationFactor,
             TabletServerInfo[] servers,
diff --git 
a/fluss-server/src/test/java/org/apache/fluss/server/coordinator/CoordinatorEventProcessorTest.java
 
b/fluss-server/src/test/java/org/apache/fluss/server/coordinator/CoordinatorEventProcessorTest.java
index c9e8bf5a0..1f19d4da7 100644
--- 
a/fluss-server/src/test/java/org/apache/fluss/server/coordinator/CoordinatorEventProcessorTest.java
+++ 
b/fluss-server/src/test/java/org/apache/fluss/server/coordinator/CoordinatorEventProcessorTest.java
@@ -1370,8 +1370,8 @@ class CoordinatorEventProcessorTest {
                 });
     }
 
-    private <T> T fromCtx(Function<CoordinatorContext, T> retriveFunction) 
throws Exception {
-        AccessContextEvent<T> event = new 
AccessContextEvent<>(retriveFunction);
+    private <T> T fromCtx(Function<CoordinatorContext, T> retrieveFunction) 
throws Exception {
+        AccessContextEvent<T> event = new 
AccessContextEvent<>(retrieveFunction);
         eventProcessor.getCoordinatorEventManager().put(event);
         return event.getResultFuture().get(30, TimeUnit.SECONDS);
     }

Reply via email to