This is an automated email from the ASF dual-hosted git repository.
yuxia pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/fluss.git
The following commit(s) were added to refs/heads/main by this push:
new c5e651313 [hotfix] Fix retrieve typos (#2727)
c5e651313 is described below
commit c5e651313286c26217fd8629eb344a0fc629e721
Author: xiaozhou <[email protected]>
AuthorDate: Thu Feb 26 19:54:35 2026 +0800
[hotfix] Fix retrieve typos (#2727)
---
.../lake/paimon/tiering/PaimonLakeCommitter.java | 2 +-
.../DvTableReadableSnapshotRetrieverTest.java | 36 +++++++++++-----------
.../metadata/CoordinatorMetadataProvider.java | 6 ++--
.../fluss/server/utils/TableAssignmentUtils.java | 16 +++++-----
.../coordinator/CoordinatorEventProcessorTest.java | 4 +--
5 files changed, 32 insertions(+), 32 deletions(-)
diff --git
a/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeCommitter.java
b/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeCommitter.java
index 5a98dcd73..03ced0db2 100644
---
a/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeCommitter.java
+++
b/fluss-lake/fluss-lake-paimon/src/main/java/org/apache/fluss/lake/paimon/tiering/PaimonLakeCommitter.java
@@ -116,7 +116,7 @@ public class PaimonLakeCommitter implements
LakeCommitter<PaimonWriteResult, Pai
if (!fileStoreTable.coreOptions().deletionVectorsEnabled()) {
return
LakeCommitResult.committedIsReadable(committedSnapshotId);
} else {
- // retrive the readable snapshot during commit
+ // retrieve the readable snapshot during commit
try (DvTableReadableSnapshotRetriever retriever =
new DvTableReadableSnapshotRetriever(
tablePath, tableId, fileStoreTable,
flussClientConfig)) {
diff --git
a/fluss-lake/fluss-lake-paimon/src/test/java/org/apache/fluss/lake/paimon/utils/DvTableReadableSnapshotRetrieverTest.java
b/fluss-lake/fluss-lake-paimon/src/test/java/org/apache/fluss/lake/paimon/utils/DvTableReadableSnapshotRetrieverTest.java
index a5a12e214..0ce3d5b8e 100644
---
a/fluss-lake/fluss-lake-paimon/src/test/java/org/apache/fluss/lake/paimon/utils/DvTableReadableSnapshotRetrieverTest.java
+++
b/fluss-lake/fluss-lake-paimon/src/test/java/org/apache/fluss/lake/paimon/utils/DvTableReadableSnapshotRetrieverTest.java
@@ -168,7 +168,7 @@ class DvTableReadableSnapshotRetrieverTest {
// No readable_snapshot yet (all buckets have L0 files)
DvTableReadableSnapshotRetriever.ReadableSnapshotResult
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot1);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot1);
assertThat(readableSnapshotAndOffsets).isNull();
// commit tiered snapshot and readable snapshot to fluss
// (simulate TieringCommitOperator) behavior
@@ -221,9 +221,9 @@ class DvTableReadableSnapshotRetrieverTest {
fileStoreTable,
Collections.singletonMap(bucket0,
generateRows(bucket0, 3, 8)));
tieredLakeSnapshotEndOffset.put(tb0, 8L);
- // retrive readable snapshot and offsets
+ // retrieve readable snapshot and offsets
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot4);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot4);
// (simulate TieringCommitOperator) behavior
commitSnapshot(
tableId,
@@ -272,7 +272,7 @@ class DvTableReadableSnapshotRetrieverTest {
Collections.singletonMap(bucket1,
generateRows(bucket1, 3, 10)));
tieredLakeSnapshotEndOffset.put(tb1, 10L);
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot6);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot6);
// readable_snapshot = snapshot5 (latest compacted snapshot)
// readable_offsets: all buckets use snapshot1's offsets (base
snapshot for flushed L0)
assertThat(readableSnapshotAndOffsets.getReadableSnapshotId()).isEqualTo(snapshot5);
@@ -309,7 +309,7 @@ class DvTableReadableSnapshotRetrieverTest {
// retrieve the readable snapshot and offsets
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot8);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot8);
// readable_snapshot = snapshot7
// readable_offsets: bucket0 uses snapshot4's offset (8L), others use
snapshot1's offset
assertThat(readableSnapshotAndOffsets.getReadableSnapshotId()).isEqualTo(snapshot7);
@@ -362,7 +362,7 @@ class DvTableReadableSnapshotRetrieverTest {
appendRows12.put(bucket2, generateRows(bucket2, 3, 11));
long snapshot12 = writeAndCommitData(fileStoreTable, appendRows12);
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot12);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot12);
assertThat(readableSnapshotAndOffsets.getReadableSnapshotId()).isEqualTo(snapshot11);
tieredLakeSnapshotEndOffset.put(tb0, 13L);
tieredLakeSnapshotEndOffset.put(tb1, 20L);
@@ -398,7 +398,7 @@ class DvTableReadableSnapshotRetrieverTest {
// Create an empty tiered snapshot (snapshot14) to simulate tiered
snapshot commit
long snapshot14 = writeAndCommitData(fileStoreTable,
Collections.emptyMap());
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot14);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot14);
// readable_snapshot = snapshot13
// readable_offsets: bucket0 uses snapshot4's offset (8L), bucket1
uses snapshot6's offset
// (10L), bucket2 uses snapshot12's offset (11L)
@@ -418,13 +418,13 @@ class DvTableReadableSnapshotRetrieverTest {
tablePath,
snapshot14,
tieredLakeSnapshotEndOffset,
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot14));
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot14));
// when the compacted snapshot is already registered in ZK,
// getReadableSnapshotAndOffsets skips recomputation and returns null.
long snapshot15 = writeAndCommitData(fileStoreTable,
Collections.emptyMap());
DvTableReadableSnapshotRetriever.ReadableSnapshotResult result15 =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot15);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot15);
assertThat(result15)
.as(
"Compacted snapshot 13 is already in ZK, should skip
recomputation and return null")
@@ -474,7 +474,7 @@ class DvTableReadableSnapshotRetrieverTest {
// No readable_snapshot yet (all buckets have L0 files)
DvTableReadableSnapshotRetriever.ReadableSnapshotResult
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot1);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot1);
assertThat(readableSnapshotAndOffsets).isNull();
commitSnapshot(
tableId,
@@ -518,7 +518,7 @@ class DvTableReadableSnapshotRetrieverTest {
// files flushed from the first tiered snapshot to latest compacted
snapshot
assertThat(readableSnapshotAndOffsets).isNull();
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot3);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot3);
assertThat(readableSnapshotAndOffsets).isNull();
commitSnapshot(
tableId,
@@ -559,7 +559,7 @@ class DvTableReadableSnapshotRetrieverTest {
long snapshot6 = writeAndCommitData(fileStoreTable, appendRowsP0More);
tieredLakeSnapshotEndOffset.put(tbP0B0, 6L);
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot6);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot6);
// readable_snapshot = snapshot5
// readable_offsets: partition0/bucket0 uses snapshot1's offset (3L),
// partition0/bucket1 uses snapshot1's offset (3L),
@@ -596,7 +596,7 @@ class DvTableReadableSnapshotRetrieverTest {
long snapshot8 = writeAndCommitData(fileStoreTable,
Collections.emptyMap());
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot8);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot8);
// readable_snapshot = snapshot7
// readable_offsets: partition0/bucket0 uses snapshot6's offset (6L),
// partition0/bucket1 uses snapshot1's offset (3L),
@@ -634,7 +634,7 @@ class DvTableReadableSnapshotRetrieverTest {
// Snapshot 7 is already registered in Fluss cluster, so the retrieve
result is null
// (no update needed).
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot9);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot9);
assertThat(readableSnapshotAndOffsets).isNull();
commitSnapshot(
tableId,
@@ -660,7 +660,7 @@ class DvTableReadableSnapshotRetrieverTest {
long snapshot11 = writeAndCommitData(fileStoreTable,
Collections.emptyMap());
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot11);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot11);
// readable_snapshot = snapshot10
// readable_offsets: partition0/bucket0 uses snapshot6's offset (6L),
// partition0/bucket1 uses snapshot1's offset (3L),
@@ -699,7 +699,7 @@ class DvTableReadableSnapshotRetrieverTest {
tablePath,
snapshot12,
tieredLakeSnapshotEndOffset,
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot12));
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot12));
// Step 10: COMPACT snapshot 13 - compact partition0, bucket1 again
(flushes snapshot12's
// L0)
@@ -719,7 +719,7 @@ class DvTableReadableSnapshotRetrieverTest {
long snapshot14 = writeAndCommitData(fileStoreTable,
Collections.emptyMap());
readableSnapshotAndOffsets =
- retriveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot14);
+ retrieveReadableSnapshotAndOffsets(tablePath, fileStoreTable,
snapshot14);
// readable_snapshot = snapshot13
// readable_offsets: partition0/bucket0 uses snapshot6's offset (6L),
// partition0/bucket1 uses snapshot12's offset (6L)
- flushed in s13,
@@ -862,7 +862,7 @@ class DvTableReadableSnapshotRetrieverTest {
}
private DvTableReadableSnapshotRetriever.ReadableSnapshotResult
- retriveReadableSnapshotAndOffsets(
+ retrieveReadableSnapshotAndOffsets(
TablePath tablePath, FileStoreTable fileStoreTable, long
tieredSnapshot)
throws Exception {
try (DvTableReadableSnapshotRetriever retriever =
diff --git
a/fluss-server/src/main/java/org/apache/fluss/server/metadata/CoordinatorMetadataProvider.java
b/fluss-server/src/main/java/org/apache/fluss/server/metadata/CoordinatorMetadataProvider.java
index f119c49db..27511d054 100644
---
a/fluss-server/src/main/java/org/apache/fluss/server/metadata/CoordinatorMetadataProvider.java
+++
b/fluss-server/src/main/java/org/apache/fluss/server/metadata/CoordinatorMetadataProvider.java
@@ -122,16 +122,16 @@ public class CoordinatorMetadataProvider extends
ZkBasedMetadataProvider {
* @param ctx the coordinator context containing leader and epoch
information
* @param tableId the table identifier
* @param partitionId the partition identifier, null for non-partitioned
tables
- * @param tableAssigment the assignment map from bucket ID to list of
replica server IDs
+ * @param tableAssignment the assignment map from bucket ID to list of
replica server IDs
* @return a list of bucket metadata objects containing complete bucket
information
*/
private static List<BucketMetadata> getBucketMetadataFromContext(
CoordinatorContext ctx,
long tableId,
@Nullable Long partitionId,
- Map<Integer, List<Integer>> tableAssigment) {
+ Map<Integer, List<Integer>> tableAssignment) {
List<BucketMetadata> bucketMetadataList = new ArrayList<>();
- tableAssigment.forEach(
+ tableAssignment.forEach(
(bucketId, serverIds) -> {
TableBucket tableBucket = new TableBucket(tableId,
partitionId, bucketId);
Optional<LeaderAndIsr> optLeaderAndIsr =
ctx.getBucketLeaderAndIsr(tableBucket);
diff --git
a/fluss-server/src/main/java/org/apache/fluss/server/utils/TableAssignmentUtils.java
b/fluss-server/src/main/java/org/apache/fluss/server/utils/TableAssignmentUtils.java
index eca96ddf5..e69b295f3 100644
---
a/fluss-server/src/main/java/org/apache/fluss/server/utils/TableAssignmentUtils.java
+++
b/fluss-server/src/main/java/org/apache/fluss/server/utils/TableAssignmentUtils.java
@@ -63,7 +63,7 @@ public class TableAssignmentUtils {
}
if (Arrays.stream(servers).noneMatch(tsInfo -> tsInfo.getRack() !=
null)) {
- return generateRackUnawareAssigment(
+ return generateRackUnawareAssignment(
nBuckets,
replicationFactor,
Arrays.stream(servers).mapToInt(TabletServerInfo::getId).toArray(),
@@ -74,7 +74,7 @@ public class TableAssignmentUtils {
throw new InvalidServerRackInfoException(
"Not all tabletServers have rack information for
replica rack aware assignment.");
} else {
- return generateRackAwareAssigment(
+ return generateRackAwareAssignment(
nBuckets, replicationFactor, servers, startIndex,
nextReplicaShift);
}
}
@@ -107,12 +107,12 @@ public class TableAssignmentUtils {
* <tr><td>bucket5 </td><td>bucket6 </td><td>bucket7
</td><td>bucket8 </td><td>bucket9 </td><td>(1st replica)</td></tr>
* <tr><td>bucket4 </td><td>bucket0 </td><td>bucket1
</td><td>bucket2 </td><td>bucket3 </td><td>(2nd replica)</td></tr>
* <tr><td>bucket8 </td><td>bucket9 </td><td>bucket5
</td><td>bucket6 </td><td>bucket7 </td><td>(2nd replica)</td></tr>
- * <tr><td>bucket3 </td><td>bucket4 </td><td>bucket0
</td><td>bucket1 </td><td>bucket2 </td><td>(3nd replica)</td></tr>
- * <tr><td>bucket7 </td><td>bucket8 </td><td>bucket9
</td><td>bucket5 </td><td>bucket6 </td><td>(3nd replica)</td></tr>
+ * <tr><td>bucket3 </td><td>bucket4 </td><td>bucket0
</td><td>bucket1 </td><td>bucket2 </td><td>(3rd replica)</td></tr>
+ * <tr><td>bucket7 </td><td>bucket8 </td><td>bucket9
</td><td>bucket5 </td><td>bucket6 </td><td>(3rd replica)</td></tr>
* </table>
*
- * <p>To create rack aware assigment, this API will first create a rack
alternated tabletServers
- * list. For example, from this tabletServerId -> rack mapping:
+ * <p>To create rack aware assignment, this API will first create a rack
alternated
+ * tabletServers list. For example, from this tabletServerId -> rack
mapping:
*
* <pre>
* 0 -> "rack1"
@@ -172,7 +172,7 @@ public class TableAssignmentUtils {
randomInt(servers.length));
}
- private static TableAssignment generateRackUnawareAssigment(
+ private static TableAssignment generateRackUnawareAssignment(
int nBuckets,
int replicationFactor,
int[] serverIds,
@@ -198,7 +198,7 @@ public class TableAssignmentUtils {
return new TableAssignment(assignments);
}
- private static TableAssignment generateRackAwareAssigment(
+ private static TableAssignment generateRackAwareAssignment(
int nBuckets,
int replicationFactor,
TabletServerInfo[] servers,
diff --git
a/fluss-server/src/test/java/org/apache/fluss/server/coordinator/CoordinatorEventProcessorTest.java
b/fluss-server/src/test/java/org/apache/fluss/server/coordinator/CoordinatorEventProcessorTest.java
index c9e8bf5a0..1f19d4da7 100644
---
a/fluss-server/src/test/java/org/apache/fluss/server/coordinator/CoordinatorEventProcessorTest.java
+++
b/fluss-server/src/test/java/org/apache/fluss/server/coordinator/CoordinatorEventProcessorTest.java
@@ -1370,8 +1370,8 @@ class CoordinatorEventProcessorTest {
});
}
- private <T> T fromCtx(Function<CoordinatorContext, T> retriveFunction)
throws Exception {
- AccessContextEvent<T> event = new
AccessContextEvent<>(retriveFunction);
+ private <T> T fromCtx(Function<CoordinatorContext, T> retrieveFunction)
throws Exception {
+ AccessContextEvent<T> event = new
AccessContextEvent<>(retrieveFunction);
eventProcessor.getCoordinatorEventManager().put(event);
return event.getResultFuture().get(30, TimeUnit.SECONDS);
}