This is an automated email from the ASF dual-hosted git repository.
vbalaji pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/master by this push:
new 9d9a9a30207 [HUDI-7225] Correcting spelling errors or annotations with
non-standard spelling (#10317)
9d9a9a30207 is described below
commit 9d9a9a302077b7beb7b1f309c1278638a3f11fb8
Author: Sigma <[email protected]>
AuthorDate: Wed Dec 13 07:32:11 2023 +0800
[HUDI-7225] Correcting spelling errors or annotations with non-standard
spelling (#10317)
Co-authored-by: mazhengxuan <[email protected]>
---
.../org/apache/hudi/table/upgrade/ZeroToOneUpgradeHandler.java | 8 ++++----
.../main/java/org/apache/hudi/avro/AvroSchemaCompatibility.java | 8 ++++----
.../src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java | 2 +-
.../hudi/common/model/debezium/AbstractDebeziumAvroPayload.java | 2 +-
.../hudi/common/table/view/AbstractTableFileSystemView.java | 2 +-
.../hudi/internal/schema/convert/AvroInternalSchemaConverter.java | 6 +++---
.../hudi/internal/schema/utils/AvroSchemaEvolutionUtils.java | 2 +-
.../hudi/table/format/cow/vector/reader/AbstractColumnReader.java | 2 +-
.../hudi/table/format/cow/vector/reader/AbstractColumnReader.java | 2 +-
.../hudi/table/format/cow/vector/reader/AbstractColumnReader.java | 2 +-
.../hudi/table/format/cow/vector/reader/AbstractColumnReader.java | 2 +-
.../hudi/table/format/cow/vector/reader/AbstractColumnReader.java | 2 +-
.../src/main/java/org/apache/hudi/sync/adb/AdbSyncTool.java | 2 +-
13 files changed, 21 insertions(+), 21 deletions(-)
diff --git
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/upgrade/ZeroToOneUpgradeHandler.java
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/upgrade/ZeroToOneUpgradeHandler.java
index 772afe71b02..69e4c2b80fb 100644
---
a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/upgrade/ZeroToOneUpgradeHandler.java
+++
b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/upgrade/ZeroToOneUpgradeHandler.java
@@ -59,8 +59,8 @@ public class ZeroToOneUpgradeHandler implements
UpgradeHandler {
HoodieTimeline inflightTimeline =
table.getMetaClient().getCommitsTimeline().filterPendingExcludingMajorAndMinorCompaction();
List<String> commits =
inflightTimeline.getReverseOrderedInstants().map(HoodieInstant::getTimestamp)
.collect(Collectors.toList());
- if (commits.size() > 0 && instantTime != null) {
- // ignore the latest inflight commit since a new commit would have been
started and we need to fix any pending commits from previous launch
+ if (!commits.isEmpty() && instantTime != null) {
+ // ignore the latest inflight commit since a new commit would have been
started, and we need to fix any pending commits from previous launch
commits.remove(instantTime);
}
for (String commit : commits) {
@@ -125,8 +125,8 @@ public class ZeroToOneUpgradeHandler implements
UpgradeHandler {
/**
* Curates file name for marker from existing log file path.
- * log file format : partitionpath/.fileid_baseInstant.log.writetoken
- * marker file format :
partitionpath/fileId_writetoken_baseinstant.basefileExtn.marker.APPEND
+ * log file format : partitionPath/.fileid_baseInstant.log.writeToken
+ * marker file format :
partitionPath/fileId_writeToken_baseInstant.baseFileExtn.marker.APPEND
*
* @param logFilePath log file path for which marker file name needs to be
generated.
* @param table {@link HoodieTable} instance to use
diff --git
a/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaCompatibility.java
b/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaCompatibility.java
index f25824dbd4a..63629b31ddd 100644
---
a/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaCompatibility.java
+++
b/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaCompatibility.java
@@ -548,7 +548,7 @@ public class AvroSchemaCompatibility {
}
/**
- * Identifies the type of a schema compatibility result.
+ * Identifies the type of schema compatibility result.
*/
public enum SchemaCompatibilityType {
COMPATIBLE, INCOMPATIBLE,
@@ -885,7 +885,7 @@ public class AvroSchemaCompatibility {
private final Schema mWriter;
/**
- * Human readable description of this result.
+ * Human-readable description of this result.
*/
private final String mDescription;
@@ -942,9 +942,9 @@ public class AvroSchemaCompatibility {
}
/**
- * Gets a human readable description of this validation result.
+ * Gets a human-readable description of this validation result.
*
- * @return a human readable description of this validation result.
+ * @return a human-readable description of this validation result.
*/
public String getDescription() {
return mDescription;
diff --git
a/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java
b/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java
index a8a72e77717..b0489d75ae0 100644
--- a/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java
+++ b/hudi-common/src/main/java/org/apache/hudi/avro/AvroSchemaUtils.java
@@ -150,7 +150,7 @@ public class AvroSchemaUtils {
* Validate whether the {@code targetSchema} is a "compatible" projection of
{@code sourceSchema}.
* Only difference of this method from {@link #isStrictProjectionOf(Schema,
Schema)} is
* the fact that it allows some legitimate type promotions (like {@code int
-> long},
- * {@code decimal(3, 2) -> decimal(5, 2)}, etc) that allows projection to
have a "wider"
+ * {@code decimal(3, 2) -> decimal(5, 2)}, etc.) that allows projection to
have a "wider"
* atomic type (whereas strict projection requires atomic type to be
identical)
*/
public static boolean isCompatibleProjectionOf(Schema sourceSchema, Schema
targetSchema) {
diff --git
a/hudi-common/src/main/java/org/apache/hudi/common/model/debezium/AbstractDebeziumAvroPayload.java
b/hudi-common/src/main/java/org/apache/hudi/common/model/debezium/AbstractDebeziumAvroPayload.java
index 69cefc73d72..82395d6f606 100644
---
a/hudi-common/src/main/java/org/apache/hudi/common/model/debezium/AbstractDebeziumAvroPayload.java
+++
b/hudi-common/src/main/java/org/apache/hudi/common/model/debezium/AbstractDebeziumAvroPayload.java
@@ -72,7 +72,7 @@ public abstract class AbstractDebeziumAvroPayload extends
OverwriteWithLatestAvr
if (shouldPickCurrentRecord(currentValue, insertValue.get(), schema)) {
return Option.of(currentValue);
}
- // Step 2: Pick the insert record (as a delete record if its a deleted
event)
+ // Step 2: Pick the insert record (as a delete record if it is a deleted
event)
return getInsertValue(schema);
}
diff --git
a/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java
b/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java
index 14f55b76889..76750171fb5 100644
---
a/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java
+++
b/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java
@@ -1270,7 +1270,7 @@ public abstract class AbstractTableFileSystemView
implements SyncableFileSystemV
abstract Stream<Pair<String, CompactionOperation>>
fetchPendingLogCompactionOperations();
/**
- * Check if there is an bootstrap base file present for this file.
+ * Check if there is a bootstrap base file present for this file.
*
* @param fgId File-Group Id
* @return true if there is associated bootstrap base-file, false otherwise
diff --git
a/hudi-common/src/main/java/org/apache/hudi/internal/schema/convert/AvroInternalSchemaConverter.java
b/hudi-common/src/main/java/org/apache/hudi/internal/schema/convert/AvroInternalSchemaConverter.java
index 786ac538271..601718e6702 100644
---
a/hudi-common/src/main/java/org/apache/hudi/internal/schema/convert/AvroInternalSchemaConverter.java
+++
b/hudi-common/src/main/java/org/apache/hudi/internal/schema/convert/AvroInternalSchemaConverter.java
@@ -160,7 +160,7 @@ public class AvroInternalSchemaConverter {
* @param schema a avro schema.
* @param visited track the visit node when do traversal for avro schema;
used to check if the name of avro record schema is correct.
* @param firstVisitRoot track whether the current visited schema node is a
root node.
- * @param nextId a initial id which used to create id for all fields.
+ * @param nextId an initial id which used to create id for all fields.
* @return a hudi type match avro schema.
*/
private static Type visitAvroSchemaToBuildType(Schema schema, Deque<String>
visited, Boolean firstVisitRoot, AtomicInteger nextId) {
@@ -275,7 +275,7 @@ public class AvroInternalSchemaConverter {
*
* @param type a hudi type.
* @param recordName the record name
- * @return a Avro schema match this type
+ * @return an Avro schema match this type
*/
public static Schema buildAvroSchemaFromType(Type type, String recordName) {
Map<Type, Schema> cache = new HashMap<>();
@@ -301,7 +301,7 @@ public class AvroInternalSchemaConverter {
* @param cache use to cache intermediate convert result to save cost.
* @param recordName auto-generated record name used as a fallback, in case
* {@link org.apache.hudi.internal.schema.Types.RecordType} doesn't bear
original record-name
- * @return a Avro schema match this type
+ * @return an Avro schema match this type
*/
private static Schema visitInternalSchemaToBuildAvroSchema(Type type,
Map<Type, Schema> cache, String recordName) {
switch (type.typeId()) {
diff --git
a/hudi-common/src/main/java/org/apache/hudi/internal/schema/utils/AvroSchemaEvolutionUtils.java
b/hudi-common/src/main/java/org/apache/hudi/internal/schema/utils/AvroSchemaEvolutionUtils.java
index 35ca13820f2..c09b21ad3ae 100644
---
a/hudi-common/src/main/java/org/apache/hudi/internal/schema/utils/AvroSchemaEvolutionUtils.java
+++
b/hudi-common/src/main/java/org/apache/hudi/internal/schema/utils/AvroSchemaEvolutionUtils.java
@@ -77,7 +77,7 @@ public class AvroSchemaEvolutionUtils {
// Remove redundancy from diffFromEvolutionSchema.
// for example, now we add a struct col in evolvedSchema, the struct col
is " user struct<name:string, age:int> "
- // when we do diff operation: user, user.name, user.age will appeared in
the resultSet which is redundancy, user.name and user.age should be excluded.
+ // when we do diff operation: user, user.name, user.age will appear in the
resultSet which is redundancy, user.name and user.age should be excluded.
// deal with add operation
TreeMap<Integer, String> finalAddAction = new TreeMap<>();
for (int i = 0; i < diffFromEvolutionColumns.size(); i++) {
diff --git
a/hudi-flink-datasource/hudi-flink1.14.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
b/hudi-flink-datasource/hudi-flink1.14.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
index 07416a37171..3fee3b4b630 100644
---
a/hudi-flink-datasource/hudi-flink1.14.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
+++
b/hudi-flink-datasource/hudi-flink1.14.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
@@ -98,7 +98,7 @@ public abstract class AbstractColumnReader<V extends
WritableColumnVector>
* Input streams:
* 1.Run length encoder to encode every data, so we have run length stream
to get
* run length information.
- * 2.Data maybe is real data, maybe is dictionary ids which need be decode
to real
+ * 2.Data maybe is real data, maybe is dictionary ids which need be decoded
to real
* data from Dictionary.
*
* Run length stream ------> Data stream
diff --git
a/hudi-flink-datasource/hudi-flink1.15.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
b/hudi-flink-datasource/hudi-flink1.15.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
index a8b733de636..ab21f257692 100644
---
a/hudi-flink-datasource/hudi-flink1.15.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
+++
b/hudi-flink-datasource/hudi-flink1.15.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
@@ -98,7 +98,7 @@ public abstract class AbstractColumnReader<V extends
WritableColumnVector>
* Input streams:
* 1.Run length encoder to encode every data, so we have run length stream
to get
* run length information.
- * 2.Data maybe is real data, maybe is dictionary ids which need be decode
to real
+ * 2.Data maybe is real data, maybe is dictionary ids which need be decoded
to real
* data from Dictionary.
*
* Run length stream ------> Data stream
diff --git
a/hudi-flink-datasource/hudi-flink1.16.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
b/hudi-flink-datasource/hudi-flink1.16.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
index a8b733de636..ab21f257692 100644
---
a/hudi-flink-datasource/hudi-flink1.16.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
+++
b/hudi-flink-datasource/hudi-flink1.16.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
@@ -98,7 +98,7 @@ public abstract class AbstractColumnReader<V extends
WritableColumnVector>
* Input streams:
* 1.Run length encoder to encode every data, so we have run length stream
to get
* run length information.
- * 2.Data maybe is real data, maybe is dictionary ids which need be decode
to real
+ * 2.Data maybe is real data, maybe is dictionary ids which need be decoded
to real
* data from Dictionary.
*
* Run length stream ------> Data stream
diff --git
a/hudi-flink-datasource/hudi-flink1.17.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
b/hudi-flink-datasource/hudi-flink1.17.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
index a8b733de636..ab21f257692 100644
---
a/hudi-flink-datasource/hudi-flink1.17.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
+++
b/hudi-flink-datasource/hudi-flink1.17.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
@@ -98,7 +98,7 @@ public abstract class AbstractColumnReader<V extends
WritableColumnVector>
* Input streams:
* 1.Run length encoder to encode every data, so we have run length stream
to get
* run length information.
- * 2.Data maybe is real data, maybe is dictionary ids which need be decode
to real
+ * 2.Data maybe is real data, maybe is dictionary ids which need be decoded
to real
* data from Dictionary.
*
* Run length stream ------> Data stream
diff --git
a/hudi-flink-datasource/hudi-flink1.18.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
b/hudi-flink-datasource/hudi-flink1.18.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
index a8b733de636..ab21f257692 100644
---
a/hudi-flink-datasource/hudi-flink1.18.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
+++
b/hudi-flink-datasource/hudi-flink1.18.x/src/main/java/org/apache/hudi/table/format/cow/vector/reader/AbstractColumnReader.java
@@ -98,7 +98,7 @@ public abstract class AbstractColumnReader<V extends
WritableColumnVector>
* Input streams:
* 1.Run length encoder to encode every data, so we have run length stream
to get
* run length information.
- * 2.Data maybe is real data, maybe is dictionary ids which need be decode
to real
+ * 2.Data maybe is real data, maybe is dictionary ids which need be decoded
to real
* data from Dictionary.
*
* Run length stream ------> Data stream
diff --git
a/hudi-sync/hudi-adb-sync/src/main/java/org/apache/hudi/sync/adb/AdbSyncTool.java
b/hudi-sync/hudi-adb-sync/src/main/java/org/apache/hudi/sync/adb/AdbSyncTool.java
index 7b4e42b7e83..50213a51973 100644
---
a/hudi-sync/hudi-adb-sync/src/main/java/org/apache/hudi/sync/adb/AdbSyncTool.java
+++
b/hudi-sync/hudi-adb-sync/src/main/java/org/apache/hudi/sync/adb/AdbSyncTool.java
@@ -189,7 +189,7 @@ public class AdbSyncTool extends HoodieSyncTool {
syncPartitions(tableName, writtenPartitionsSince);
// Update sync commit time
- // whether to skip syncing commit time stored in tbl properties, since it
is time consuming.
+ // whether to skip syncing commit time stored in tbl properties, since it
is time-consuming.
if (!config.getBoolean(ADB_SYNC_SKIP_LAST_COMMIT_TIME_SYNC)) {
syncClient.updateLastCommitTimeSynced(tableName);
}