This is an automated email from the ASF dual-hosted git repository.
vinoyang pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-hudi.git
The following commit(s) were added to refs/heads/master by this push:
new 78b3194 [HUDI-751] Fix some coding issues reported by FindBugs (#1470)
78b3194 is described below
commit 78b3194e8241c519a85310997f31b2b55df487e1
Author: Shaofeng Shi <[email protected]>
AuthorDate: Tue Mar 31 21:19:32 2020 +0800
[HUDI-751] Fix some coding issues reported by FindBugs (#1470)
---
.../org/apache/hudi/cli/commands/RollbacksCommand.java | 2 +-
.../org/apache/hudi/cli/commands/SparkEnvCommand.java | 6 +++---
.../java/org/apache/hudi/cli/commands/StatsCommand.java | 6 +++---
.../main/java/org/apache/hudi/cli/utils/HiveUtil.java | 16 ++++------------
.../java/org/apache/hudi/client/HoodieCleanClient.java | 1 +
.../java/org/apache/hudi/client/HoodieReadClient.java | 1 +
.../java/org/apache/hudi/client/HoodieWriteClient.java | 1 +
.../main/java/org/apache/hudi/client/WriteStatus.java | 1 +
.../org/apache/hudi/config/HoodieHBaseIndexConfig.java | 2 +-
.../BoundedPartitionAwareCompactionStrategy.java | 2 +-
.../compact/strategy/DayBasedCompactionStrategy.java | 6 +++---
.../hudi/common/config/SerializableConfiguration.java | 1 +
.../org/apache/hudi/common/model/HoodieBaseFile.java | 1 +
.../java/org/apache/hudi/common/model/HoodieLogFile.java | 2 ++
.../java/org/apache/hudi/common/model/HoodieRecord.java | 10 +++++-----
.../hudi/common/model/HoodieRollingStatMetadata.java | 2 +-
.../apache/hudi/common/table/HoodieTableMetaClient.java | 7 ++++---
.../common/table/timeline/HoodieDefaultTimeline.java | 1 +
.../java/org/apache/hudi/common/util/HoodieTimer.java | 2 +-
.../apache/hudi/common/util/collection/DiskBasedMap.java | 2 +-
.../org/apache/hudi/hadoop/HoodieROTablePathFilter.java | 1 +
.../hive/SlashEncodedDayPartitionValueExtractor.java | 1 +
.../org/apache/hudi/utilities/HDFSParquetImporter.java | 1 +
.../apache/hudi/utilities/deltastreamer/Compactor.java | 1 +
.../apache/hudi/utilities/deltastreamer/DeltaSync.java | 5 +++--
.../utilities/deltastreamer/HoodieDeltaStreamer.java | 3 +++
.../apache/hudi/utilities/perf/TimelineServerPerf.java | 1 +
.../org/apache/hudi/utilities/sources/CsvDFSSource.java | 8 +++++---
.../hudi/utilities/sources/HiveIncrPullSource.java | 2 ++
.../hudi/utilities/sources/helpers/AvroConvertor.java | 1 +
.../hudi/utilities/sources/helpers/DFSPathSelector.java | 2 +-
31 files changed, 57 insertions(+), 41 deletions(-)
diff --git
a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java
b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java
index 9e4bf28..70b34bc 100644
--- a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java
+++ b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java
@@ -120,7 +120,7 @@ public class RollbacksCommand implements CommandMarker {
/**
* An Active timeline containing only rollbacks.
*/
- class RollbackTimeline extends HoodieActiveTimeline {
+ static class RollbackTimeline extends HoodieActiveTimeline {
public RollbackTimeline(HoodieTableMetaClient metaClient) {
super(metaClient,
CollectionUtils.createImmutableSet(HoodieTimeline.ROLLBACK_EXTENSION));
diff --git
a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java
b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java
index d209a08..7969808 100644
--- a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java
+++ b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java
@@ -34,7 +34,7 @@ import java.util.Map;
@Component
public class SparkEnvCommand implements CommandMarker {
- public static Map<String, String> env = new HashMap<String, String>();
+ public static Map<String, String> env = new HashMap<>();
@CliCommand(value = "set", help = "Set spark launcher env to cli")
public void setEnv(@CliOption(key = {"conf"}, help = "Env config to be set")
final String confMap) {
@@ -49,8 +49,8 @@ public class SparkEnvCommand implements CommandMarker {
public String showAllEnv() {
String[][] rows = new String[env.size()][2];
int i = 0;
- for (String key: env.keySet()) {
- rows[i] = new String[]{key, env.get(key)};
+ for (Map.Entry<String, String> entry: env.entrySet()) {
+ rows[i] = new String[]{entry.getKey(), entry.getValue()};
i++;
}
return HoodiePrintHelper.print(new String[] {"key", "value"}, rows);
diff --git
a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java
b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java
index 9db544c..e5be0e4 100644
--- a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java
+++ b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java
@@ -131,9 +131,9 @@ public class StatsCommand implements CommandMarker {
}
List<Comparable[]> rows = new ArrayList<>();
- for (String instantTime : commitHistoMap.keySet()) {
- Snapshot s = commitHistoMap.get(instantTime).getSnapshot();
- rows.add(printFileSizeHistogram(instantTime, s));
+ for (Map.Entry<String, Histogram> entry : commitHistoMap.entrySet()) {
+ Snapshot s = entry.getValue().getSnapshot();
+ rows.add(printFileSizeHistogram(entry.getKey(), s));
}
Snapshot s = globalHistogram.getSnapshot();
rows.add(printFileSizeHistogram("ALL", s));
diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java
b/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java
index 291ac74..37f705e 100644
--- a/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java
+++ b/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java
@@ -49,10 +49,9 @@ public class HiveUtil {
public static long countRecords(String jdbcUrl, HoodieTableMetaClient
source, String dbName, String user, String pass)
throws SQLException {
- Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass);
ResultSet rs = null;
- Statement stmt = conn.createStatement();
- try {
+ try (Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass);
+ Statement stmt = conn.createStatement()) {
// stmt.execute("set mapred.job.queue.name=<queue_name>");
stmt.execute("set
hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat");
stmt.execute("set hive.stats.autogather=false");
@@ -68,9 +67,6 @@ public class HiveUtil {
if (rs != null) {
rs.close();
}
- if (stmt != null) {
- stmt.close();
- }
}
}
@@ -88,10 +84,9 @@ public class HiveUtil {
private static long countRecords(String jdbcUrl, HoodieTableMetaClient
source, String srcDb, String startDateStr,
String endDateStr, String user, String pass) throws SQLException {
- Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass);
ResultSet rs = null;
- Statement stmt = conn.createStatement();
- try {
+ try (Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass);
+ Statement stmt = conn.createStatement()) {
// stmt.execute("set mapred.job.queue.name=<queue_name>");
stmt.execute("set
hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat");
stmt.execute("set hive.stats.autogather=false");
@@ -106,9 +101,6 @@ public class HiveUtil {
if (rs != null) {
rs.close();
}
- if (stmt != null) {
- stmt.close();
- }
}
}
}
diff --git
a/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java
b/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java
index 43a3655..d622f70 100644
--- a/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java
+++ b/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java
@@ -47,6 +47,7 @@ import java.util.List;
public class HoodieCleanClient<T extends HoodieRecordPayload> extends
AbstractHoodieClient {
+ private static final long serialVersionUID = 1L;
private static final Logger LOG =
LogManager.getLogger(HoodieCleanClient.class);
private final transient HoodieMetrics metrics;
diff --git
a/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java
b/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java
index d1e92b5..2dedfaf 100644
--- a/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java
+++ b/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java
@@ -57,6 +57,7 @@ import scala.Tuple2;
*/
public class HoodieReadClient<T extends HoodieRecordPayload> implements
Serializable {
+ private static final long serialVersionUID = 1L;
private static final Logger LOG =
LogManager.getLogger(HoodieReadClient.class);
/**
diff --git
a/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java
b/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java
index a0d1867..2a25a76 100644
--- a/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java
+++ b/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java
@@ -93,6 +93,7 @@ import scala.Tuple2;
*/
public class HoodieWriteClient<T extends HoodieRecordPayload> extends
AbstractHoodieWriteClient<T> {
+ private static final long serialVersionUID = 1L;
private static final Logger LOG =
LogManager.getLogger(HoodieWriteClient.class);
private static final String LOOKUP_STR = "lookup";
private final boolean rollbackPending;
diff --git a/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java
b/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java
index 7dd89e0..a93f268 100644
--- a/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java
+++ b/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java
@@ -35,6 +35,7 @@ import java.util.Random;
*/
public class WriteStatus implements Serializable {
+ private static final long serialVersionUID = 1L;
private static final long RANDOM_SEED = 9038412832L;
private final HashMap<HoodieKey, Throwable> errors = new HashMap<>();
diff --git
a/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java
b/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java
index 68ada22..709f08c 100644
---
a/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java
+++
b/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java
@@ -61,7 +61,7 @@ public class HoodieHBaseIndexConfig extends
DefaultHoodieConfig {
* value based on global indexing throughput needs and most importantly, how
much the HBase installation in use is
* able to tolerate without Region Servers going down.
*/
- public static String HBASE_MAX_QPS_PER_REGION_SERVER_PROP =
"hoodie.index.hbase.max.qps.per.region.server";
+ public static final String HBASE_MAX_QPS_PER_REGION_SERVER_PROP =
"hoodie.index.hbase.max.qps.per.region.server";
/**
* Default batch size, used only for Get, but computed for Put.
*/
diff --git
a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java
b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java
index 0472008..24da226 100644
---
a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java
+++
b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java
@@ -39,7 +39,7 @@ import java.util.stream.Collectors;
*/
public class BoundedPartitionAwareCompactionStrategy extends
DayBasedCompactionStrategy {
- SimpleDateFormat dateFormat = new SimpleDateFormat(datePartitionFormat);
+ SimpleDateFormat dateFormat = new SimpleDateFormat(DATE_PARTITION_FORMAT);
@Override
public List<HoodieCompactionOperation> orderAndFilter(HoodieWriteConfig
writeConfig,
diff --git
a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java
b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java
index cf7727b..ea29fdd 100644
---
a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java
+++
b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java
@@ -39,14 +39,14 @@ import java.util.stream.Collectors;
public class DayBasedCompactionStrategy extends CompactionStrategy {
// For now, use SimpleDateFormat as default partition format
- protected static String datePartitionFormat = "yyyy/MM/dd";
+ protected static final String DATE_PARTITION_FORMAT = "yyyy/MM/dd";
// Sorts compaction in LastInFirstCompacted order
protected static Comparator<String> comparator = (String leftPartition,
String rightPartition) -> {
try {
leftPartition = getPartitionPathWithoutPartitionKeys(leftPartition);
rightPartition = getPartitionPathWithoutPartitionKeys(rightPartition);
- Date left = new SimpleDateFormat(datePartitionFormat,
Locale.ENGLISH).parse(leftPartition);
- Date right = new SimpleDateFormat(datePartitionFormat,
Locale.ENGLISH).parse(rightPartition);
+ Date left = new SimpleDateFormat(DATE_PARTITION_FORMAT,
Locale.ENGLISH).parse(leftPartition);
+ Date right = new SimpleDateFormat(DATE_PARTITION_FORMAT,
Locale.ENGLISH).parse(rightPartition);
return left.after(right) ? -1 : right.after(left) ? 1 : 0;
} catch (ParseException e) {
throw new HoodieException("Invalid Partition Date Format", e);
diff --git
a/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java
b/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java
index 8ad82d4..408cb17 100644
---
a/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java
+++
b/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java
@@ -30,6 +30,7 @@ import java.io.Serializable;
*/
public class SerializableConfiguration implements Serializable {
+ private static final long serialVersionUID = 1L;
private transient Configuration configuration;
public SerializableConfiguration(Configuration configuration) {
diff --git
a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java
b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java
index ee20136..5ccc2a8 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java
@@ -31,6 +31,7 @@ import java.util.Objects;
*/
public class HoodieBaseFile implements Serializable {
+ private static final long serialVersionUID = 1L;
private transient FileStatus fileStatus;
private final String fullPath;
private long fileLen;
diff --git
a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java
b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java
index a0a22aa..973fd51 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java
@@ -37,6 +37,7 @@ import java.util.Objects;
*/
public class HoodieLogFile implements Serializable {
+ private static final long serialVersionUID = 1L;
public static final String DELTA_EXTENSION = ".log";
public static final Integer LOGFILE_BASE_VERSION = 1;
@@ -129,6 +130,7 @@ public class HoodieLogFile implements Serializable {
*/
public static class LogFileComparator implements Comparator<HoodieLogFile>,
Serializable {
+ private static final long serialVersionUID = 1L;
private transient Comparator<String> writeTokenComparator;
private Comparator<String> getWriteTokenComparator() {
diff --git
a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java
b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java
index 57f5fa7..df7843f 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java
@@ -30,11 +30,11 @@ import java.util.Objects;
*/
public class HoodieRecord<T extends HoodieRecordPayload> implements
Serializable {
- public static String COMMIT_TIME_METADATA_FIELD = "_hoodie_commit_time";
- public static String COMMIT_SEQNO_METADATA_FIELD = "_hoodie_commit_seqno";
- public static String RECORD_KEY_METADATA_FIELD = "_hoodie_record_key";
- public static String PARTITION_PATH_METADATA_FIELD =
"_hoodie_partition_path";
- public static String FILENAME_METADATA_FIELD = "_hoodie_file_name";
+ public static final String COMMIT_TIME_METADATA_FIELD =
"_hoodie_commit_time";
+ public static final String COMMIT_SEQNO_METADATA_FIELD =
"_hoodie_commit_seqno";
+ public static final String RECORD_KEY_METADATA_FIELD = "_hoodie_record_key";
+ public static final String PARTITION_PATH_METADATA_FIELD =
"_hoodie_partition_path";
+ public static final String FILENAME_METADATA_FIELD = "_hoodie_file_name";
public static final List<String> HOODIE_META_COLUMNS =
CollectionUtils.createImmutableList(COMMIT_TIME_METADATA_FIELD,
COMMIT_SEQNO_METADATA_FIELD,
diff --git
a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java
b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java
index fc0ac19..a354092 100644
---
a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java
+++
b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java
@@ -52,7 +52,7 @@ public class HoodieRollingStatMetadata implements
Serializable {
this.actionType = actionType;
}
- class RollingStatsHashMap<K, V> extends HashMap<K, V> {
+ static class RollingStatsHashMap<K, V> extends HashMap<K, V> {
@Override
public V put(K key, V value) {
diff --git
a/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java
b/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java
index 4ad8b51..03e91c2 100644
---
a/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java
+++
b/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java
@@ -68,10 +68,11 @@ import java.util.stream.Stream;
*/
public class HoodieTableMetaClient implements Serializable {
+ private static final long serialVersionUID = 1L;
private static final Logger LOG =
LogManager.getLogger(HoodieTableMetaClient.class);
- public static String METAFOLDER_NAME = ".hoodie";
- public static String TEMPFOLDER_NAME = METAFOLDER_NAME + File.separator +
".temp";
- public static String AUXILIARYFOLDER_NAME = METAFOLDER_NAME + File.separator
+ ".aux";
+ public static final String METAFOLDER_NAME = ".hoodie";
+ public static final String TEMPFOLDER_NAME = METAFOLDER_NAME +
File.separator + ".temp";
+ public static final String AUXILIARYFOLDER_NAME = METAFOLDER_NAME +
File.separator + ".aux";
public static final String MARKER_EXTN = ".marker";
private String basePath;
diff --git
a/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java
b/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java
index 3f400be..5fa4a02 100644
---
a/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java
+++
b/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java
@@ -48,6 +48,7 @@ import static java.util.Collections.reverse;
*/
public class HoodieDefaultTimeline implements HoodieTimeline {
+ private static final long serialVersionUID = 1L;
private static final Logger LOG =
LogManager.getLogger(HoodieDefaultTimeline.class);
private static final String HASHING_ALGORITHM = "SHA-256";
diff --git
a/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java
b/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java
index 2b8fac8..0ccc7ca 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java
@@ -32,7 +32,7 @@ public class HoodieTimer {
// Ordered stack of TimeInfo's to make sure stopping the timer returns the
correct elapsed time
Deque<TimeInfo> timeInfoDeque = new ArrayDeque<>();
- class TimeInfo {
+ static class TimeInfo {
// captures the startTime of the code block
long startTime;
diff --git
a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java
b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java
index 90977dd..f4b5924 100644
---
a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java
+++
b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java
@@ -55,7 +55,7 @@ import java.util.stream.Stream;
*/
public final class DiskBasedMap<T extends Serializable, R extends
Serializable> implements Map<T, R>, Iterable<R> {
- public static int BUFFER_SIZE = 128 * 1024; // 128 KB
+ public static final int BUFFER_SIZE = 128 * 1024; // 128 KB
private static final Logger LOG = LogManager.getLogger(DiskBasedMap.class);
// Stores the key and corresponding value's latest metadata spilled to disk
private final Map<T, ValueMetadata> valueMetadataMap;
diff --git
a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java
b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java
index c0bc13b..4058875 100644
---
a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java
+++
b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java
@@ -51,6 +51,7 @@ import java.util.stream.Collectors;
*/
public class HoodieROTablePathFilter implements PathFilter, Serializable {
+ private static final long serialVersionUID = 1L;
private static final Logger LOG =
LogManager.getLogger(HoodieROTablePathFilter.class);
/**
diff --git
a/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java
b/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java
index 4dba0fe..f4a4c90 100644
---
a/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java
+++
b/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java
@@ -33,6 +33,7 @@ import java.util.List;
*/
public class SlashEncodedDayPartitionValueExtractor implements
PartitionValueExtractor {
+ private static final long serialVersionUID = 1L;
private transient DateTimeFormatter dtfOut;
public SlashEncodedDayPartitionValueExtractor() {
diff --git
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java
index 0ed09cd..f389c58 100644
---
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java
+++
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java
@@ -66,6 +66,7 @@ import scala.Tuple2;
*/
public class HDFSParquetImporter implements Serializable {
+ private static final long serialVersionUID = 1L;
private static final Logger LOG =
LogManager.getLogger(HDFSParquetImporter.class);
private static final DateTimeFormatter PARTITION_FORMATTER =
DateTimeFormatter.ofPattern("yyyy/MM/dd")
diff --git
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java
index eafe6c8..4c23537 100644
---
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java
+++
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java
@@ -37,6 +37,7 @@ import java.io.Serializable;
*/
public class Compactor implements Serializable {
+ private static final long serialVersionUID = 1L;
private static final Logger LOG = LogManager.getLogger(Compactor.class);
private transient HoodieWriteClient compactionClient;
diff --git
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java
index a29db9b..99cb497 100644
---
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java
+++
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java
@@ -83,9 +83,10 @@ import static
org.apache.hudi.utilities.schema.RowBasedSchemaProvider.HOODIE_REC
*/
public class DeltaSync implements Serializable {
+ private static final long serialVersionUID = 1L;
private static final Logger LOG = LogManager.getLogger(DeltaSync.class);
- public static String CHECKPOINT_KEY = "deltastreamer.checkpoint.key";
- public static String CHECKPOINT_RESET_KEY =
"deltastreamer.checkpoint.reset_key";
+ public static final String CHECKPOINT_KEY = "deltastreamer.checkpoint.key";
+ public static final String CHECKPOINT_RESET_KEY =
"deltastreamer.checkpoint.reset_key";
/**
* Delta Sync Config.
diff --git
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java
index 4fa43f4..948033f 100644
---
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java
+++
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java
@@ -77,6 +77,7 @@ import java.util.stream.IntStream;
*/
public class HoodieDeltaStreamer implements Serializable {
+ private static final long serialVersionUID = 1L;
private static final Logger LOG =
LogManager.getLogger(HoodieDeltaStreamer.class);
public static String CHECKPOINT_KEY = "deltastreamer.checkpoint.key";
@@ -302,6 +303,7 @@ public class HoodieDeltaStreamer implements Serializable {
*/
public static class DeltaSyncService extends AbstractDeltaStreamerService {
+ private static final long serialVersionUID = 1L;
/**
* Delta Sync Config.
*/
@@ -489,6 +491,7 @@ public class HoodieDeltaStreamer implements Serializable {
*/
public static class AsyncCompactService extends AbstractDeltaStreamerService
{
+ private static final long serialVersionUID = 1L;
private final int maxConcurrentCompaction;
private transient Compactor compactor;
private transient JavaSparkContext jssc;
diff --git
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java
index ca5ab24..a6692cc 100644
---
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java
+++
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java
@@ -58,6 +58,7 @@ import java.util.stream.IntStream;
public class TimelineServerPerf implements Serializable {
+ private static final long serialVersionUID = 1L;
private static final Logger LOG =
LogManager.getLogger(TimelineServerPerf.class);
private final Config cfg;
private transient TimelineService timelineServer;
diff --git
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java
index 228d46d..3d158ba 100644
---
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java
+++
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java
@@ -58,10 +58,12 @@ import java.util.List;
* passed to the CSV reader without inferring the schema from the CSV file.
*/
public class CsvDFSSource extends RowSource {
+
+ private static final long serialVersionUID = 1L;
// CsvSource config prefix
- public static final String CSV_SRC_CONFIG_PREFIX =
"hoodie.deltastreamer.csv.";
+ protected static final String CSV_SRC_CONFIG_PREFIX =
"hoodie.deltastreamer.csv.";
// CSV-specific configurations to pass in from Hudi to Spark
- public static final List<String> CSV_CONFIG_KEYS = Arrays.asList(
+ protected static final List<String> CSV_CONFIG_KEYS = Arrays.asList(
"sep", "encoding", "quote", "escape", "charToEscapeQuoteEscaping",
"comment",
"header", "enforceSchema", "inferSchema", "samplingRatio",
"ignoreLeadingWhiteSpace",
"ignoreTrailingWhiteSpace", "nullValue", "emptyValue", "nanValue",
"positiveInf",
@@ -69,7 +71,7 @@ public class CsvDFSSource extends RowSource {
"mode", "columnNameOfCorruptRecord", "multiLine"
);
- private final DFSPathSelector pathSelector;
+ private final transient DFSPathSelector pathSelector;
private final StructType sourceSchema;
public CsvDFSSource(TypedProperties props,
diff --git
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java
index 76f2187..a8bea98 100644
---
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java
+++
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java
@@ -59,6 +59,8 @@ import java.util.stream.Collectors;
*/
public class HiveIncrPullSource extends AvroSource {
+ private static final long serialVersionUID = 1L;
+
private static final Logger LOG =
LogManager.getLogger(HiveIncrPullSource.class);
private final transient FileSystem fs;
diff --git
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java
index 988dbd8..449db10 100644
---
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java
+++
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java
@@ -33,6 +33,7 @@ import java.io.Serializable;
*/
public class AvroConvertor implements Serializable {
+ private static final long serialVersionUID = 1L;
/**
* To be lazily inited on executors.
*/
diff --git
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java
index 665720b..cb0100a 100644
---
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java
+++
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java
@@ -85,7 +85,7 @@ public class DFSPathSelector {
long maxModificationTime = Long.MIN_VALUE;
List<FileStatus> filteredFiles = new ArrayList<>();
for (FileStatus f : eligibleFiles) {
- if (lastCheckpointStr.isPresent() && f.getModificationTime() <=
Long.valueOf(lastCheckpointStr.get())) {
+ if (lastCheckpointStr.isPresent() && f.getModificationTime() <=
Long.valueOf(lastCheckpointStr.get()).longValue()) {
// skip processed files
continue;
}