This is an automated email from the ASF dual-hosted git repository.

vinoyang pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-hudi.git


The following commit(s) were added to refs/heads/master by this push:
     new 78b3194  [HUDI-751] Fix some coding issues reported by FindBugs (#1470)
78b3194 is described below

commit 78b3194e8241c519a85310997f31b2b55df487e1
Author: Shaofeng Shi <shaofeng...@gmail.com>
AuthorDate: Tue Mar 31 21:19:32 2020 +0800

    [HUDI-751] Fix some coding issues reported by FindBugs (#1470)
---
 .../org/apache/hudi/cli/commands/RollbacksCommand.java   |  2 +-
 .../org/apache/hudi/cli/commands/SparkEnvCommand.java    |  6 +++---
 .../java/org/apache/hudi/cli/commands/StatsCommand.java  |  6 +++---
 .../main/java/org/apache/hudi/cli/utils/HiveUtil.java    | 16 ++++------------
 .../java/org/apache/hudi/client/HoodieCleanClient.java   |  1 +
 .../java/org/apache/hudi/client/HoodieReadClient.java    |  1 +
 .../java/org/apache/hudi/client/HoodieWriteClient.java   |  1 +
 .../main/java/org/apache/hudi/client/WriteStatus.java    |  1 +
 .../org/apache/hudi/config/HoodieHBaseIndexConfig.java   |  2 +-
 .../BoundedPartitionAwareCompactionStrategy.java         |  2 +-
 .../compact/strategy/DayBasedCompactionStrategy.java     |  6 +++---
 .../hudi/common/config/SerializableConfiguration.java    |  1 +
 .../org/apache/hudi/common/model/HoodieBaseFile.java     |  1 +
 .../java/org/apache/hudi/common/model/HoodieLogFile.java |  2 ++
 .../java/org/apache/hudi/common/model/HoodieRecord.java  | 10 +++++-----
 .../hudi/common/model/HoodieRollingStatMetadata.java     |  2 +-
 .../apache/hudi/common/table/HoodieTableMetaClient.java  |  7 ++++---
 .../common/table/timeline/HoodieDefaultTimeline.java     |  1 +
 .../java/org/apache/hudi/common/util/HoodieTimer.java    |  2 +-
 .../apache/hudi/common/util/collection/DiskBasedMap.java |  2 +-
 .../org/apache/hudi/hadoop/HoodieROTablePathFilter.java  |  1 +
 .../hive/SlashEncodedDayPartitionValueExtractor.java     |  1 +
 .../org/apache/hudi/utilities/HDFSParquetImporter.java   |  1 +
 .../apache/hudi/utilities/deltastreamer/Compactor.java   |  1 +
 .../apache/hudi/utilities/deltastreamer/DeltaSync.java   |  5 +++--
 .../utilities/deltastreamer/HoodieDeltaStreamer.java     |  3 +++
 .../apache/hudi/utilities/perf/TimelineServerPerf.java   |  1 +
 .../org/apache/hudi/utilities/sources/CsvDFSSource.java  |  8 +++++---
 .../hudi/utilities/sources/HiveIncrPullSource.java       |  2 ++
 .../hudi/utilities/sources/helpers/AvroConvertor.java    |  1 +
 .../hudi/utilities/sources/helpers/DFSPathSelector.java  |  2 +-
 31 files changed, 57 insertions(+), 41 deletions(-)

diff --git 
a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java 
b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java
index 9e4bf28..70b34bc 100644
--- a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java
+++ b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java
@@ -120,7 +120,7 @@ public class RollbacksCommand implements CommandMarker {
   /**
    * An Active timeline containing only rollbacks.
    */
-  class RollbackTimeline extends HoodieActiveTimeline {
+  static class RollbackTimeline extends HoodieActiveTimeline {
 
     public RollbackTimeline(HoodieTableMetaClient metaClient) {
       super(metaClient, 
CollectionUtils.createImmutableSet(HoodieTimeline.ROLLBACK_EXTENSION));
diff --git 
a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java 
b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java
index d209a08..7969808 100644
--- a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java
+++ b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java
@@ -34,7 +34,7 @@ import java.util.Map;
 @Component
 public class SparkEnvCommand implements CommandMarker {
 
-  public static Map<String, String> env = new HashMap<String, String>();
+  public static Map<String, String> env = new HashMap<>();
 
   @CliCommand(value = "set", help = "Set spark launcher env to cli")
   public void setEnv(@CliOption(key = {"conf"}, help = "Env config to be set") 
final String confMap) {
@@ -49,8 +49,8 @@ public class SparkEnvCommand implements CommandMarker {
   public String showAllEnv() {
     String[][] rows = new String[env.size()][2];
     int i = 0;
-    for (String key: env.keySet()) {
-      rows[i] = new String[]{key, env.get(key)};
+    for (Map.Entry<String, String> entry: env.entrySet()) {
+      rows[i] = new String[]{entry.getKey(), entry.getValue()};
       i++;
     }
     return HoodiePrintHelper.print(new String[] {"key", "value"}, rows);
diff --git 
a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java 
b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java
index 9db544c..e5be0e4 100644
--- a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java
+++ b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java
@@ -131,9 +131,9 @@ public class StatsCommand implements CommandMarker {
     }
 
     List<Comparable[]> rows = new ArrayList<>();
-    for (String instantTime : commitHistoMap.keySet()) {
-      Snapshot s = commitHistoMap.get(instantTime).getSnapshot();
-      rows.add(printFileSizeHistogram(instantTime, s));
+    for (Map.Entry<String, Histogram> entry : commitHistoMap.entrySet()) {
+      Snapshot s = entry.getValue().getSnapshot();
+      rows.add(printFileSizeHistogram(entry.getKey(), s));
     }
     Snapshot s = globalHistogram.getSnapshot();
     rows.add(printFileSizeHistogram("ALL", s));
diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java 
b/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java
index 291ac74..37f705e 100644
--- a/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java
+++ b/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java
@@ -49,10 +49,9 @@ public class HiveUtil {
 
   public static long countRecords(String jdbcUrl, HoodieTableMetaClient 
source, String dbName, String user, String pass)
       throws SQLException {
-    Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass);
     ResultSet rs = null;
-    Statement stmt = conn.createStatement();
-    try {
+    try (Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass);
+         Statement stmt = conn.createStatement()) {
       // stmt.execute("set mapred.job.queue.name=<queue_name>");
       stmt.execute("set 
hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat");
       stmt.execute("set hive.stats.autogather=false");
@@ -68,9 +67,6 @@ public class HiveUtil {
       if (rs != null) {
         rs.close();
       }
-      if (stmt != null) {
-        stmt.close();
-      }
     }
   }
 
@@ -88,10 +84,9 @@ public class HiveUtil {
 
   private static long countRecords(String jdbcUrl, HoodieTableMetaClient 
source, String srcDb, String startDateStr,
       String endDateStr, String user, String pass) throws SQLException {
-    Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass);
     ResultSet rs = null;
-    Statement stmt = conn.createStatement();
-    try {
+    try (Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass);
+         Statement stmt = conn.createStatement()) {
       // stmt.execute("set mapred.job.queue.name=<queue_name>");
       stmt.execute("set 
hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat");
       stmt.execute("set hive.stats.autogather=false");
@@ -106,9 +101,6 @@ public class HiveUtil {
       if (rs != null) {
         rs.close();
       }
-      if (stmt != null) {
-        stmt.close();
-      }
     }
   }
 }
diff --git 
a/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java 
b/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java
index 43a3655..d622f70 100644
--- a/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java
+++ b/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java
@@ -47,6 +47,7 @@ import java.util.List;
 
 public class HoodieCleanClient<T extends HoodieRecordPayload> extends 
AbstractHoodieClient {
 
+  private static final long serialVersionUID = 1L;
   private static final Logger LOG = 
LogManager.getLogger(HoodieCleanClient.class);
   private final transient HoodieMetrics metrics;
 
diff --git 
a/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java 
b/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java
index d1e92b5..2dedfaf 100644
--- a/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java
+++ b/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java
@@ -57,6 +57,7 @@ import scala.Tuple2;
  */
 public class HoodieReadClient<T extends HoodieRecordPayload> implements 
Serializable {
 
+  private static final long serialVersionUID = 1L;
   private static final Logger LOG = 
LogManager.getLogger(HoodieReadClient.class);
 
   /**
diff --git 
a/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java 
b/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java
index a0d1867..2a25a76 100644
--- a/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java
+++ b/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java
@@ -93,6 +93,7 @@ import scala.Tuple2;
  */
 public class HoodieWriteClient<T extends HoodieRecordPayload> extends 
AbstractHoodieWriteClient<T> {
 
+  private static final long serialVersionUID = 1L;
   private static final Logger LOG = 
LogManager.getLogger(HoodieWriteClient.class);
   private static final String LOOKUP_STR = "lookup";
   private final boolean rollbackPending;
diff --git a/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java 
b/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java
index 7dd89e0..a93f268 100644
--- a/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java
+++ b/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java
@@ -35,6 +35,7 @@ import java.util.Random;
  */
 public class WriteStatus implements Serializable {
 
+  private static final long serialVersionUID = 1L;
   private static final long RANDOM_SEED = 9038412832L;
 
   private final HashMap<HoodieKey, Throwable> errors = new HashMap<>();
diff --git 
a/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java 
b/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java
index 68ada22..709f08c 100644
--- 
a/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java
+++ 
b/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java
@@ -61,7 +61,7 @@ public class HoodieHBaseIndexConfig extends 
DefaultHoodieConfig {
    * value based on global indexing throughput needs and most importantly, how 
much the HBase installation in use is
    * able to tolerate without Region Servers going down.
    */
-  public static String HBASE_MAX_QPS_PER_REGION_SERVER_PROP = 
"hoodie.index.hbase.max.qps.per.region.server";
+  public static final String HBASE_MAX_QPS_PER_REGION_SERVER_PROP = 
"hoodie.index.hbase.max.qps.per.region.server";
   /**
    * Default batch size, used only for Get, but computed for Put.
    */
diff --git 
a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java
 
b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java
index 0472008..24da226 100644
--- 
a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java
+++ 
b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java
@@ -39,7 +39,7 @@ import java.util.stream.Collectors;
  */
 public class BoundedPartitionAwareCompactionStrategy extends 
DayBasedCompactionStrategy {
 
-  SimpleDateFormat dateFormat = new SimpleDateFormat(datePartitionFormat);
+  SimpleDateFormat dateFormat = new SimpleDateFormat(DATE_PARTITION_FORMAT);
 
   @Override
   public List<HoodieCompactionOperation> orderAndFilter(HoodieWriteConfig 
writeConfig,
diff --git 
a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java
 
b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java
index cf7727b..ea29fdd 100644
--- 
a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java
+++ 
b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java
@@ -39,14 +39,14 @@ import java.util.stream.Collectors;
 public class DayBasedCompactionStrategy extends CompactionStrategy {
 
   // For now, use SimpleDateFormat as default partition format
-  protected static String datePartitionFormat = "yyyy/MM/dd";
+  protected static final String DATE_PARTITION_FORMAT = "yyyy/MM/dd";
   // Sorts compaction in LastInFirstCompacted order
   protected static Comparator<String> comparator = (String leftPartition, 
String rightPartition) -> {
     try {
       leftPartition = getPartitionPathWithoutPartitionKeys(leftPartition);
       rightPartition = getPartitionPathWithoutPartitionKeys(rightPartition);
-      Date left = new SimpleDateFormat(datePartitionFormat, 
Locale.ENGLISH).parse(leftPartition);
-      Date right = new SimpleDateFormat(datePartitionFormat, 
Locale.ENGLISH).parse(rightPartition);
+      Date left = new SimpleDateFormat(DATE_PARTITION_FORMAT, 
Locale.ENGLISH).parse(leftPartition);
+      Date right = new SimpleDateFormat(DATE_PARTITION_FORMAT, 
Locale.ENGLISH).parse(rightPartition);
       return left.after(right) ? -1 : right.after(left) ? 1 : 0;
     } catch (ParseException e) {
       throw new HoodieException("Invalid Partition Date Format", e);
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java
 
b/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java
index 8ad82d4..408cb17 100644
--- 
a/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java
+++ 
b/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java
@@ -30,6 +30,7 @@ import java.io.Serializable;
  */
 public class SerializableConfiguration implements Serializable {
 
+  private static final long serialVersionUID = 1L;
   private transient Configuration configuration;
 
   public SerializableConfiguration(Configuration configuration) {
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java 
b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java
index ee20136..5ccc2a8 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java
@@ -31,6 +31,7 @@ import java.util.Objects;
  */
 public class HoodieBaseFile implements Serializable {
 
+  private static final long serialVersionUID = 1L;
   private transient FileStatus fileStatus;
   private final String fullPath;
   private long fileLen;
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java 
b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java
index a0a22aa..973fd51 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java
@@ -37,6 +37,7 @@ import java.util.Objects;
  */
 public class HoodieLogFile implements Serializable {
 
+  private static final long serialVersionUID = 1L;
   public static final String DELTA_EXTENSION = ".log";
   public static final Integer LOGFILE_BASE_VERSION = 1;
 
@@ -129,6 +130,7 @@ public class HoodieLogFile implements Serializable {
    */
   public static class LogFileComparator implements Comparator<HoodieLogFile>, 
Serializable {
 
+    private static final long serialVersionUID = 1L;
     private transient Comparator<String> writeTokenComparator;
 
     private Comparator<String> getWriteTokenComparator() {
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java 
b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java
index 57f5fa7..df7843f 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java
@@ -30,11 +30,11 @@ import java.util.Objects;
  */
 public class HoodieRecord<T extends HoodieRecordPayload> implements 
Serializable {
 
-  public static String COMMIT_TIME_METADATA_FIELD = "_hoodie_commit_time";
-  public static String COMMIT_SEQNO_METADATA_FIELD = "_hoodie_commit_seqno";
-  public static String RECORD_KEY_METADATA_FIELD = "_hoodie_record_key";
-  public static String PARTITION_PATH_METADATA_FIELD = 
"_hoodie_partition_path";
-  public static String FILENAME_METADATA_FIELD = "_hoodie_file_name";
+  public static final String COMMIT_TIME_METADATA_FIELD = 
"_hoodie_commit_time";
+  public static final String COMMIT_SEQNO_METADATA_FIELD = 
"_hoodie_commit_seqno";
+  public static final String RECORD_KEY_METADATA_FIELD = "_hoodie_record_key";
+  public static final String PARTITION_PATH_METADATA_FIELD = 
"_hoodie_partition_path";
+  public static final String FILENAME_METADATA_FIELD = "_hoodie_file_name";
 
   public static final List<String> HOODIE_META_COLUMNS =
       CollectionUtils.createImmutableList(COMMIT_TIME_METADATA_FIELD, 
COMMIT_SEQNO_METADATA_FIELD,
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java
 
b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java
index fc0ac19..a354092 100644
--- 
a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java
+++ 
b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java
@@ -52,7 +52,7 @@ public class HoodieRollingStatMetadata implements 
Serializable {
     this.actionType = actionType;
   }
 
-  class RollingStatsHashMap<K, V> extends HashMap<K, V> {
+  static class RollingStatsHashMap<K, V> extends HashMap<K, V> {
 
     @Override
     public V put(K key, V value) {
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java
 
b/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java
index 4ad8b51..03e91c2 100644
--- 
a/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java
+++ 
b/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java
@@ -68,10 +68,11 @@ import java.util.stream.Stream;
  */
 public class HoodieTableMetaClient implements Serializable {
 
+  private static final long serialVersionUID = 1L;
   private static final Logger LOG = 
LogManager.getLogger(HoodieTableMetaClient.class);
-  public static String METAFOLDER_NAME = ".hoodie";
-  public static String TEMPFOLDER_NAME = METAFOLDER_NAME + File.separator + 
".temp";
-  public static String AUXILIARYFOLDER_NAME = METAFOLDER_NAME + File.separator 
+ ".aux";
+  public static final String METAFOLDER_NAME = ".hoodie";
+  public static final String TEMPFOLDER_NAME = METAFOLDER_NAME + 
File.separator + ".temp";
+  public static final String AUXILIARYFOLDER_NAME = METAFOLDER_NAME + 
File.separator + ".aux";
   public static final String MARKER_EXTN = ".marker";
 
   private String basePath;
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java
 
b/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java
index 3f400be..5fa4a02 100644
--- 
a/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java
+++ 
b/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java
@@ -48,6 +48,7 @@ import static java.util.Collections.reverse;
  */
 public class HoodieDefaultTimeline implements HoodieTimeline {
 
+  private static final long serialVersionUID = 1L;
   private static final Logger LOG = 
LogManager.getLogger(HoodieDefaultTimeline.class);
 
   private static final String HASHING_ALGORITHM = "SHA-256";
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java 
b/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java
index 2b8fac8..0ccc7ca 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java
@@ -32,7 +32,7 @@ public class HoodieTimer {
   // Ordered stack of TimeInfo's to make sure stopping the timer returns the 
correct elapsed time
   Deque<TimeInfo> timeInfoDeque = new ArrayDeque<>();
 
-  class TimeInfo {
+  static class TimeInfo {
 
     // captures the startTime of the code block
     long startTime;
diff --git 
a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java
 
b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java
index 90977dd..f4b5924 100644
--- 
a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java
+++ 
b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java
@@ -55,7 +55,7 @@ import java.util.stream.Stream;
  */
 public final class DiskBasedMap<T extends Serializable, R extends 
Serializable> implements Map<T, R>, Iterable<R> {
 
-  public static int BUFFER_SIZE = 128 * 1024;  // 128 KB
+  public static final int BUFFER_SIZE = 128 * 1024;  // 128 KB
   private static final Logger LOG = LogManager.getLogger(DiskBasedMap.class);
   // Stores the key and corresponding value's latest metadata spilled to disk
   private final Map<T, ValueMetadata> valueMetadataMap;
diff --git 
a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java
 
b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java
index c0bc13b..4058875 100644
--- 
a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java
+++ 
b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java
@@ -51,6 +51,7 @@ import java.util.stream.Collectors;
  */
 public class HoodieROTablePathFilter implements PathFilter, Serializable {
 
+  private static final long serialVersionUID = 1L;
   private static final Logger LOG = 
LogManager.getLogger(HoodieROTablePathFilter.class);
 
   /**
diff --git 
a/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java
 
b/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java
index 4dba0fe..f4a4c90 100644
--- 
a/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java
+++ 
b/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java
@@ -33,6 +33,7 @@ import java.util.List;
  */
 public class SlashEncodedDayPartitionValueExtractor implements 
PartitionValueExtractor {
 
+  private static final long serialVersionUID = 1L;
   private transient DateTimeFormatter dtfOut;
 
   public SlashEncodedDayPartitionValueExtractor() {
diff --git 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java
 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java
index 0ed09cd..f389c58 100644
--- 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java
+++ 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java
@@ -66,6 +66,7 @@ import scala.Tuple2;
  */
 public class HDFSParquetImporter implements Serializable {
 
+  private static final long serialVersionUID = 1L;
   private static final Logger LOG = 
LogManager.getLogger(HDFSParquetImporter.class);
 
   private static final DateTimeFormatter PARTITION_FORMATTER = 
DateTimeFormatter.ofPattern("yyyy/MM/dd")
diff --git 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java
 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java
index eafe6c8..4c23537 100644
--- 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java
+++ 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java
@@ -37,6 +37,7 @@ import java.io.Serializable;
  */
 public class Compactor implements Serializable {
 
+  private static final long serialVersionUID = 1L;
   private static final Logger LOG = LogManager.getLogger(Compactor.class);
 
   private transient HoodieWriteClient compactionClient;
diff --git 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java
 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java
index a29db9b..99cb497 100644
--- 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java
+++ 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java
@@ -83,9 +83,10 @@ import static 
org.apache.hudi.utilities.schema.RowBasedSchemaProvider.HOODIE_REC
  */
 public class DeltaSync implements Serializable {
 
+  private static final long serialVersionUID = 1L;
   private static final Logger LOG = LogManager.getLogger(DeltaSync.class);
-  public static String CHECKPOINT_KEY = "deltastreamer.checkpoint.key";
-  public static String CHECKPOINT_RESET_KEY = 
"deltastreamer.checkpoint.reset_key";
+  public static final String CHECKPOINT_KEY = "deltastreamer.checkpoint.key";
+  public static final String CHECKPOINT_RESET_KEY = 
"deltastreamer.checkpoint.reset_key";
 
   /**
    * Delta Sync Config.
diff --git 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java
 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java
index 4fa43f4..948033f 100644
--- 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java
+++ 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java
@@ -77,6 +77,7 @@ import java.util.stream.IntStream;
  */
 public class HoodieDeltaStreamer implements Serializable {
 
+  private static final long serialVersionUID = 1L;
   private static final Logger LOG = 
LogManager.getLogger(HoodieDeltaStreamer.class);
 
   public static String CHECKPOINT_KEY = "deltastreamer.checkpoint.key";
@@ -302,6 +303,7 @@ public class HoodieDeltaStreamer implements Serializable {
    */
   public static class DeltaSyncService extends AbstractDeltaStreamerService {
 
+    private static final long serialVersionUID = 1L;
     /**
      * Delta Sync Config.
      */
@@ -489,6 +491,7 @@ public class HoodieDeltaStreamer implements Serializable {
    */
   public static class AsyncCompactService extends AbstractDeltaStreamerService 
{
 
+    private static final long serialVersionUID = 1L;
     private final int maxConcurrentCompaction;
     private transient Compactor compactor;
     private transient JavaSparkContext jssc;
diff --git 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java
 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java
index ca5ab24..a6692cc 100644
--- 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java
+++ 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java
@@ -58,6 +58,7 @@ import java.util.stream.IntStream;
 
 public class TimelineServerPerf implements Serializable {
 
+  private static final long serialVersionUID = 1L;
   private static final Logger LOG = 
LogManager.getLogger(TimelineServerPerf.class);
   private final Config cfg;
   private transient TimelineService timelineServer;
diff --git 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java
 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java
index 228d46d..3d158ba 100644
--- 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java
+++ 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java
@@ -58,10 +58,12 @@ import java.util.List;
  * passed to the CSV reader without inferring the schema from the CSV file.
  */
 public class CsvDFSSource extends RowSource {
+
+  private static final long serialVersionUID = 1L;
   // CsvSource config prefix
-  public static final String CSV_SRC_CONFIG_PREFIX = 
"hoodie.deltastreamer.csv.";
+  protected static final String CSV_SRC_CONFIG_PREFIX = 
"hoodie.deltastreamer.csv.";
   // CSV-specific configurations to pass in from Hudi to Spark
-  public static final List<String> CSV_CONFIG_KEYS = Arrays.asList(
+  protected static final List<String> CSV_CONFIG_KEYS = Arrays.asList(
       "sep", "encoding", "quote", "escape", "charToEscapeQuoteEscaping", 
"comment",
       "header", "enforceSchema", "inferSchema", "samplingRatio", 
"ignoreLeadingWhiteSpace",
       "ignoreTrailingWhiteSpace", "nullValue", "emptyValue", "nanValue", 
"positiveInf",
@@ -69,7 +71,7 @@ public class CsvDFSSource extends RowSource {
       "mode", "columnNameOfCorruptRecord", "multiLine"
   );
 
-  private final DFSPathSelector pathSelector;
+  private final transient DFSPathSelector pathSelector;
   private final StructType sourceSchema;
 
   public CsvDFSSource(TypedProperties props,
diff --git 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java
 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java
index 76f2187..a8bea98 100644
--- 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java
+++ 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java
@@ -59,6 +59,8 @@ import java.util.stream.Collectors;
  */
 public class HiveIncrPullSource extends AvroSource {
 
+  private static final long serialVersionUID = 1L;
+
   private static final Logger LOG = 
LogManager.getLogger(HiveIncrPullSource.class);
 
   private final transient FileSystem fs;
diff --git 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java
 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java
index 988dbd8..449db10 100644
--- 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java
+++ 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java
@@ -33,6 +33,7 @@ import java.io.Serializable;
  */
 public class AvroConvertor implements Serializable {
 
+  private static final long serialVersionUID = 1L;
   /**
    * To be lazily inited on executors.
    */
diff --git 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java
 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java
index 665720b..cb0100a 100644
--- 
a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java
+++ 
b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java
@@ -85,7 +85,7 @@ public class DFSPathSelector {
       long maxModificationTime = Long.MIN_VALUE;
       List<FileStatus> filteredFiles = new ArrayList<>();
       for (FileStatus f : eligibleFiles) {
-        if (lastCheckpointStr.isPresent() && f.getModificationTime() <= 
Long.valueOf(lastCheckpointStr.get())) {
+        if (lastCheckpointStr.isPresent() && f.getModificationTime() <= 
Long.valueOf(lastCheckpointStr.get()).longValue()) {
           // skip processed files
           continue;
         }

Reply via email to