This is an automated email from the ASF dual-hosted git repository. vinoyang pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/incubator-hudi.git
The following commit(s) were added to refs/heads/master by this push: new 78b3194 [HUDI-751] Fix some coding issues reported by FindBugs (#1470) 78b3194 is described below commit 78b3194e8241c519a85310997f31b2b55df487e1 Author: Shaofeng Shi <shaofeng...@gmail.com> AuthorDate: Tue Mar 31 21:19:32 2020 +0800 [HUDI-751] Fix some coding issues reported by FindBugs (#1470) --- .../org/apache/hudi/cli/commands/RollbacksCommand.java | 2 +- .../org/apache/hudi/cli/commands/SparkEnvCommand.java | 6 +++--- .../java/org/apache/hudi/cli/commands/StatsCommand.java | 6 +++--- .../main/java/org/apache/hudi/cli/utils/HiveUtil.java | 16 ++++------------ .../java/org/apache/hudi/client/HoodieCleanClient.java | 1 + .../java/org/apache/hudi/client/HoodieReadClient.java | 1 + .../java/org/apache/hudi/client/HoodieWriteClient.java | 1 + .../main/java/org/apache/hudi/client/WriteStatus.java | 1 + .../org/apache/hudi/config/HoodieHBaseIndexConfig.java | 2 +- .../BoundedPartitionAwareCompactionStrategy.java | 2 +- .../compact/strategy/DayBasedCompactionStrategy.java | 6 +++--- .../hudi/common/config/SerializableConfiguration.java | 1 + .../org/apache/hudi/common/model/HoodieBaseFile.java | 1 + .../java/org/apache/hudi/common/model/HoodieLogFile.java | 2 ++ .../java/org/apache/hudi/common/model/HoodieRecord.java | 10 +++++----- .../hudi/common/model/HoodieRollingStatMetadata.java | 2 +- .../apache/hudi/common/table/HoodieTableMetaClient.java | 7 ++++--- .../common/table/timeline/HoodieDefaultTimeline.java | 1 + .../java/org/apache/hudi/common/util/HoodieTimer.java | 2 +- .../apache/hudi/common/util/collection/DiskBasedMap.java | 2 +- .../org/apache/hudi/hadoop/HoodieROTablePathFilter.java | 1 + .../hive/SlashEncodedDayPartitionValueExtractor.java | 1 + .../org/apache/hudi/utilities/HDFSParquetImporter.java | 1 + .../apache/hudi/utilities/deltastreamer/Compactor.java | 1 + .../apache/hudi/utilities/deltastreamer/DeltaSync.java | 5 +++-- .../utilities/deltastreamer/HoodieDeltaStreamer.java | 3 +++ .../apache/hudi/utilities/perf/TimelineServerPerf.java | 1 + .../org/apache/hudi/utilities/sources/CsvDFSSource.java | 8 +++++--- .../hudi/utilities/sources/HiveIncrPullSource.java | 2 ++ .../hudi/utilities/sources/helpers/AvroConvertor.java | 1 + .../hudi/utilities/sources/helpers/DFSPathSelector.java | 2 +- 31 files changed, 57 insertions(+), 41 deletions(-) diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java index 9e4bf28..70b34bc 100644 --- a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java +++ b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/RollbacksCommand.java @@ -120,7 +120,7 @@ public class RollbacksCommand implements CommandMarker { /** * An Active timeline containing only rollbacks. */ - class RollbackTimeline extends HoodieActiveTimeline { + static class RollbackTimeline extends HoodieActiveTimeline { public RollbackTimeline(HoodieTableMetaClient metaClient) { super(metaClient, CollectionUtils.createImmutableSet(HoodieTimeline.ROLLBACK_EXTENSION)); diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java index d209a08..7969808 100644 --- a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java +++ b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/SparkEnvCommand.java @@ -34,7 +34,7 @@ import java.util.Map; @Component public class SparkEnvCommand implements CommandMarker { - public static Map<String, String> env = new HashMap<String, String>(); + public static Map<String, String> env = new HashMap<>(); @CliCommand(value = "set", help = "Set spark launcher env to cli") public void setEnv(@CliOption(key = {"conf"}, help = "Env config to be set") final String confMap) { @@ -49,8 +49,8 @@ public class SparkEnvCommand implements CommandMarker { public String showAllEnv() { String[][] rows = new String[env.size()][2]; int i = 0; - for (String key: env.keySet()) { - rows[i] = new String[]{key, env.get(key)}; + for (Map.Entry<String, String> entry: env.entrySet()) { + rows[i] = new String[]{entry.getKey(), entry.getValue()}; i++; } return HoodiePrintHelper.print(new String[] {"key", "value"}, rows); diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java index 9db544c..e5be0e4 100644 --- a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java +++ b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/StatsCommand.java @@ -131,9 +131,9 @@ public class StatsCommand implements CommandMarker { } List<Comparable[]> rows = new ArrayList<>(); - for (String instantTime : commitHistoMap.keySet()) { - Snapshot s = commitHistoMap.get(instantTime).getSnapshot(); - rows.add(printFileSizeHistogram(instantTime, s)); + for (Map.Entry<String, Histogram> entry : commitHistoMap.entrySet()) { + Snapshot s = entry.getValue().getSnapshot(); + rows.add(printFileSizeHistogram(entry.getKey(), s)); } Snapshot s = globalHistogram.getSnapshot(); rows.add(printFileSizeHistogram("ALL", s)); diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java b/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java index 291ac74..37f705e 100644 --- a/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java +++ b/hudi-cli/src/main/java/org/apache/hudi/cli/utils/HiveUtil.java @@ -49,10 +49,9 @@ public class HiveUtil { public static long countRecords(String jdbcUrl, HoodieTableMetaClient source, String dbName, String user, String pass) throws SQLException { - Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass); ResultSet rs = null; - Statement stmt = conn.createStatement(); - try { + try (Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass); + Statement stmt = conn.createStatement()) { // stmt.execute("set mapred.job.queue.name=<queue_name>"); stmt.execute("set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat"); stmt.execute("set hive.stats.autogather=false"); @@ -68,9 +67,6 @@ public class HiveUtil { if (rs != null) { rs.close(); } - if (stmt != null) { - stmt.close(); - } } } @@ -88,10 +84,9 @@ public class HiveUtil { private static long countRecords(String jdbcUrl, HoodieTableMetaClient source, String srcDb, String startDateStr, String endDateStr, String user, String pass) throws SQLException { - Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass); ResultSet rs = null; - Statement stmt = conn.createStatement(); - try { + try (Connection conn = HiveUtil.getConnection(jdbcUrl, user, pass); + Statement stmt = conn.createStatement()) { // stmt.execute("set mapred.job.queue.name=<queue_name>"); stmt.execute("set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat"); stmt.execute("set hive.stats.autogather=false"); @@ -106,9 +101,6 @@ public class HiveUtil { if (rs != null) { rs.close(); } - if (stmt != null) { - stmt.close(); - } } } } diff --git a/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java b/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java index 43a3655..d622f70 100644 --- a/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java +++ b/hudi-client/src/main/java/org/apache/hudi/client/HoodieCleanClient.java @@ -47,6 +47,7 @@ import java.util.List; public class HoodieCleanClient<T extends HoodieRecordPayload> extends AbstractHoodieClient { + private static final long serialVersionUID = 1L; private static final Logger LOG = LogManager.getLogger(HoodieCleanClient.class); private final transient HoodieMetrics metrics; diff --git a/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java b/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java index d1e92b5..2dedfaf 100644 --- a/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java +++ b/hudi-client/src/main/java/org/apache/hudi/client/HoodieReadClient.java @@ -57,6 +57,7 @@ import scala.Tuple2; */ public class HoodieReadClient<T extends HoodieRecordPayload> implements Serializable { + private static final long serialVersionUID = 1L; private static final Logger LOG = LogManager.getLogger(HoodieReadClient.class); /** diff --git a/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java b/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java index a0d1867..2a25a76 100644 --- a/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java +++ b/hudi-client/src/main/java/org/apache/hudi/client/HoodieWriteClient.java @@ -93,6 +93,7 @@ import scala.Tuple2; */ public class HoodieWriteClient<T extends HoodieRecordPayload> extends AbstractHoodieWriteClient<T> { + private static final long serialVersionUID = 1L; private static final Logger LOG = LogManager.getLogger(HoodieWriteClient.class); private static final String LOOKUP_STR = "lookup"; private final boolean rollbackPending; diff --git a/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java b/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java index 7dd89e0..a93f268 100644 --- a/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java +++ b/hudi-client/src/main/java/org/apache/hudi/client/WriteStatus.java @@ -35,6 +35,7 @@ import java.util.Random; */ public class WriteStatus implements Serializable { + private static final long serialVersionUID = 1L; private static final long RANDOM_SEED = 9038412832L; private final HashMap<HoodieKey, Throwable> errors = new HashMap<>(); diff --git a/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java b/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java index 68ada22..709f08c 100644 --- a/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java +++ b/hudi-client/src/main/java/org/apache/hudi/config/HoodieHBaseIndexConfig.java @@ -61,7 +61,7 @@ public class HoodieHBaseIndexConfig extends DefaultHoodieConfig { * value based on global indexing throughput needs and most importantly, how much the HBase installation in use is * able to tolerate without Region Servers going down. */ - public static String HBASE_MAX_QPS_PER_REGION_SERVER_PROP = "hoodie.index.hbase.max.qps.per.region.server"; + public static final String HBASE_MAX_QPS_PER_REGION_SERVER_PROP = "hoodie.index.hbase.max.qps.per.region.server"; /** * Default batch size, used only for Get, but computed for Put. */ diff --git a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java index 0472008..24da226 100644 --- a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java +++ b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/BoundedPartitionAwareCompactionStrategy.java @@ -39,7 +39,7 @@ import java.util.stream.Collectors; */ public class BoundedPartitionAwareCompactionStrategy extends DayBasedCompactionStrategy { - SimpleDateFormat dateFormat = new SimpleDateFormat(datePartitionFormat); + SimpleDateFormat dateFormat = new SimpleDateFormat(DATE_PARTITION_FORMAT); @Override public List<HoodieCompactionOperation> orderAndFilter(HoodieWriteConfig writeConfig, diff --git a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java index cf7727b..ea29fdd 100644 --- a/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java +++ b/hudi-client/src/main/java/org/apache/hudi/table/compact/strategy/DayBasedCompactionStrategy.java @@ -39,14 +39,14 @@ import java.util.stream.Collectors; public class DayBasedCompactionStrategy extends CompactionStrategy { // For now, use SimpleDateFormat as default partition format - protected static String datePartitionFormat = "yyyy/MM/dd"; + protected static final String DATE_PARTITION_FORMAT = "yyyy/MM/dd"; // Sorts compaction in LastInFirstCompacted order protected static Comparator<String> comparator = (String leftPartition, String rightPartition) -> { try { leftPartition = getPartitionPathWithoutPartitionKeys(leftPartition); rightPartition = getPartitionPathWithoutPartitionKeys(rightPartition); - Date left = new SimpleDateFormat(datePartitionFormat, Locale.ENGLISH).parse(leftPartition); - Date right = new SimpleDateFormat(datePartitionFormat, Locale.ENGLISH).parse(rightPartition); + Date left = new SimpleDateFormat(DATE_PARTITION_FORMAT, Locale.ENGLISH).parse(leftPartition); + Date right = new SimpleDateFormat(DATE_PARTITION_FORMAT, Locale.ENGLISH).parse(rightPartition); return left.after(right) ? -1 : right.after(left) ? 1 : 0; } catch (ParseException e) { throw new HoodieException("Invalid Partition Date Format", e); diff --git a/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java b/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java index 8ad82d4..408cb17 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/config/SerializableConfiguration.java @@ -30,6 +30,7 @@ import java.io.Serializable; */ public class SerializableConfiguration implements Serializable { + private static final long serialVersionUID = 1L; private transient Configuration configuration; public SerializableConfiguration(Configuration configuration) { diff --git a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java index ee20136..5ccc2a8 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieBaseFile.java @@ -31,6 +31,7 @@ import java.util.Objects; */ public class HoodieBaseFile implements Serializable { + private static final long serialVersionUID = 1L; private transient FileStatus fileStatus; private final String fullPath; private long fileLen; diff --git a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java index a0a22aa..973fd51 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieLogFile.java @@ -37,6 +37,7 @@ import java.util.Objects; */ public class HoodieLogFile implements Serializable { + private static final long serialVersionUID = 1L; public static final String DELTA_EXTENSION = ".log"; public static final Integer LOGFILE_BASE_VERSION = 1; @@ -129,6 +130,7 @@ public class HoodieLogFile implements Serializable { */ public static class LogFileComparator implements Comparator<HoodieLogFile>, Serializable { + private static final long serialVersionUID = 1L; private transient Comparator<String> writeTokenComparator; private Comparator<String> getWriteTokenComparator() { diff --git a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java index 57f5fa7..df7843f 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRecord.java @@ -30,11 +30,11 @@ import java.util.Objects; */ public class HoodieRecord<T extends HoodieRecordPayload> implements Serializable { - public static String COMMIT_TIME_METADATA_FIELD = "_hoodie_commit_time"; - public static String COMMIT_SEQNO_METADATA_FIELD = "_hoodie_commit_seqno"; - public static String RECORD_KEY_METADATA_FIELD = "_hoodie_record_key"; - public static String PARTITION_PATH_METADATA_FIELD = "_hoodie_partition_path"; - public static String FILENAME_METADATA_FIELD = "_hoodie_file_name"; + public static final String COMMIT_TIME_METADATA_FIELD = "_hoodie_commit_time"; + public static final String COMMIT_SEQNO_METADATA_FIELD = "_hoodie_commit_seqno"; + public static final String RECORD_KEY_METADATA_FIELD = "_hoodie_record_key"; + public static final String PARTITION_PATH_METADATA_FIELD = "_hoodie_partition_path"; + public static final String FILENAME_METADATA_FIELD = "_hoodie_file_name"; public static final List<String> HOODIE_META_COLUMNS = CollectionUtils.createImmutableList(COMMIT_TIME_METADATA_FIELD, COMMIT_SEQNO_METADATA_FIELD, diff --git a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java index fc0ac19..a354092 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/model/HoodieRollingStatMetadata.java @@ -52,7 +52,7 @@ public class HoodieRollingStatMetadata implements Serializable { this.actionType = actionType; } - class RollingStatsHashMap<K, V> extends HashMap<K, V> { + static class RollingStatsHashMap<K, V> extends HashMap<K, V> { @Override public V put(K key, V value) { diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java b/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java index 4ad8b51..03e91c2 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/HoodieTableMetaClient.java @@ -68,10 +68,11 @@ import java.util.stream.Stream; */ public class HoodieTableMetaClient implements Serializable { + private static final long serialVersionUID = 1L; private static final Logger LOG = LogManager.getLogger(HoodieTableMetaClient.class); - public static String METAFOLDER_NAME = ".hoodie"; - public static String TEMPFOLDER_NAME = METAFOLDER_NAME + File.separator + ".temp"; - public static String AUXILIARYFOLDER_NAME = METAFOLDER_NAME + File.separator + ".aux"; + public static final String METAFOLDER_NAME = ".hoodie"; + public static final String TEMPFOLDER_NAME = METAFOLDER_NAME + File.separator + ".temp"; + public static final String AUXILIARYFOLDER_NAME = METAFOLDER_NAME + File.separator + ".aux"; public static final String MARKER_EXTN = ".marker"; private String basePath; diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java b/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java index 3f400be..5fa4a02 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieDefaultTimeline.java @@ -48,6 +48,7 @@ import static java.util.Collections.reverse; */ public class HoodieDefaultTimeline implements HoodieTimeline { + private static final long serialVersionUID = 1L; private static final Logger LOG = LogManager.getLogger(HoodieDefaultTimeline.class); private static final String HASHING_ALGORITHM = "SHA-256"; diff --git a/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java b/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java index 2b8fac8..0ccc7ca 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/util/HoodieTimer.java @@ -32,7 +32,7 @@ public class HoodieTimer { // Ordered stack of TimeInfo's to make sure stopping the timer returns the correct elapsed time Deque<TimeInfo> timeInfoDeque = new ArrayDeque<>(); - class TimeInfo { + static class TimeInfo { // captures the startTime of the code block long startTime; diff --git a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java index 90977dd..f4b5924 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java @@ -55,7 +55,7 @@ import java.util.stream.Stream; */ public final class DiskBasedMap<T extends Serializable, R extends Serializable> implements Map<T, R>, Iterable<R> { - public static int BUFFER_SIZE = 128 * 1024; // 128 KB + public static final int BUFFER_SIZE = 128 * 1024; // 128 KB private static final Logger LOG = LogManager.getLogger(DiskBasedMap.class); // Stores the key and corresponding value's latest metadata spilled to disk private final Map<T, ValueMetadata> valueMetadataMap; diff --git a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java index c0bc13b..4058875 100644 --- a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java +++ b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieROTablePathFilter.java @@ -51,6 +51,7 @@ import java.util.stream.Collectors; */ public class HoodieROTablePathFilter implements PathFilter, Serializable { + private static final long serialVersionUID = 1L; private static final Logger LOG = LogManager.getLogger(HoodieROTablePathFilter.class); /** diff --git a/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java b/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java index 4dba0fe..f4a4c90 100644 --- a/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java +++ b/hudi-hive-sync/src/main/java/org/apache/hudi/hive/SlashEncodedDayPartitionValueExtractor.java @@ -33,6 +33,7 @@ import java.util.List; */ public class SlashEncodedDayPartitionValueExtractor implements PartitionValueExtractor { + private static final long serialVersionUID = 1L; private transient DateTimeFormatter dtfOut; public SlashEncodedDayPartitionValueExtractor() { diff --git a/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java b/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java index 0ed09cd..f389c58 100644 --- a/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java +++ b/hudi-utilities/src/main/java/org/apache/hudi/utilities/HDFSParquetImporter.java @@ -66,6 +66,7 @@ import scala.Tuple2; */ public class HDFSParquetImporter implements Serializable { + private static final long serialVersionUID = 1L; private static final Logger LOG = LogManager.getLogger(HDFSParquetImporter.class); private static final DateTimeFormatter PARTITION_FORMATTER = DateTimeFormatter.ofPattern("yyyy/MM/dd") diff --git a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java index eafe6c8..4c23537 100644 --- a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java +++ b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/Compactor.java @@ -37,6 +37,7 @@ import java.io.Serializable; */ public class Compactor implements Serializable { + private static final long serialVersionUID = 1L; private static final Logger LOG = LogManager.getLogger(Compactor.class); private transient HoodieWriteClient compactionClient; diff --git a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java index a29db9b..99cb497 100644 --- a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java +++ b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/DeltaSync.java @@ -83,9 +83,10 @@ import static org.apache.hudi.utilities.schema.RowBasedSchemaProvider.HOODIE_REC */ public class DeltaSync implements Serializable { + private static final long serialVersionUID = 1L; private static final Logger LOG = LogManager.getLogger(DeltaSync.class); - public static String CHECKPOINT_KEY = "deltastreamer.checkpoint.key"; - public static String CHECKPOINT_RESET_KEY = "deltastreamer.checkpoint.reset_key"; + public static final String CHECKPOINT_KEY = "deltastreamer.checkpoint.key"; + public static final String CHECKPOINT_RESET_KEY = "deltastreamer.checkpoint.reset_key"; /** * Delta Sync Config. diff --git a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java index 4fa43f4..948033f 100644 --- a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java +++ b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java @@ -77,6 +77,7 @@ import java.util.stream.IntStream; */ public class HoodieDeltaStreamer implements Serializable { + private static final long serialVersionUID = 1L; private static final Logger LOG = LogManager.getLogger(HoodieDeltaStreamer.class); public static String CHECKPOINT_KEY = "deltastreamer.checkpoint.key"; @@ -302,6 +303,7 @@ public class HoodieDeltaStreamer implements Serializable { */ public static class DeltaSyncService extends AbstractDeltaStreamerService { + private static final long serialVersionUID = 1L; /** * Delta Sync Config. */ @@ -489,6 +491,7 @@ public class HoodieDeltaStreamer implements Serializable { */ public static class AsyncCompactService extends AbstractDeltaStreamerService { + private static final long serialVersionUID = 1L; private final int maxConcurrentCompaction; private transient Compactor compactor; private transient JavaSparkContext jssc; diff --git a/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java b/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java index ca5ab24..a6692cc 100644 --- a/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java +++ b/hudi-utilities/src/main/java/org/apache/hudi/utilities/perf/TimelineServerPerf.java @@ -58,6 +58,7 @@ import java.util.stream.IntStream; public class TimelineServerPerf implements Serializable { + private static final long serialVersionUID = 1L; private static final Logger LOG = LogManager.getLogger(TimelineServerPerf.class); private final Config cfg; private transient TimelineService timelineServer; diff --git a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java index 228d46d..3d158ba 100644 --- a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java +++ b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/CsvDFSSource.java @@ -58,10 +58,12 @@ import java.util.List; * passed to the CSV reader without inferring the schema from the CSV file. */ public class CsvDFSSource extends RowSource { + + private static final long serialVersionUID = 1L; // CsvSource config prefix - public static final String CSV_SRC_CONFIG_PREFIX = "hoodie.deltastreamer.csv."; + protected static final String CSV_SRC_CONFIG_PREFIX = "hoodie.deltastreamer.csv."; // CSV-specific configurations to pass in from Hudi to Spark - public static final List<String> CSV_CONFIG_KEYS = Arrays.asList( + protected static final List<String> CSV_CONFIG_KEYS = Arrays.asList( "sep", "encoding", "quote", "escape", "charToEscapeQuoteEscaping", "comment", "header", "enforceSchema", "inferSchema", "samplingRatio", "ignoreLeadingWhiteSpace", "ignoreTrailingWhiteSpace", "nullValue", "emptyValue", "nanValue", "positiveInf", @@ -69,7 +71,7 @@ public class CsvDFSSource extends RowSource { "mode", "columnNameOfCorruptRecord", "multiLine" ); - private final DFSPathSelector pathSelector; + private final transient DFSPathSelector pathSelector; private final StructType sourceSchema; public CsvDFSSource(TypedProperties props, diff --git a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java index 76f2187..a8bea98 100644 --- a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java +++ b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/HiveIncrPullSource.java @@ -59,6 +59,8 @@ import java.util.stream.Collectors; */ public class HiveIncrPullSource extends AvroSource { + private static final long serialVersionUID = 1L; + private static final Logger LOG = LogManager.getLogger(HiveIncrPullSource.class); private final transient FileSystem fs; diff --git a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java index 988dbd8..449db10 100644 --- a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java +++ b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/AvroConvertor.java @@ -33,6 +33,7 @@ import java.io.Serializable; */ public class AvroConvertor implements Serializable { + private static final long serialVersionUID = 1L; /** * To be lazily inited on executors. */ diff --git a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java index 665720b..cb0100a 100644 --- a/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java +++ b/hudi-utilities/src/main/java/org/apache/hudi/utilities/sources/helpers/DFSPathSelector.java @@ -85,7 +85,7 @@ public class DFSPathSelector { long maxModificationTime = Long.MIN_VALUE; List<FileStatus> filteredFiles = new ArrayList<>(); for (FileStatus f : eligibleFiles) { - if (lastCheckpointStr.isPresent() && f.getModificationTime() <= Long.valueOf(lastCheckpointStr.get())) { + if (lastCheckpointStr.isPresent() && f.getModificationTime() <= Long.valueOf(lastCheckpointStr.get()).longValue()) { // skip processed files continue; }