This is an automated email from the ASF dual-hosted git repository.

vinoth pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/master by this push:
     new 68a656b  [HUDI-1032] Remove unused code in HoodieCopyOnWriteTable and 
code clean (#1750)
68a656b is described below

commit 68a656b01650cfcd8c099dd04cea59fb083dca69
Author: wangxianghu <[email protected]>
AuthorDate: Sun Jun 21 22:34:47 2020 +0800

    [HUDI-1032] Remove unused code in HoodieCopyOnWriteTable and code clean 
(#1750)
---
 .../apache/hudi/table/HoodieCopyOnWriteTable.java  | 65 +---------------------
 .../apache/hudi/table/HoodieMergeOnReadTable.java  |  5 +-
 2 files changed, 2 insertions(+), 68 deletions(-)

diff --git 
a/hudi-client/src/main/java/org/apache/hudi/table/HoodieCopyOnWriteTable.java 
b/hudi-client/src/main/java/org/apache/hudi/table/HoodieCopyOnWriteTable.java
index 974d847..64b3bf0 100644
--- 
a/hudi-client/src/main/java/org/apache/hudi/table/HoodieCopyOnWriteTable.java
+++ 
b/hudi-client/src/main/java/org/apache/hudi/table/HoodieCopyOnWriteTable.java
@@ -31,7 +31,6 @@ import org.apache.hudi.client.utils.ParquetReaderIterator;
 import org.apache.hudi.common.model.HoodieBaseFile;
 import org.apache.hudi.common.model.HoodieKey;
 import org.apache.hudi.common.model.HoodieRecord;
-import org.apache.hudi.common.model.HoodieRecordLocation;
 import org.apache.hudi.common.model.HoodieRecordPayload;
 import org.apache.hudi.common.table.HoodieTableMetaClient;
 import org.apache.hudi.common.table.timeline.HoodieInstant;
@@ -66,7 +65,6 @@ import org.apache.spark.api.java.JavaRDD;
 import org.apache.spark.api.java.JavaSparkContext;
 
 import java.io.IOException;
-import java.io.Serializable;
 import java.util.Collections;
 import java.util.Iterator;
 import java.util.List;
@@ -75,7 +73,6 @@ import java.util.Map;
 /**
  * Implementation of a very heavily read-optimized Hoodie Table where, all 
data is stored in base files, with
  * zero read amplification.
- *
  * <p>
  * INSERTS - Produce new files, block aligned to desired size (or) Merge with 
the smallest existing file, to expand it
  * <p>
@@ -207,14 +204,11 @@ public class HoodieCopyOnWriteTable<T extends 
HoodieRecordPayload> extends Hoodi
     return new SavepointActionExecutor(jsc, config, this, instantToSavepoint, 
user, comment).execute();
   }
 
+  @Override
   public HoodieRestoreMetadata restore(JavaSparkContext jsc, String 
restoreInstantTime, String instantToRestore) {
     return new CopyOnWriteRestoreActionExecutor(jsc, config, this, 
restoreInstantTime, instantToRestore).execute();
   }
 
-  enum BucketType {
-    UPDATE, INSERT
-  }
-
   /**
    * Consumer that dequeues records from queue and sends to Merge Handle.
    */
@@ -240,61 +234,4 @@ public class HoodieCopyOnWriteTable<T extends 
HoodieRecordPayload> extends Hoodi
     }
   }
 
-  /**
-   * Helper class for a small file's location and its actual size on disk.
-   */
-  static class SmallFile implements Serializable {
-
-    HoodieRecordLocation location;
-    long sizeBytes;
-
-    @Override
-    public String toString() {
-      final StringBuilder sb = new StringBuilder("SmallFile {");
-      sb.append("location=").append(location).append(", ");
-      sb.append("sizeBytes=").append(sizeBytes);
-      sb.append('}');
-      return sb.toString();
-    }
-  }
-
-  /**
-   * Helper class for an insert bucket along with the weight [0.0, 1.0] that 
defines the amount of incoming inserts that
-   * should be allocated to the bucket.
-   */
-  class InsertBucket implements Serializable {
-
-    int bucketNumber;
-    // fraction of total inserts, that should go into this bucket
-    double weight;
-
-    @Override
-    public String toString() {
-      final StringBuilder sb = new StringBuilder("WorkloadStat {");
-      sb.append("bucketNumber=").append(bucketNumber).append(", ");
-      sb.append("weight=").append(weight);
-      sb.append('}');
-      return sb.toString();
-    }
-  }
-
-  /**
-   * Helper class for a bucket's type (INSERT and UPDATE) and its file 
location.
-   */
-  class BucketInfo implements Serializable {
-
-    BucketType bucketType;
-    String fileIdPrefix;
-    String partitionPath;
-
-    @Override
-    public String toString() {
-      final StringBuilder sb = new StringBuilder("BucketInfo {");
-      sb.append("bucketType=").append(bucketType).append(", ");
-      sb.append("fileIdPrefix=").append(fileIdPrefix).append(", ");
-      sb.append("partitionPath=").append(partitionPath);
-      sb.append('}');
-      return sb.toString();
-    }
-  }
 }
diff --git 
a/hudi-client/src/main/java/org/apache/hudi/table/HoodieMergeOnReadTable.java 
b/hudi-client/src/main/java/org/apache/hudi/table/HoodieMergeOnReadTable.java
index 65981c2..8496ea4 100644
--- 
a/hudi-client/src/main/java/org/apache/hudi/table/HoodieMergeOnReadTable.java
+++ 
b/hudi-client/src/main/java/org/apache/hudi/table/HoodieMergeOnReadTable.java
@@ -43,8 +43,6 @@ import 
org.apache.hudi.table.action.deltacommit.UpsertPreppedDeltaCommitActionEx
 import org.apache.hudi.table.action.compact.ScheduleCompactionActionExecutor;
 import org.apache.hudi.table.action.restore.MergeOnReadRestoreActionExecutor;
 import org.apache.hudi.table.action.rollback.MergeOnReadRollbackActionExecutor;
-import org.apache.log4j.LogManager;
-import org.apache.log4j.Logger;
 import org.apache.spark.api.java.JavaRDD;
 import org.apache.spark.api.java.JavaSparkContext;
 
@@ -69,8 +67,6 @@ import java.util.Map;
  */
 public class HoodieMergeOnReadTable<T extends HoodieRecordPayload> extends 
HoodieCopyOnWriteTable<T> {
 
-  private static final Logger LOG = 
LogManager.getLogger(HoodieMergeOnReadTable.class);
-
   HoodieMergeOnReadTable(HoodieWriteConfig config, Configuration hadoopConf, 
HoodieTableMetaClient metaClient) {
     super(config, hadoopConf, metaClient);
   }
@@ -137,6 +133,7 @@ public class HoodieMergeOnReadTable<T extends 
HoodieRecordPayload> extends Hoodi
     return new MergeOnReadRollbackActionExecutor(jsc, config, this, 
rollbackInstantTime, commitInstant, deleteInstants).execute();
   }
 
+  @Override
   public HoodieRestoreMetadata restore(JavaSparkContext jsc, String 
restoreInstantTime, String instantToRestore) {
     return new MergeOnReadRestoreActionExecutor(jsc, config, this, 
restoreInstantTime, instantToRestore).execute();
   }

Reply via email to