This is an automated email from the ASF dual-hosted git repository.
zihanli58 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/gobblin.git
The following commit(s) were added to refs/heads/master by this push:
new 501e0e0 do not delete data while dropping a hive table because data
is deleted, if needed, separately (#3431)
501e0e0 is described below
commit 501e0e09e520978b3bbb9f2d07b07fc611f7dffb
Author: Arjun Singh Bora <[email protected]>
AuthorDate: Thu Nov 18 11:17:28 2021 -0800
do not delete data while dropping a hive table because data is deleted, if
needed, separately (#3431)
---
.../hive/metastore/HiveMetaStoreBasedRegister.java | 24 +++++++++++-----------
1 file changed, 12 insertions(+), 12 deletions(-)
diff --git
a/gobblin-hive-registration/src/main/java/org/apache/gobblin/hive/metastore/HiveMetaStoreBasedRegister.java
b/gobblin-hive-registration/src/main/java/org/apache/gobblin/hive/metastore/HiveMetaStoreBasedRegister.java
index 3d61634..020eba7 100644
---
a/gobblin-hive-registration/src/main/java/org/apache/gobblin/hive/metastore/HiveMetaStoreBasedRegister.java
+++
b/gobblin-hive-registration/src/main/java/org/apache/gobblin/hive/metastore/HiveMetaStoreBasedRegister.java
@@ -17,25 +17,15 @@
package org.apache.gobblin.hive.metastore;
-import com.google.common.annotations.VisibleForTesting;
-import com.google.common.cache.Cache;
-import com.google.common.cache.CacheBuilder;
-import com.google.common.cache.CacheLoader;
import java.io.IOException;
import java.util.Arrays;
import java.util.List;
-
import java.util.Map;
import java.util.concurrent.Callable;
import java.util.concurrent.ExecutionException;
-import lombok.extern.slf4j.Slf4j;
import org.apache.avro.Schema;
import org.apache.commons.pool2.impl.GenericObjectPoolConfig;
-import org.apache.gobblin.hive.AutoCloseableHiveLock;
-import org.apache.gobblin.metrics.kafka.KafkaSchemaRegistry;
-import org.apache.gobblin.source.extractor.extract.kafka.KafkaSource;
-import org.apache.gobblin.util.AvroUtils;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.hive.metastore.IMetaStoreClient;
import org.apache.hadoop.hive.metastore.TableType;
@@ -49,14 +39,21 @@ import org.apache.thrift.TException;
import org.joda.time.DateTime;
import com.codahale.metrics.Timer;
+import com.google.common.annotations.VisibleForTesting;
import com.google.common.base.Optional;
import com.google.common.base.Preconditions;
+import com.google.common.cache.Cache;
+import com.google.common.cache.CacheBuilder;
+import com.google.common.cache.CacheLoader;
import com.google.common.primitives.Ints;
+import lombok.extern.slf4j.Slf4j;
+
import org.apache.gobblin.annotation.Alpha;
import org.apache.gobblin.configuration.State;
-import org.apache.gobblin.hive.HiveMetaStoreClientFactory;
+import org.apache.gobblin.hive.AutoCloseableHiveLock;
import org.apache.gobblin.hive.HiveLock;
+import org.apache.gobblin.hive.HiveMetaStoreClientFactory;
import org.apache.gobblin.hive.HiveMetastoreClientPool;
import org.apache.gobblin.hive.HivePartition;
import org.apache.gobblin.hive.HiveRegProps;
@@ -68,7 +65,10 @@ import org.apache.gobblin.metrics.GobblinMetrics;
import org.apache.gobblin.metrics.GobblinMetricsRegistry;
import org.apache.gobblin.metrics.MetricContext;
import org.apache.gobblin.metrics.event.EventSubmitter;
+import org.apache.gobblin.metrics.kafka.KafkaSchemaRegistry;
+import org.apache.gobblin.source.extractor.extract.kafka.KafkaSource;
import org.apache.gobblin.util.AutoReturnableObject;
+import org.apache.gobblin.util.AvroUtils;
/**
@@ -525,7 +525,7 @@ public class HiveMetaStoreBasedRegister extends
HiveRegister {
}
if (tableExists) {
try (Timer.Context context =
this.metricContext.timer(DROP_TABLE).time()) {
- client.get().dropTable(dbName, tableName);
+ client.get().dropTable(dbName, tableName, false, false);
}
String metastoreURI =
this.clientPool.getHiveConf().get(HiveMetaStoreClientFactory.HIVE_METASTORE_TOKEN_SIGNATURE,
"null");
HiveMetaStoreEventHelper.submitSuccessfulTableDrop(eventSubmitter,
dbName, tableName, metastoreURI);