newsbreak-tonglin opened a new issue, #8586:
URL: https://github.com/apache/hudi/issues/8586

   use Flink Mongo CDC fetch data from mongo to Hudi MOR table, sync ro table 
success, but sync rt table failed
   
   with error message:
   
   2023-04-27 07:46:27,967 INFO  
org.apache.hadoop.hive.metastore.HiveMetaStoreClient         [] - Trying to 
connect to metastore with URI thrift://ip-xxx-xx-xxx-xxx:9083
   2023-04-27 07:46:27,992 INFO  
org.apache.hadoop.hive.metastore.HiveMetaStoreClient         [] - Opened a 
connection to metastore, current connections: 1
   2023-04-27 07:46:28,001 INFO  
org.apache.hadoop.hive.metastore.HiveMetaStoreClient         [] - Connected to 
metastore.
   2023-04-27 07:46:28,001 INFO  
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient     [] - 
RetryingMetaStoreClient proxy=class 
org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient ugi=hadoop 
(auth:SIMPLE) retries=1 delay=1 lifetime=0
   2023-04-27 07:46:28,176 INFO  org.apache.hudi.hive.HiveSyncTool              
              [] - Syncing target hoodie table with hive 
table(default.mongo_cdc_hudi_xxxx_test25). Hive metastore URL 
:thrift://xxx:9083, basePath :s3://xxxx/hudi_test25
   2023-04-27 07:46:28,176 INFO  org.apache.hudi.hive.HiveSyncTool              
              [] - Trying to sync hoodie table mongo_cdc_hudi_xxx_test25_ro 
with base path s3://xxxx/hudi_test25 of type MERGE_ON_READ
   2023-04-27 07:46:28,206 ERROR org.apache.hudi.hive.ddl.HMSDDLExecutor        
              [] - Failed to create database default
   org.apache.hadoop.hive.metastore.api.AlreadyExistsException: Database 
default already exists
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result$create_database_resultStandardScheme.read(ThriftHiveMetastore.java:39325)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result$create_database_resultStandardScheme.read(ThriftHiveMetastore.java:39311)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result.read(ThriftHiveMetastore.java:39245)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_create_database(ThriftHiveMetastore.java:1106)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.create_database(ThriftHiveMetastore.java:1093)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createDatabase(HiveMetaStoreClient.java:809)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
~[?:1.8.0_352]
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
~[?:1.8.0_352]
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 ~[?:1.8.0_352]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_352]
        at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:212)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at com.sun.proxy.$Proxy121.createDatabase(Unknown Source) ~[?:?]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
~[?:1.8.0_352]
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
~[?:1.8.0_352]
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 ~[?:1.8.0_352]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_352]
        at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2773)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at com.sun.proxy.$Proxy121.createDatabase(Unknown Source) ~[?:?]
        at 
org.apache.hudi.hive.ddl.HMSDDLExecutor.createDatabase(HMSDDLExecutor.java:95) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HoodieHiveSyncClient.createDatabase(HoodieHiveSyncClient.java:224)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncHoodieTable(HiveSyncTool.java:187) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at org.apache.hudi.hive.HiveSyncTool.doSync(HiveSyncTool.java:158) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncHoodieTable(HiveSyncTool.java:142) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.sink.StreamWriteOperatorCoordinator.doSyncHive(StreamWriteOperatorCoordinator.java:335)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:0.12.1]
        at 
org.apache.hudi.sink.utils.NonThrownExecutor.lambda$wrapAction$0(NonThrownExecutor.java:130)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:0.12.1]
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
[?:1.8.0_352]
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
[?:1.8.0_352]
        at java.lang.Thread.run(Thread.java:750) [?:1.8.0_352]
   2023-04-27 07:46:28,207 WARN  org.apache.hudi.hive.HiveSyncTool              
              [] - Unable to create database
   org.apache.hudi.hive.HoodieHiveSyncException: Failed to create database 
default
        at 
org.apache.hudi.hive.ddl.HMSDDLExecutor.createDatabase(HMSDDLExecutor.java:98) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HoodieHiveSyncClient.createDatabase(HoodieHiveSyncClient.java:224)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncHoodieTable(HiveSyncTool.java:187) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at org.apache.hudi.hive.HiveSyncTool.doSync(HiveSyncTool.java:158) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncHoodieTable(HiveSyncTool.java:142) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.sink.StreamWriteOperatorCoordinator.doSyncHive(StreamWriteOperatorCoordinator.java:335)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:0.12.1]
        at 
org.apache.hudi.sink.utils.NonThrownExecutor.lambda$wrapAction$0(NonThrownExecutor.java:130)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:0.12.1]
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
[?:1.8.0_352]
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
[?:1.8.0_352]
        at java.lang.Thread.run(Thread.java:750) [?:1.8.0_352]
   Caused by: org.apache.hadoop.hive.metastore.api.AlreadyExistsException: 
Database default already exists
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result$create_database_resultStandardScheme.read(ThriftHiveMetastore.java:39325)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result$create_database_resultStandardScheme.read(ThriftHiveMetastore.java:39311)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$create_database_result.read(ThriftHiveMetastore.java:39245)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_create_database(ThriftHiveMetastore.java:1106)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.create_database(ThriftHiveMetastore.java:1093)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createDatabase(HiveMetaStoreClient.java:809)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
~[?:1.8.0_352]
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
~[?:1.8.0_352]
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 ~[?:1.8.0_352]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_352]
        at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:212)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at com.sun.proxy.$Proxy121.createDatabase(Unknown Source) ~[?:?]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
~[?:1.8.0_352]
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
~[?:1.8.0_352]
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 ~[?:1.8.0_352]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_352]
        at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2773)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at com.sun.proxy.$Proxy121.createDatabase(Unknown Source) ~[?:?]
        at 
org.apache.hudi.hive.ddl.HMSDDLExecutor.createDatabase(HMSDDLExecutor.java:95) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        ... 9 more
   2023-04-27 07:46:28,244 INFO  
com.amazon.ws.emr.hadoop.fs.s3n.S3NativeFileSystem           [] - Opening 
's3://xxxx/hudi_test25/.hoodie/20230427074259124.deltacommit' for reading
   2023-04-27 07:46:28,263 INFO  
com.amazon.ws.emr.hadoop.fs.s3n.S3NativeFileSystem           [] - Opening 
's3://xxxx/hudi_test25/.hoodie/20230427074259124.deltacommit' for reading
   2023-04-27 07:46:28,331 INFO  
com.amazon.ws.emr.hadoop.fs.s3n.S3NativeFileSystem           [] - Opening 
's3://xxxx/hudi_test25/.245cbf6d-8d54-40ed-908a-6b28503cee9f_20230427074259124.log.3_0-5-0'
 for reading
   2023-04-27 07:46:28,612 INFO  
com.ververica.cdc.connectors.base.source.enumerator.IncrementalSourceEnumerator 
[] - The enumerator receives finished split offsets 
FinishedSnapshotSplitsReportEvent{finishedOffsets={xxxx.xxxx:49={resumeToken={"_data":
 "82644A2839000000022B0229296E04"}, timestamp=7226632777347629058}}} from 
subtask 3.
   2023-04-27 07:46:28,613 INFO  
org.apache.flink.runtime.source.coordinator.SourceCoordinator [] - Source 
Source: mongo_cdc_test[1] received split request from parallel task 3
   2023-04-27 07:46:28,613 INFO  
com.ververica.cdc.connectors.base.source.enumerator.IncrementalSourceEnumerator 
[] - Assign split SnapshotSplit{tableId=xxxx.xxxx, splitId='xxxx.xxxx:54', 
splitKeyType=[`_id` INT], splitStart=[{"_id": 1.0}, {"_id": "xxxx"}], 
splitEnd=[{"_id": 1.0}, {"_id": "xxxx"}], highWatermark=null} to subtask 3
   2023-04-27 07:46:28,614 INFO  org.apache.hudi.hive.HiveSyncTool              
              [] - Hive table mongo_cdc_hudi_xxxx_test25_ro is not found. 
Creating it
   2023-04-27 07:46:28,823 INFO  org.apache.hudi.hive.HiveSyncTool              
              [] - Schema sync complete. Syncing partitions for 
mongo_cdc_hudi_xxxx_test25_ro
   2023-04-27 07:46:28,823 INFO  org.apache.hudi.hive.HiveSyncTool              
              [] - Last commit time synced was found to be null
   2023-04-27 07:46:28,823 INFO  org.apache.hudi.sync.common.HoodieSyncClient   
              [] - Last commit time synced is not known, listing all partitions 
in s3://xxxx/hudi_test25,FS 
:com.amazon.ws.emr.hadoop.fs.s3n.S3NativeFileSystem@5e4e3478
   2023-04-27 07:46:28,957 INFO  org.apache.hudi.hive.HiveSyncTool              
              [] - Storage partitions scan complete. Found 1
   2023-04-27 07:46:28,970 INFO  
org.apache.hadoop.hive.metastore.HiveMetaStoreClient         [] - Closed a 
connection to metastore, current connections: 0
   2023-04-27 07:46:28,971 ERROR 
org.apache.hudi.sink.StreamWriteOperatorCoordinator          [] - Executor 
executes action [sync hive metadata for instant 20230427074627270] error
   org.apache.hudi.exception.HoodieException: Got runtime exception when hive 
syncing mongo_cdc_hudi_xxxx_test25
        at 
org.apache.hudi.hive.HiveSyncTool.syncHoodieTable(HiveSyncTool.java:145) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.sink.StreamWriteOperatorCoordinator.doSyncHive(StreamWriteOperatorCoordinator.java:335)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:0.12.1]
        at 
org.apache.hudi.sink.utils.NonThrownExecutor.lambda$wrapAction$0(NonThrownExecutor.java:130)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:0.12.1]
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
[?:1.8.0_352]
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
[?:1.8.0_352]
        at java.lang.Thread.run(Thread.java:750) [?:1.8.0_352]
   Caused by: org.apache.hudi.hive.HoodieHiveSyncException: Failed to sync 
partitions for table mongo_cdc_hudi_xxxx_test25_ro
        at 
org.apache.hudi.hive.HiveSyncTool.syncPartitions(HiveSyncTool.java:341) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncHoodieTable(HiveSyncTool.java:232) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at org.apache.hudi.hive.HiveSyncTool.doSync(HiveSyncTool.java:158) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncHoodieTable(HiveSyncTool.java:142) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        ... 5 more
   Caused by: org.apache.hudi.hive.HoodieHiveSyncException: Failed to get all 
partitions for table default.mongo_cdc_hudi_xxxx_test25_ro
        at 
org.apache.hudi.hive.HoodieHiveSyncClient.getAllPartitions(HoodieHiveSyncClient.java:180)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncPartitions(HiveSyncTool.java:317) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncHoodieTable(HiveSyncTool.java:232) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at org.apache.hudi.hive.HiveSyncTool.doSync(HiveSyncTool.java:158) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncHoodieTable(HiveSyncTool.java:142) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        ... 5 more
   Caused by: org.apache.hadoop.hive.metastore.api.NoSuchObjectException: 
@hive#default.mongo_cdc_hudi_xxxx_test25_ro table not found
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_partitions_result$get_partitions_resultStandardScheme.read(ThriftHiveMetastore.java)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_partitions_result$get_partitions_resultStandardScheme.read(ThriftHiveMetastore.java)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$get_partitions_result.read(ThriftHiveMetastore.java)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_get_partitions(ThriftHiveMetastore.java:2958)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_partitions(ThriftHiveMetastore.java:2943)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.listPartitions(HiveMetaStoreClient.java:1368)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient.listPartitions(HiveMetaStoreClient.java:1362)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
~[?:1.8.0_352]
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
~[?:1.8.0_352]
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 ~[?:1.8.0_352]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_352]
        at 
org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:212)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at com.sun.proxy.$Proxy121.listPartitions(Unknown Source) ~[?:?]
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
~[?:1.8.0_352]
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
~[?:1.8.0_352]
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 ~[?:1.8.0_352]
        at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_352]
        at 
org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2773)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at com.sun.proxy.$Proxy121.listPartitions(Unknown Source) ~[?:?]
        at 
org.apache.hudi.hive.HoodieHiveSyncClient.getAllPartitions(HoodieHiveSyncClient.java:175)
 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncPartitions(HiveSyncTool.java:317) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncHoodieTable(HiveSyncTool.java:232) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at org.apache.hudi.hive.HiveSyncTool.doSync(HiveSyncTool.java:158) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        at 
org.apache.hudi.hive.HiveSyncTool.syncHoodieTable(HiveSyncTool.java:142) 
~[blob_p-b71662d6c81e3e8943644ff10285f67a0c201af0-ff618a8f3d6ea6018fee8fd3b95760be:?]
        ... 5 more
   2023-04-27 07:46:29,306 INFO  
com.ververica.cdc.connectors.base.source.enumerator.IncrementalSourceEnumerator 
[] - The enumerator receives finished split offsets 
FinishedSnapshotSplitsReportEvent{finishedOffsets={user_feature.user_secondcat_emb_v2:51={resumeToken={"_data":
 "82644A2848000000022B0229296E04"}, timestamp=7226632841772138498}}} from 
subtask 0.
   
   
   **Environment Description**
   
   * Hudi version : 0.12.1
   
   * Spark version : use Flink 15.2
   
   * Hive version : 2.3.7
   
   * Hadoop version : use s3
   
   * Storage (HDFS/S3/GCS..) : S3
   
   * Running on Docker? (yes/no) : no
   
   
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to