2416210017 opened a new issue, #5216:
URL: https://github.com/apache/seatunnel/issues/5216

   ### Search before asking
   
   - [X] I had searched in the 
[issues](https://github.com/apache/seatunnel/issues?q=is%3Aissue+label%3A%22bug%22)
 and found no similar issues.
   
   
   ### What happened
   
   The seatunel Zeta engine failed to connect to hive. I have placed the 
relevant jar package under lib in the seatunel installation directory, but 
attempting to connect to hive still reported an error. What is the reason?
   
   
![image](https://github.com/apache/seatunnel/assets/52597892/54e9b7ca-cd50-41c3-a744-46ac941f4657)
   
   
   ### SeaTunnel Version
   
   seatunel 2.3.2
   
   ### SeaTunnel Config
   
   ```conf
   env {
     execution.parallelism = 2
     job.mode = "BATCH"
     checkpoint.interval = 10000
   }
   
   source {
     Hive {
       table_name = "mid.ads_damai_day_ec_pro_si_ds"
       metastore_uri = "thrift://10.10.223.15:9083"
       result_table_name = "hive_starrocks_ds_t1"
       hdfs_site_path = "/etc/hadoop/conf/hdfs-site.xml"
     }
   }
   
   
   sink {
     Console {
     }
   
   }
   ```
   
   
   ### Running Command
   
   ```shell
   su hive -s /bin/bash -c "/data1/SEATUNNEL2-2.3.2/bin/seatunnel.sh --config 
/tmp/hive.conf"
   ```
   
   
   ### Error Exception
   
   ```log
   2023-08-03 17:46:37,655 INFO  com.hazelcast.core.LifecycleService - 
hz.client_1 [seatunnel-sit] [5.1] HazelcastClient 5.1 (20220228 - 21f20e7) is 
CLIENT_DISCONNECTED
   2023-08-03 17:46:37,659 INFO  com.hazelcast.core.LifecycleService - 
hz.client_1 [seatunnel-sit] [5.1] HazelcastClient 5.1 (20220228 - 21f20e7) is 
SHUTDOWN
   2023-08-03 17:46:37,659 INFO  
org.apache.seatunnel.core.starter.seatunnel.command.ClientExecuteCommand - 
Closed SeaTunnel client......
   2023-08-03 17:46:37,659 ERROR org.apache.seatunnel.core.starter.SeaTunnel - 
   
   
===============================================================================
   
   
   2023-08-03 17:46:37,659 ERROR org.apache.seatunnel.core.starter.SeaTunnel - 
Fatal Error, 
   
   2023-08-03 17:46:37,660 ERROR org.apache.seatunnel.core.starter.SeaTunnel - 
Please submit bug report in https://github.com/apache/seatunnel/issues
   
   2023-08-03 17:46:37,660 ERROR org.apache.seatunnel.core.starter.SeaTunnel - 
Reason:SeaTunnel job executed failed 
   
   2023-08-03 17:46:37,663 ERROR org.apache.seatunnel.core.starter.SeaTunnel - 
Exception 
StackTrace:org.apache.seatunnel.core.starter.exception.CommandExecuteException: 
SeaTunnel job executed failed
           at 
org.apache.seatunnel.core.starter.seatunnel.command.ClientExecuteCommand.execute(ClientExecuteCommand.java:188)
           at org.apache.seatunnel.core.starter.SeaTunnel.run(SeaTunnel.java:40)
           at 
org.apache.seatunnel.core.starter.seatunnel.SeaTunnelClient.main(SeaTunnelClient.java:34)
   Caused by: java.util.concurrent.CompletionException: 
java.lang.NoClassDefFoundError: Lorg/apache/hadoop/hive/metastore/api/Table;
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.wrapInCompletionException(AbstractInvocationFuture.java:1347)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.cascadeException(AbstractInvocationFuture.java:1340)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.access$200(AbstractInvocationFuture.java:65)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture$ApplyNode.execute(AbstractInvocationFuture.java:1478)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.unblockOtherNode(AbstractInvocationFuture.java:797)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.unblockAll(AbstractInvocationFuture.java:759)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.complete0(AbstractInvocationFuture.java:1235)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.completeExceptionallyInternal(AbstractInvocationFuture.java:1223)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.completeExceptionally(AbstractInvocationFuture.java:709)
           at 
com.hazelcast.client.impl.spi.impl.ClientInvocation.completeExceptionally(ClientInvocation.java:294)
           at 
com.hazelcast.client.impl.spi.impl.ClientInvocation.notifyExceptionWithOwnedPermission(ClientInvocation.java:321)
           at 
com.hazelcast.client.impl.spi.impl.ClientInvocation.notifyException(ClientInvocation.java:304)
           at 
com.hazelcast.client.impl.spi.impl.ClientResponseHandlerSupplier.handleResponse(ClientResponseHandlerSupplier.java:164)
           at 
com.hazelcast.client.impl.spi.impl.ClientResponseHandlerSupplier.process(ClientResponseHandlerSupplier.java:141)
           at 
com.hazelcast.client.impl.spi.impl.ClientResponseHandlerSupplier.access$300(ClientResponseHandlerSupplier.java:60)
           at 
com.hazelcast.client.impl.spi.impl.ClientResponseHandlerSupplier$DynamicResponseHandler.accept(ClientResponseHandlerSupplier.java:251)
           at 
com.hazelcast.client.impl.spi.impl.ClientResponseHandlerSupplier$DynamicResponseHandler.accept(ClientResponseHandlerSupplier.java:243)
           at 
com.hazelcast.client.impl.connection.tcp.TcpClientConnection.handleClientMessage(TcpClientConnection.java:245)
           at 
com.hazelcast.client.impl.protocol.util.ClientMessageDecoder.handleMessage(ClientMessageDecoder.java:135)
           at 
com.hazelcast.client.impl.protocol.util.ClientMessageDecoder.onRead(ClientMessageDecoder.java:89)
           at 
com.hazelcast.internal.networking.nio.NioInboundPipeline.process(NioInboundPipeline.java:136)
           at 
com.hazelcast.internal.networking.nio.NioThread.processSelectionKey(NioThread.java:383)
           at 
com.hazelcast.internal.networking.nio.NioThread.processSelectionKeys(NioThread.java:368)
           at 
com.hazelcast.internal.networking.nio.NioThread.selectLoop(NioThread.java:294)
           at 
com.hazelcast.internal.networking.nio.NioThread.executeRun(NioThread.java:249)
           at 
com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102)
   Caused by: java.lang.NoClassDefFoundError: 
Lorg/apache/hadoop/hive/metastore/api/Table;
           at java.lang.Class.getDeclaredFields0(Native Method)
           at java.lang.Class.privateGetDeclaredFields(Class.java:2583)
           at java.lang.Class.getDeclaredField(Class.java:2068)
           at 
java.io.ObjectStreamClass.getDeclaredSUID(ObjectStreamClass.java:1871)
           at java.io.ObjectStreamClass.access$700(ObjectStreamClass.java:79)
           at java.io.ObjectStreamClass$3.run(ObjectStreamClass.java:506)
           at java.io.ObjectStreamClass$3.run(ObjectStreamClass.java:494)
           at java.security.AccessController.doPrivileged(Native Method)
           at java.io.ObjectStreamClass.<init>(ObjectStreamClass.java:494)
           at java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:391)
           at java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:681)
           at 
java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:2002)
           at 
java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1849)
           at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2159)
           at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1666)
           at 
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2404)
           at 
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2328)
           at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2186)
           at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1666)
           at java.io.ObjectInputStream.readObject(ObjectInputStream.java:502)
           at java.io.ObjectInputStream.readObject(ObjectInputStream.java:460)
           at 
com.hazelcast.internal.serialization.impl.defaultserializers.JavaDefaultSerializers$JavaSerializer.read(JavaDefaultSerializers.java:92)
           at 
com.hazelcast.internal.serialization.impl.defaultserializers.JavaDefaultSerializers$JavaSerializer.read(JavaDefaultSerializers.java:85)
           at 
com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44)
           at 
com.hazelcast.internal.serialization.impl.AbstractSerializationService.readObject(AbstractSerializationService.java:349)
           at 
com.hazelcast.internal.serialization.impl.ByteArrayObjectDataInput.readObject(ByteArrayObjectDataInput.java:600)
           at 
org.apache.seatunnel.engine.core.dag.logical.LogicalVertex.readData(LogicalVertex.java:99)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.readInternal(DataSerializableSerializer.java:160)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.read(DataSerializableSerializer.java:106)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.read(DataSerializableSerializer.java:51)
           at 
com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44)
           at 
com.hazelcast.internal.serialization.impl.AbstractSerializationService.readObject(AbstractSerializationService.java:349)
           at 
com.hazelcast.internal.serialization.impl.ByteArrayObjectDataInput.readObject(ByteArrayObjectDataInput.java:600)
           at 
org.apache.seatunnel.engine.core.dag.logical.LogicalDag.readData(LogicalDag.java:154)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.readInternal(DataSerializableSerializer.java:160)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.read(DataSerializableSerializer.java:106)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.read(DataSerializableSerializer.java:51)
           at 
com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44)
           at 
com.hazelcast.internal.serialization.impl.AbstractSerializationService.toObject(AbstractSerializationService.java:268)
           at 
com.hazelcast.jet.impl.execution.init.CustomClassLoadedObject.deserializeWithCustomClassLoader(CustomClassLoadedObject.java:66)
           at 
org.apache.seatunnel.engine.server.master.JobMaster.init(JobMaster.java:203)
           at 
org.apache.seatunnel.engine.server.CoordinatorService.lambda$submitJob$5(CoordinatorService.java:438)
           at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
           at java.util.concurrent.FutureTask.run(FutureTask.java:266)
           at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
           at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
           at java.lang.Thread.run(Thread.java:748)
    
   2023-08-03 17:46:37,664 ERROR org.apache.seatunnel.core.starter.SeaTunnel - 
   
===============================================================================
   
   
   
   Exception in thread "main" 
org.apache.seatunnel.core.starter.exception.CommandExecuteException: SeaTunnel 
job executed failed
           at 
org.apache.seatunnel.core.starter.seatunnel.command.ClientExecuteCommand.execute(ClientExecuteCommand.java:188)
           at org.apache.seatunnel.core.starter.SeaTunnel.run(SeaTunnel.java:40)
           at 
org.apache.seatunnel.core.starter.seatunnel.SeaTunnelClient.main(SeaTunnelClient.java:34)
   Caused by: java.util.concurrent.CompletionException: 
java.lang.NoClassDefFoundError: Lorg/apache/hadoop/hive/metastore/api/Table;
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.wrapInCompletionException(AbstractInvocationFuture.java:1347)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.cascadeException(AbstractInvocationFuture.java:1340)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.access$200(AbstractInvocationFuture.java:65)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture$ApplyNode.execute(AbstractInvocationFuture.java:1478)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.unblockOtherNode(AbstractInvocationFuture.java:797)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.unblockAll(AbstractInvocationFuture.java:759)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.complete0(AbstractInvocationFuture.java:1235)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.completeExceptionallyInternal(AbstractInvocationFuture.java:1223)
           at 
com.hazelcast.spi.impl.AbstractInvocationFuture.completeExceptionally(AbstractInvocationFuture.java:709)
           at 
com.hazelcast.client.impl.spi.impl.ClientInvocation.completeExceptionally(ClientInvocation.java:294)
           at 
com.hazelcast.client.impl.spi.impl.ClientInvocation.notifyExceptionWithOwnedPermission(ClientInvocation.java:321)
           at 
com.hazelcast.client.impl.spi.impl.ClientInvocation.notifyException(ClientInvocation.java:304)
           at 
com.hazelcast.client.impl.spi.impl.ClientResponseHandlerSupplier.handleResponse(ClientResponseHandlerSupplier.java:164)
           at 
com.hazelcast.client.impl.spi.impl.ClientResponseHandlerSupplier.process(ClientResponseHandlerSupplier.java:141)
           at 
com.hazelcast.client.impl.spi.impl.ClientResponseHandlerSupplier.access$300(ClientResponseHandlerSupplier.java:60)
           at 
com.hazelcast.client.impl.spi.impl.ClientResponseHandlerSupplier$DynamicResponseHandler.accept(ClientResponseHandlerSupplier.java:251)
           at 
com.hazelcast.client.impl.spi.impl.ClientResponseHandlerSupplier$DynamicResponseHandler.accept(ClientResponseHandlerSupplier.java:243)
           at 
com.hazelcast.client.impl.connection.tcp.TcpClientConnection.handleClientMessage(TcpClientConnection.java:245)
           at 
com.hazelcast.client.impl.protocol.util.ClientMessageDecoder.handleMessage(ClientMessageDecoder.java:135)
           at 
com.hazelcast.client.impl.protocol.util.ClientMessageDecoder.onRead(ClientMessageDecoder.java:89)
           at 
com.hazelcast.internal.networking.nio.NioInboundPipeline.process(NioInboundPipeline.java:136)
           at 
com.hazelcast.internal.networking.nio.NioThread.processSelectionKey(NioThread.java:383)
           at 
com.hazelcast.internal.networking.nio.NioThread.processSelectionKeys(NioThread.java:368)
           at 
com.hazelcast.internal.networking.nio.NioThread.selectLoop(NioThread.java:294)
           at 
com.hazelcast.internal.networking.nio.NioThread.executeRun(NioThread.java:249)
           at 
com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102)
   Caused by: java.lang.NoClassDefFoundError: 
Lorg/apache/hadoop/hive/metastore/api/Table;
           at java.lang.Class.getDeclaredFields0(Native Method)
           at java.lang.Class.privateGetDeclaredFields(Class.java:2583)
           at java.lang.Class.getDeclaredField(Class.java:2068)
           at 
java.io.ObjectStreamClass.getDeclaredSUID(ObjectStreamClass.java:1871)
           at java.io.ObjectStreamClass.access$700(ObjectStreamClass.java:79)
           at java.io.ObjectStreamClass$3.run(ObjectStreamClass.java:506)
           at java.io.ObjectStreamClass$3.run(ObjectStreamClass.java:494)
           at java.security.AccessController.doPrivileged(Native Method)
           at java.io.ObjectStreamClass.<init>(ObjectStreamClass.java:494)
           at java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:391)
           at java.io.ObjectStreamClass.initNonProxy(ObjectStreamClass.java:681)
           at 
java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:2002)
           at 
java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1849)
           at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2159)
           at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1666)
           at 
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2404)
           at 
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2328)
           at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2186)
           at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1666)
           at java.io.ObjectInputStream.readObject(ObjectInputStream.java:502)
           at java.io.ObjectInputStream.readObject(ObjectInputStream.java:460)
           at 
com.hazelcast.internal.serialization.impl.defaultserializers.JavaDefaultSerializers$JavaSerializer.read(JavaDefaultSerializers.java:92)
           at 
com.hazelcast.internal.serialization.impl.defaultserializers.JavaDefaultSerializers$JavaSerializer.read(JavaDefaultSerializers.java:85)
           at 
com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44)
           at 
com.hazelcast.internal.serialization.impl.AbstractSerializationService.readObject(AbstractSerializationService.java:349)
           at 
com.hazelcast.internal.serialization.impl.ByteArrayObjectDataInput.readObject(ByteArrayObjectDataInput.java:600)
           at 
org.apache.seatunnel.engine.core.dag.logical.LogicalVertex.readData(LogicalVertex.java:99)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.readInternal(DataSerializableSerializer.java:160)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.read(DataSerializableSerializer.java:106)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.read(DataSerializableSerializer.java:51)
           at 
com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44)
           at 
com.hazelcast.internal.serialization.impl.AbstractSerializationService.readObject(AbstractSerializationService.java:349)
           at 
com.hazelcast.internal.serialization.impl.ByteArrayObjectDataInput.readObject(ByteArrayObjectDataInput.java:600)
           at 
org.apache.seatunnel.engine.core.dag.logical.LogicalDag.readData(LogicalDag.java:154)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.readInternal(DataSerializableSerializer.java:160)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.read(DataSerializableSerializer.java:106)
           at 
com.hazelcast.internal.serialization.impl.DataSerializableSerializer.read(DataSerializableSerializer.java:51)
           at 
com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44)
           at 
com.hazelcast.internal.serialization.impl.AbstractSerializationService.toObject(AbstractSerializationService.java:268)
           at 
com.hazelcast.jet.impl.execution.init.CustomClassLoadedObject.deserializeWithCustomClassLoader(CustomClassLoadedObject.java:66)
           at 
org.apache.seatunnel.engine.server.master.JobMaster.init(JobMaster.java:203)
           at 
org.apache.seatunnel.engine.server.CoordinatorService.lambda$submitJob$5(CoordinatorService.java:438)
           at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
           at java.util.concurrent.FutureTask.run(FutureTask.java:266)
           at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
           at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
           at java.lang.Thread.run(Thread.java:748)
   ```
   
   
   ### Zeta or Flink or Spark Version
   
   Zeta 
   
   ### Java or Scala Version
   
   java version "1.8.0_181"
   
   ### Screenshots
   
   
![image](https://github.com/apache/seatunnel/assets/52597892/298d4a93-bee0-4f1d-9856-648f332b6935)
   
   
   ### Are you willing to submit PR?
   
   - [X] Yes I am willing to submit a PR!
   
   ### Code of Conduct
   
   - [X] I agree to follow this project's [Code of 
Conduct](https://www.apache.org/foundation/policies/conduct)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to