Jetiaime opened a new issue, #10921:
URL: https://github.com/apache/hudi/issues/10921

   **Describe the problem you faced**
   
   I'm trying to compile Hudi 0.12.2 with Spark 3.3.1, Hadoop 3.3.2, Hive 
3.1.2. And I compiled HBase 2.4.9 With Hadoop3 in order to avoid 
[HUDI-4805](https://github.com/apache/hudi/pull/6756), And I excluded a lot of 
jetty dependencies to avoid `java.lang.NoSuchMethodError: 
org.apache.hudi.org.apache.jetty.server.session.SessionHandler.setHttpOnly(Z)V`.
 After doing this, I compiled with `mvn clean package -Pscala-2.12 -Pspark3.3 
-Pspark-bundle-shade-hive -DskipTests -Dcheckstyle.skip` On my macbook m1pro 
chips.
   
   And I got this:
   
   <img width="1624" alt="image" 
src="https://github.com/apache/hudi/assets/93769000/d1cbc391-f1d7-4671-8884-4fa5b3c12227";>
   
   Here's My Dependencies that I exclude:
   
   ```xml
     <dependencies>
       <!-- Hoodie - Spark -->
   
       <dependency>
         <groupId>org.apache.hudi</groupId>
         <artifactId>hudi-spark_${scala.binary.version}</artifactId>
         <version>${project.version}</version>
         <exclusions>
           <exclusion>
             <artifactId>websocket-client</artifactId>
             <groupId>org.eclipse.jetty.websocket</groupId>
           </exclusion>
         </exclusions>
       </dependency>
   
       <!-- Hoodie - Other -->
   
       <dependency>
         <groupId>org.apache.hudi</groupId>
         <artifactId>hudi-hadoop-mr</artifactId>
         <version>${project.version}</version>
       </dependency>
   
       <dependency>
         <groupId>org.apache.hudi</groupId>
         <artifactId>hudi-hive-sync</artifactId>
         <version>${project.version}</version>
       </dependency>
   
       <dependency>
         <groupId>org.apache.hudi</groupId>
         <artifactId>hudi-timeline-service</artifactId>
         <version>${project.version}</version>
       </dependency>
   
       <!-- Parquet -->
       <dependency>
         <groupId>org.apache.parquet</groupId>
         <artifactId>parquet-avro</artifactId>
         <scope>compile</scope>
       </dependency>
   
       <!-- Hive -->
       <dependency>
         <groupId>${hive.groupid}</groupId>
         <artifactId>hive-service</artifactId>
         <version>${hive.version}</version>
         <scope>${spark.bundle.hive.scope}</scope>
         <exclusions>
           <exclusion>
             <artifactId>servlet-api</artifactId>
             <groupId>javax.servlet</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-http</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-io</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-server</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-servlet</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-security</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-webapp</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-util</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>websocket-server</artifactId>
             <groupId>org.eclipse.jetty.websocket</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-runner</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
         </exclusions>
       </dependency>
   
       <dependency>
         <groupId>${hive.groupid}</groupId>
         <artifactId>hive-service-rpc</artifactId>
         <version>${hive.version}</version>
         <scope>${spark.bundle.hive.scope}</scope>
       </dependency>
   
       <dependency>
         <groupId>${hive.groupid}</groupId>
         <artifactId>hive-jdbc</artifactId>
         <version>${hive.version}</version>
         <scope>${spark.bundle.hive.scope}</scope>
         <exclusions>
           <exclusion>
             <artifactId>jetty-server</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-util-ajax</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-util</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
         </exclusions>
       </dependency>
   
       <dependency>
         <groupId>${hive.groupid}</groupId>
         <artifactId>hive-metastore</artifactId>
         <version>${hive.version}</version>
         <scope>${spark.bundle.hive.scope}</scope>
       </dependency>
   
       <dependency>
         <groupId>${hive.groupid}</groupId>
         <artifactId>hive-common</artifactId>
         <version>${hive.version}</version>
         <scope>${spark.bundle.hive.scope}</scope>
         <exclusions>
           <exclusion>
             <artifactId>jetty-client</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-http</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-server</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-servlet</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-webapp</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-util-ajax</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-util</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
           <exclusion>
             <artifactId>jetty-rewrite</artifactId>
             <groupId>org.eclipse.jetty</groupId>
           </exclusion>
         </exclusions>
       </dependency>
   
       <!-- 增加hudi配置版本的jetty -->
       <dependency>
         <groupId>org.eclipse.jetty</groupId>
         <artifactId>jetty-server</artifactId>
         <version>${jetty.version}</version>
       </dependency>
       <dependency>
         <groupId>org.eclipse.jetty</groupId>
         <artifactId>jetty-util</artifactId>
         <version>${jetty.version}</version>
       </dependency>
       <dependency>
         <groupId>org.eclipse.jetty</groupId>
         <artifactId>jetty-webapp</artifactId>
         <version>${jetty.version}</version>
       </dependency>
       <dependency>
         <groupId>org.eclipse.jetty</groupId>
         <artifactId>jetty-http</artifactId>
         <version>${jetty.version}</version>
       </dependency>
   
       <!-- zookeeper -->
       <dependency>
         <groupId>org.apache.curator</groupId>
         <artifactId>curator-framework</artifactId>
         <version>${zk-curator.version}</version>
       </dependency>
   
       <dependency>
         <groupId>org.apache.curator</groupId>
         <artifactId>curator-client</artifactId>
         <version>${zk-curator.version}</version>
       </dependency>
   
       <dependency>
         <groupId>org.apache.curator</groupId>
         <artifactId>curator-recipes</artifactId>
         <version>${zk-curator.version}</version>
       </dependency>
   
       <dependency>
         <groupId>javax.servlet</groupId>
         <artifactId>javax.servlet-api</artifactId>
         <version>${javax.servlet.version}</version>
       </dependency>
   
       <!-- TODO: Reinvestigate PR 633 -->
     </dependencies>
   ```
   
   **Environment Description**
   
   * Hudi version : 0.12.2
   
   * Spark version : 3.3.1
   
   * Hive version : 3.1.2
   
   * Hadoop version : 3.3.2
   
   * Storage (HDFS/S3/GCS..) : HDFS
   
   * Running on Docker? (yes/no) : no
   
   
   **Additional context**
   
   I'm using SparkSQL to insert into a very simple Hudi table:
   
   ```sql
   create table hudi_test1(uuid int, name string) using hudi.
   ```
   
   **Stacktrace**
   
   ```
   00:06  WARN: Timeline-server-based markers are not supported for HDFS: base 
path hdfs://wdpdev:8020/warehouse/tablespace/managed/hive/hudi_test1.  Falling 
back to direct markers.
   24/03/25 21:13:25 WARN TaskSetManager: Lost task 0.0 in stage 14.0 (TID 19) 
(wdpdev executor 2): org.apache.hudi.exception.HoodieUpsertException: Error 
upserting bucketType UPDATE for partition :0
           at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.handleUpsertPartition(BaseSparkCommitActionExecutor.java:329)
           at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.lambda$mapPartitionsAsRDD$a3ab3c4$1(BaseSparkCommitActionExecutor.java:244)
           at 
org.apache.spark.api.java.JavaRDDLike.$anonfun$mapPartitionsWithIndex$1(JavaRDDLike.scala:102)
           at 
org.apache.spark.api.java.JavaRDDLike.$anonfun$mapPartitionsWithIndex$1$adapted(JavaRDDLike.scala:102)
           at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$2(RDD.scala:907)
           at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$2$adapted(RDD.scala:907)
           at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
           at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:365)
           at org.apache.spark.rdd.RDD.iterator(RDD.scala:329)
           at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
           at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:365)
           at org.apache.spark.rdd.RDD.$anonfun$getOrCompute$1(RDD.scala:378)
           at 
org.apache.spark.storage.BlockManager.$anonfun$doPutIterator$1(BlockManager.scala:1535)
           at 
org.apache.spark.storage.BlockManager.org$apache$spark$storage$BlockManager$$doPut(BlockManager.scala:1445)
           at 
org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:1509)
           at 
org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:1332)
           at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:376)
           at org.apache.spark.rdd.RDD.iterator(RDD.scala:327)
           at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
           at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:365)
           at org.apache.spark.rdd.RDD.iterator(RDD.scala:329)
           at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
           at org.apache.spark.scheduler.Task.run(Task.scala:136)
           at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:548)
           at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1504)
           at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:551)
           at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
           at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
           at java.lang.Thread.run(Thread.java:750)
   Caused by: org.apache.hudi.exception.HoodieAppendException: Failed while 
appending records to 
hdfs://wdpdev:8020/warehouse/tablespace/managed/hive/hudi_test1/.hoodie/metadata/files/.files-0000_00000000000000.log.6_0-15-20
           at 
org.apache.hudi.io.HoodieAppendHandle.appendDataAndDeleteBlocks(HoodieAppendHandle.java:410)
           at 
org.apache.hudi.io.HoodieAppendHandle.doAppend(HoodieAppendHandle.java:382)
           at 
org.apache.hudi.table.action.deltacommit.BaseSparkDeltaCommitActionExecutor.handleUpdate(BaseSparkDeltaCommitActionExecutor.java:84)
           at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.handleUpsertPartition(BaseSparkCommitActionExecutor.java:322)
           ... 28 more
   Caused by: java.io.IOException: Failed to replace a bad datanode on the 
existing pipeline due to no more good datanodes being available to try. (Nodes: 
current=[DatanodeInfoWithStorage[172.17.0.16:50010,DS-5a9ce9f1-b486-473d-9407-799757e73c5a,DISK]],
 
original=[DatanodeInfoWithStorage[172.17.0.16:50010,DS-5a9ce9f1-b486-473d-9407-799757e73c5a,DISK]]).
 The current failed datanode replacement policy is DEFAULT, and a client may 
configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' 
in its configuration.
           at 
org.apache.hadoop.hdfs.DataStreamer.findNewDatanode(DataStreamer.java:1304)
           at 
org.apache.hadoop.hdfs.DataStreamer.addDatanode2ExistingPipeline(DataStreamer.java:1372)
           at 
org.apache.hadoop.hdfs.DataStreamer.handleDatanodeReplacement(DataStreamer.java:1598)
           at 
org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1499)
           at 
org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1481)
           at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:720)
   
   00:11  WARN: Timeline-server-based markers are not supported for HDFS: base 
path hdfs://wdpdev:8020/warehouse/tablespace/managed/hive/hudi_test1.  Falling 
back to direct markers.
   24/03/25 21:13:30 ERROR SparkSQLDriver: Failed in [insert into hudi_test1 
values (5, 'a5')]
   org.apache.hudi.exception.HoodieException: Could not sync using the meta 
sync class org.apache.hudi.hive.HiveSyncTool
           at 
org.apache.hudi.sync.common.util.SyncUtilHelpers.runHoodieMetaSync(SyncUtilHelpers.java:60)
           at 
org.apache.hudi.HoodieSparkSqlWriter$.$anonfun$metaSync$2(HoodieSparkSqlWriter.scala:645)
           at 
org.apache.hudi.HoodieSparkSqlWriter$.$anonfun$metaSync$2$adapted(HoodieSparkSqlWriter.scala:644)
           at scala.collection.mutable.HashSet.foreach(HashSet.scala:79)
           at 
org.apache.hudi.HoodieSparkSqlWriter$.metaSync(HoodieSparkSqlWriter.scala:644)
           at 
org.apache.hudi.HoodieSparkSqlWriter$.commitAndPerformPostOperations(HoodieSparkSqlWriter.scala:731)
           at 
org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:345)
           at 
org.apache.spark.sql.hudi.command.InsertIntoHoodieTableCommand$.run(InsertIntoHoodieTableCommand.scala:101)
           at 
org.apache.spark.sql.hudi.command.InsertIntoHoodieTableCommand.run(InsertIntoHoodieTableCommand.scala:60)
           at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:75)
           at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:73)
           at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:84)
           at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:98)
           at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:109)
           at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:169)
           at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:95)
           at 
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779)
           at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
           at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:98)
           at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:94)
           at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:584)
           at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:176)
           at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:584)
           at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:30)
           at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267)
           at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263)
           at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)
           at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)
           at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:560)
           at 
org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:94)
           at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:81)
           at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:79)
           at org.apache.spark.sql.Dataset.<init>(Dataset.scala:220)
           at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100)
           at 
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779)
           at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97)
           at 
org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:622)
           at 
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779)
           at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:617)
           at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:651)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:67)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:384)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.$anonfun$processLine$1(SparkSQLCLIDriver.scala:504)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.$anonfun$processLine$1$adapted(SparkSQLCLIDriver.scala:498)
           at scala.collection.Iterator.foreach(Iterator.scala:943)
           at scala.collection.Iterator.foreach$(Iterator.scala:943)
           at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
           at scala.collection.IterableLike.foreach(IterableLike.scala:74)
           at scala.collection.IterableLike.foreach$(IterableLike.scala:73)
           at scala.collection.AbstractIterable.foreach(Iterable.scala:56)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processLine(SparkSQLCLIDriver.scala:498)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:286)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
           at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
           at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
           at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
           at java.lang.reflect.Method.invoke(Method.java:498)
           at 
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
           at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:958)
           at 
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
           at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
           at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
           at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1046)
           at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1055)
           at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   Caused by: org.apache.hudi.exception.HoodieException: Unable to instantiate 
class org.apache.hudi.hive.HiveSyncTool
           at 
org.apache.hudi.common.util.ReflectionUtils.loadClass(ReflectionUtils.java:91)
           at 
org.apache.hudi.sync.common.util.SyncUtilHelpers.instantiateMetaSyncTool(SyncUtilHelpers.java:84)
           at 
org.apache.hudi.sync.common.util.SyncUtilHelpers.runHoodieMetaSync(SyncUtilHelpers.java:58)
           ... 64 more
   Caused by: java.lang.reflect.InvocationTargetException
           at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
Method)
           at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
           at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
           at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
           at 
org.apache.hudi.common.util.ReflectionUtils.loadClass(ReflectionUtils.java:89)
           ... 66 more
   Caused by: java.lang.NoClassDefFoundError: 
org/apache/hudi/com/google/common/collect/ImmutableMap
           at 
org.apache.hudi.org.apache.hadoop.hive.conf.SystemVariables.<clinit>(SystemVariables.java:37)
           at 
org.apache.hudi.org.apache.hadoop.hive.conf.HiveConf$ConfVars.<init>(HiveConf.java:4489)
           at 
org.apache.hudi.org.apache.hadoop.hive.conf.HiveConf$ConfVars.<init>(HiveConf.java:4449)
           at 
org.apache.hudi.org.apache.hadoop.hive.conf.HiveConf$ConfVars.<clinit>(HiveConf.java:428)
           at 
org.apache.hudi.org.apache.hadoop.hive.conf.HiveConf.<clinit>(HiveConf.java:150)
           at org.apache.hudi.hive.HiveSyncConfig.<init>(HiveSyncConfig.java:78)
           at org.apache.hudi.hive.HiveSyncTool.<init>(HiveSyncTool.java:92)
           ... 71 more
   Caused by: java.lang.ClassNotFoundException: 
org.apache.hudi.com.google.common.collect.ImmutableMap
           at java.net.URLClassLoader.findClass(URLClassLoader.java:387)
           at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
           at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:355)
           at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
           ... 78 more
   org.apache.hudi.exception.HoodieException: Could not sync using the meta 
sync class org.apache.hudi.hive.HiveSyncTool
           at 
org.apache.hudi.sync.common.util.SyncUtilHelpers.runHoodieMetaSync(SyncUtilHelpers.java:60)
           at 
org.apache.hudi.HoodieSparkSqlWriter$.$anonfun$metaSync$2(HoodieSparkSqlWriter.scala:645)
           at 
org.apache.hudi.HoodieSparkSqlWriter$.$anonfun$metaSync$2$adapted(HoodieSparkSqlWriter.scala:644)
           at scala.collection.mutable.HashSet.foreach(HashSet.scala:79)
           at 
org.apache.hudi.HoodieSparkSqlWriter$.metaSync(HoodieSparkSqlWriter.scala:644)
           at 
org.apache.hudi.HoodieSparkSqlWriter$.commitAndPerformPostOperations(HoodieSparkSqlWriter.scala:731)
           at 
org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:345)
           at 
org.apache.spark.sql.hudi.command.InsertIntoHoodieTableCommand$.run(InsertIntoHoodieTableCommand.scala:101)
           at 
org.apache.spark.sql.hudi.command.InsertIntoHoodieTableCommand.run(InsertIntoHoodieTableCommand.scala:60)
           at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:75)
           at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:73)
           at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:84)
           at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:98)
           at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$6(SQLExecution.scala:109)
           at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:169)
           at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:95)
           at 
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779)
           at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
           at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:98)
           at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:94)
           at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:584)
           at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:176)
           at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:584)
           at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:30)
           at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267)
           at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263)
           at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)
           at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)
           at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:560)
           at 
org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:94)
           at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:81)
           at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:79)
           at org.apache.spark.sql.Dataset.<init>(Dataset.scala:220)
           at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:100)
           at 
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779)
           at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97)
           at 
org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:622)
           at 
org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779)
           at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:617)
           at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:651)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:67)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:384)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.$anonfun$processLine$1(SparkSQLCLIDriver.scala:504)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.$anonfun$processLine$1$adapted(SparkSQLCLIDriver.scala:498)
           at scala.collection.Iterator.foreach(Iterator.scala:943)
           at scala.collection.Iterator.foreach$(Iterator.scala:943)
           at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
           at scala.collection.IterableLike.foreach(IterableLike.scala:74)
           at scala.collection.IterableLike.foreach$(IterableLike.scala:73)
           at scala.collection.AbstractIterable.foreach(Iterable.scala:56)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processLine(SparkSQLCLIDriver.scala:498)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:286)
           at 
org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
           at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
           at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
           at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
           at java.lang.reflect.Method.invoke(Method.java:498)
           at 
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
           at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:958)
           at 
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
           at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
           at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
           at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1046)
           at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1055)
           at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   Caused by: org.apache.hudi.exception.HoodieException: Unable to instantiate 
class org.apache.hudi.hive.HiveSyncTool
           at 
org.apache.hudi.common.util.ReflectionUtils.loadClass(ReflectionUtils.java:91)
           at 
org.apache.hudi.sync.common.util.SyncUtilHelpers.instantiateMetaSyncTool(SyncUtilHelpers.java:84)
           at 
org.apache.hudi.sync.common.util.SyncUtilHelpers.runHoodieMetaSync(SyncUtilHelpers.java:58)
           ... 64 more
   Caused by: java.lang.reflect.InvocationTargetException
           at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
Method)
           at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
           at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
           at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
           at 
org.apache.hudi.common.util.ReflectionUtils.loadClass(ReflectionUtils.java:89)
           ... 66 more
   Caused by: java.lang.NoClassDefFoundError: 
org/apache/hudi/com/google/common/collect/ImmutableMap
           at 
org.apache.hudi.org.apache.hadoop.hive.conf.SystemVariables.<clinit>(SystemVariables.java:37)
           at 
org.apache.hudi.org.apache.hadoop.hive.conf.HiveConf$ConfVars.<init>(HiveConf.java:4489)
           at 
org.apache.hudi.org.apache.hadoop.hive.conf.HiveConf$ConfVars.<init>(HiveConf.java:4449)
           at 
org.apache.hudi.org.apache.hadoop.hive.conf.HiveConf$ConfVars.<clinit>(HiveConf.java:428)
           at 
org.apache.hudi.org.apache.hadoop.hive.conf.HiveConf.<clinit>(HiveConf.java:150)
           at org.apache.hudi.hive.HiveSyncConfig.<init>(HiveSyncConfig.java:78)
           at org.apache.hudi.hive.HiveSyncTool.<init>(HiveSyncTool.java:92)
           ... 71 more
   Caused by: java.lang.ClassNotFoundException: 
org.apache.hudi.com.google.common.collect.ImmutableMap
           at java.net.URLClassLoader.findClass(URLClassLoader.java:387)
           at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
           at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:355)
           at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
           ... 78 more
   ```
   
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to