gbcoder2020 commented on issue #13002:
URL: https://github.com/apache/hudi/issues/13002#issuecomment-2743922673

   @nsivabalan 
   
   ```
   2025-03-11T15:04:00.241Z
   25/03/11 15:03:59 ERROR YarnScheduler: Lost executor 1492 on <masked ***>: 
Executor heartbeat timed out after 149584 ms
   
   25/03/11 15:03:59 ERROR YarnScheduler: Lost executor 1492 on <masked ***> 
Executor heartbeat timed out after 149584 ms
   2025-03-11T15:04:00.241Z
   25/03/11 15:03:59 INFO DAGScheduler: ResultStage 260 (collect at 
HoodieJavaRDD.java:177) failed in 7303.427 s due to Job aborted due to stage 
failure: Task 2075 in stage 260.0 failed 4 times, most recent failure: Lost 
task 2075.3 in stage 260.0 (TID 1152449) (<masked ***> executor 1493): 
ExecutorLostFailure (executor 1493 exited caused by one of the running tasks) 
Reason: Executor heartbeat timed out after 131337 ms
   
   25/03/11 15:03:59 INFO DAGScheduler: ResultStage 260 (collect at 
HoodieJavaRDD.java:177) failed in 7303.427 s due to Job aborted due to stage 
failure: Task 2075 in stage 260.0 failed 4 times, most recent failure: Lost 
task 2075.3 in stage 260.0 (TID 1152449) (<masked ***> executor 1493): 
ExecutorLostFailure (executor 1493 exited caused by one of the running tasks) 
Reason: Executor heartbeat timed out after 131337 ms
   2025-03-11T15:04:00.241Z
   Driver stacktrace:
   
   Driver stacktrace:
   2025-03-11T15:04:00.241Z
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 1473.2 in stage 260.0 (TID 
1152417) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 1473.2 in stage 260.0 (TID 
1152417) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   2025-03-11T15:04:00.241Z
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 873.2 in stage 260.0 (TID 
1152416) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 873.2 in stage 260.0 (TID 
1152416) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   2025-03-11T15:04:00.241Z
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 2073.2 in stage 260.0 (TID 
1152419) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 2073.2 in stage 260.0 (TID 
1152419) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   2025-03-11T15:04:00.241Z
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 2373.2 in stage 260.0 (TID 
1152413) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 2373.2 in stage 260.0 (TID 
1152413) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   2025-03-11T15:04:00.241Z
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 1773.2 in stage 260.0 (TID 
1152412) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 1773.2 in stage 260.0 (TID 
1152412) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   2025-03-11T15:04:00.241Z
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 273.2 in stage 260.0 (TID 
1152415) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 273.2 in stage 260.0 (TID 
1152415) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   2025-03-11T15:04:00.241Z
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 573.2 in stage 260.0 (TID 
1152418) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 573.2 in stage 260.0 (TID 
1152418) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   2025-03-11T15:04:00.241Z
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 1173.2 in stage 260.0 (TID 
1152414) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   
   25/03/11 15:03:59 WARN TaskSetManager: Lost task 1173.2 in stage 260.0 (TID 
1152414) (<masked ***> executor 1492): ExecutorLostFailure (executor 1492 
exited caused by one of the running tasks) Reason: Executor heartbeat timed out 
after 149584 ms
   2025-03-11T15:04:00.241Z
   25/03/11 15:03:59 INFO ExecutorMonitor: Executor 1492 is removed. Remove 
reason statistics: (gracefully decommissioned: 0, decommision unfinished: 0, 
driver killed: 603, unexpectedly exited: 877).
   
   25/03/11 15:03:59 INFO ExecutorMonitor: Executor 1492 is removed. Remove 
reason statistics: (gracefully decommissioned: 0, decommision unfinished: 0, 
driver killed: 603, unexpectedly exited: 877).
   2025-03-11T15:04:00.241Z
   25/03/11 15:04:00 INFO DAGScheduler: Executor lost: 1493 (epoch 102)
   
   25/03/11 15:04:00 INFO DAGScheduler: Executor lost: 1493 (epoch 102)
   2025-03-11T15:04:00.241Z
   25/03/11 15:04:00 INFO BlockManagerMasterEndpoint: Trying to remove executor 
1493 from BlockManagerMaster.
   
   25/03/11 15:04:00 INFO BlockManagerMasterEndpoint: Trying to remove executor 
1493 from BlockManagerMaster.
   2025-03-11T15:04:00.241Z
   25/03/11 15:04:00 INFO BlockManagerMasterEndpoint: Removing block manager 
BlockManagerId(1493, <masked ***>, 44649, None)
   
   25/03/11 15:04:00 INFO BlockManagerMasterEndpoint: Removing block manager 
BlockManagerId(1493, <masked ***>, 44649, None)
   2025-03-11T15:04:00.241Z
   25/03/11 15:04:00 INFO BlockManagerMaster: Removed 1493 successfully in 
removeExecutor
   
   25/03/11 15:04:00 INFO BlockManagerMaster: Removed 1493 successfully in 
removeExecutor
   2025-03-11T15:04:00.241Z
   25/03/11 15:04:00 INFO DAGScheduler: Job 117 failed: collect at 
HoodieJavaRDD.java:177, took 7303.439556 s
   
   25/03/11 15:04:00 INFO DAGScheduler: Job 117 failed: collect at 
HoodieJavaRDD.java:177, took 7303.439556 s
   2025-03-11T15:04:00.241Z
   25/03/11 15:04:00 INFO DAGScheduler: Executor lost: 1492 (epoch 102)
   
   25/03/11 15:04:00 INFO DAGScheduler: Executor lost: 1492 (epoch 102)
   2025-03-11T15:04:00.241Z
   25/03/11 15:04:00 INFO BlockManagerMasterEndpoint: Trying to remove executor 
1492 from BlockManagerMaster.
   
   25/03/11 15:04:00 INFO BlockManagerMasterEndpoint: Trying to remove executor 
1492 from BlockManagerMaster.
   2025-03-11T15:04:00.241Z
   25/03/11 15:04:00 INFO BlockManagerMasterEndpoint: Removing block manager 
BlockManagerId(1492, <masked ***>, 33069, None)
   
   25/03/11 15:04:00 INFO BlockManagerMasterEndpoint: Removing block manager 
BlockManagerId(1492, <masked ***>, 33069, None)
   2025-03-11T15:04:00.241Z
   25/03/11 15:04:00 INFO BlockManagerMaster: Removed 1492 successfully in 
removeExecutor
   
   25/03/11 15:04:00 INFO BlockManagerMaster: Removed 1492 successfully in 
removeExecutor
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 ERROR HoodieBackedTableMetadataWriter: Exception in 
running table services on metadata table
   
   25/03/11 15:04:00 ERROR HoodieBackedTableMetadataWriter: Exception in 
running table services on metadata table
   2025-03-11T15:04:00.242Z
   org.apache.hudi.exception.HoodieCompactionException: Could not compact 
s3://<s3 path>/.hoodie/metadata
        at 
org.apache.hudi.table.action.compact.RunCompactionActionExecutor.execute(RunCompactionActionExecutor.java:129)
        at 
org.apache.hudi.table.HoodieSparkMergeOnReadTable.compact(HoodieSparkMergeOnReadTable.java:155)
        at 
org.apache.hudi.client.BaseHoodieTableServiceClient.compact(BaseHoodieTableServiceClient.java:297)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.compact(BaseHoodieWriteClient.java:1126)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.compact(BaseHoodieWriteClient.java:1046)
        at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.compactIfNecessary(HoodieBackedTableMetadataWriter.java:1316)
        at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.performTableServices(HoodieBackedTableMetadataWriter.java:1267)
        at 
org.apache.hudi.client.SparkRDDWriteClient.initializeMetadataTable(SparkRDDWriteClient.java:323)
        at 
org.apache.hudi.client.SparkRDDWriteClient.initMetadataTable(SparkRDDWriteClient.java:288)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.doInitTable(BaseHoodieWriteClient.java:1244)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.initTable(BaseHoodieWriteClient.java:1284)
        at 
org.apache.hudi.client.SparkRDDWriteClient.upsert(SparkRDDWriteClient.java:154)
        at 
org.apache.hudi.DataSourceUtils.doWriteOperation(DataSourceUtils.java:225)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.liftedTree1$1(HoodieSparkSqlWriter.scala:492)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.writeInternal(HoodieSparkSqlWriter.scala:490)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.write(HoodieSparkSqlWriter.scala:187)
        at 
org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:125)
        at org.apache.hudi.DefaultSource.createRelation(DefaultSource.scala:168)
        at 
org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:47)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:75)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:73)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:84)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:104)
        at 
org.apache.spark.sql.catalyst.QueryPlanningTracker$.withTracker(QueryPlanningTracker.scala:107)
        at 
org.apache.spark.sql.execution.SQLExecution$.withTracker(SQLExecution.scala:250)
        at 
org.apache.spark.sql.execution.SQLExecution$.executeQuery$1(SQLExecution.scala:123)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$9(SQLExecution.scala:160)
        at 
org.apache.spark.sql.catalyst.QueryPlanningTracker$.withTracker(QueryPlanningTracker.scala:107)
        at 
org.apache.spark.sql.execution.SQLExecution$.withTracker(SQLExecution.scala:250)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$8(SQLExecution.scala:160)
        at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:271)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:159)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:827)
        at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:69)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:101)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:97)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:554)
        at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:107)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:554)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:32)
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267)
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:530)
        at 
org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:97)
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:84)
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:82)
        at 
org.apache.spark.sql.execution.QueryExecution.assertCommandExecuted(QueryExecution.scala:142)
        at 
org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:856)
        at 
org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:387)
        at 
org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:360)
        at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:239)
        .
        .
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
        at scala.util.Try$.apply(Try.scala:213)
        .
        .
        .
        .
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
        at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:1066)
        at 
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:192)
        at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:215)
        at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:91)
        at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1158)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1167)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   
   org.apache.hudi.exception.HoodieCompactionException: Could not compact 
s3://<s3 path>/.hoodie/metadata at 
org.apache.hudi.table.action.compact.RunCompactionActionExecutor.execute(RunCompactionActionExecutor.java:129)
 at 
org.apache.hudi.table.HoodieSparkMergeOnReadTable.compact(HoodieSparkMergeOnReadTable.java:155)
 at 
org.apache.hudi.client.BaseHoodieTableServiceClient.compact(BaseHoodieTableServiceClient.java:297)
 at 
org.apache.hudi.client.BaseHoodieWriteClient.compact(BaseHoodieWriteClient.java:1126)
 at 
org.apache.hudi.client.BaseHoodieWriteClient.compact(BaseHoodieWriteClient.java:1046)
 at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.compactIfNecessary(HoodieBackedTableMetadataWriter.java:1316)
 at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.performTableServices(HoodieBackedTableMetadataWriter.java:1267)
 at 
org.apache.hudi.client.SparkRDDWriteClient.initializeMetadataTable(SparkRDDWriteClient.java:323)
 at org.apache.hudi.client.SparkRDDWriteClient.initMet
 adataTable(SparkRDDWriteClient.java:288) at 
org.apache.hudi.client.BaseHoodieWriteClient.doInitTable(BaseHoodieWriteClient.java:1244)
 at 
org.apache.hudi.client.BaseHoodieWriteClient.initTable(BaseHoodieWriteClient.java:1284)
 at 
org.apache.hudi.client.SparkRDDWriteClient.upsert(SparkRDDWriteClient.java:154) 
at org.apache.hudi.DataSourceUtils.doWriteOperation(DataSourceUtils.java:225) 
at 
org.apache.hudi.HoodieSparkSqlWriterInternal.liftedTree1$1(HoodieSparkSqlWriter.scala:492)
 at 
org.apache.hudi.HoodieSparkSqlWriterInternal.writeInternal(HoodieSparkSqlWriter.scala:490)
 at 
org.apache.hudi.HoodieSparkSqlWriterInternal.write(HoodieSparkSqlWriter.scala:187)
 at org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:125) 
at org.apache.hudi.DefaultSource.createRelation(DefaultSource.scala:168) at 
org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:47)
 at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectR
 esult$lzycompute(commands.scala:75) at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:73)
 at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:84)
 at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:104)
 at 
org.apache.spark.sql.catalyst.QueryPlanningTracker$.withTracker(QueryPlanningTracker.scala:107)
 at 
org.apache.spark.sql.execution.SQLExecution$.withTracker(SQLExecution.scala:250)
 at 
org.apache.spark.sql.execution.SQLExecution$.executeQuery$1(SQLExecution.scala:123)
 at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$9(SQLExecution.scala:160)
 at 
org.apache.spark.sql.catalyst.QueryPlanningTracker$.withTracker(QueryPlanningTracker.scala:107)
 at 
org.apache.spark.sql.execution.SQLExecution$.withTracker(SQLExecution.scala:250)
 at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$8(SQ
 LExecution.scala:160) at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:271)
 at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:159)
 at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:827) at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:69)
 at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:101)
 at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:97)
 at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:554)
 at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:107)
 at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:554)
 at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans
 $logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:32) 
at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267)
 at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263)
 at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)
 at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)
 at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:530) 
at 
org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:97)
 at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:84)
 at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:82)
 at 
org.apache.spark.sql.execution.QueryExecution.assertCommandExecuted(QueryExecution.scala:142)
 at org.apache
 .spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:856) at 
org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:387) 
at org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:360) 
at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:239) 
   .
   .
   .
   sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:498) at 
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:1066)
 at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:192) at 
org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:215) at 
org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:91) at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1158) at 
org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1167) at 
org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   2025-03-11T15:04:00.242Z
   Caused by: org.apache.spark.SparkException: Job aborted due to stage 
failure: Task 2075 in stage 260.0 failed 4 times, most recent failure: Lost 
task 2075.3 in stage 260.0 (TID 1152449) (<masked ***> executor 1493): 
ExecutorLostFailure (executor 1493 exited caused by one of the running tasks) 
Reason: Executor heartbeat timed out after 131337 ms
   
   Caused by: org.apache.spark.SparkException: Job aborted due to stage 
failure: Task 2075 in stage 260.0 failed 4 times, most recent failure: Lost 
task 2075.3 in stage 260.0 (TID 1152449) (<masked ***> executor 1493): 
ExecutorLostFailure (executor 1493 exited caused by one of the running tasks) 
Reason: Executor heartbeat timed out after 131337 ms
   2025-03-11T15:04:00.242Z
   Driver stacktrace:
        at 
org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2974)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2910)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2909)
        at 
scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
        at 
scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
        at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2909)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1263)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1263)
        at scala.Option.foreach(Option.scala:407)
        at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1263)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:3173)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:3112)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:3101)
        at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
        at 
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:1028)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2271)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2292)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2311)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2336)
        at org.apache.spark.rdd.RDD.$anonfun$collect$1(RDD.scala:1019)
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
        at org.apache.spark.rdd.RDD.withScope(RDD.scala:405)
        at org.apache.spark.rdd.RDD.collect(RDD.scala:1018)
        at org.apache.spark.api.java.JavaRDDLike.collect(JavaRDDLike.scala:362)
        at org.apache.spark.api.java.JavaRDDLike.collect$(JavaRDDLike.scala:361)
        at 
org.apache.spark.api.java.AbstractJavaRDDLike.collect(JavaRDDLike.scala:45)
        at 
org.apache.hudi.data.HoodieJavaRDD.collectAsList(HoodieJavaRDD.java:177)
        at 
org.apache.hudi.table.action.compact.RunCompactionActionExecutor.execute(RunCompactionActionExecutor.java:113)
        ... 74 more
   
   Driver stacktrace: at 
org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2974)
 at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2910)
 at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2909)
 at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) at 
scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) at 
scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2909) at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1263)
 at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1263)
 at scala.Option.foreach(Option.scala:407) at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1263)
 at org.apache.spark.scheduler.DAGSchedulerEventProcess
 Loop.doOnReceive(DAGScheduler.scala:3173) at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:3112)
 at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:3101)
 at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49) at 
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:1028) at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2271) at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2292) at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2311) at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2336) at 
org.apache.spark.rdd.RDD.$anonfun$collect$1(RDD.scala:1019) at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) 
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) 
at org.apache.spark.rdd.RDD.withScope(RDD.scala:405) at 
org.apache.spark.rdd.RDD.collect(RDD.scala:1018) at 
org.apache.spark.api.java.Java
 RDDLike.collect(JavaRDDLike.scala:362) at 
org.apache.spark.api.java.JavaRDDLike.collect$(JavaRDDLike.scala:361) at 
org.apache.spark.api.java.AbstractJavaRDDLike.collect(JavaRDDLike.scala:45) at 
org.apache.hudi.data.HoodieJavaRDD.collectAsList(HoodieJavaRDD.java:177) at 
org.apache.hudi.table.action.compact.RunCompactionActionExecutor.execute(RunCompactionActionExecutor.java:113)
 ... 74 more
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 WARN YarnSchedulerBackend$YarnDriverEndpoint: Attempted to 
kill task 1152414 for unknown executor 1492.
   
   25/03/11 15:04:00 WARN YarnSchedulerBackend$YarnDriverEndpoint: Attempted to 
kill task 1152414 for unknown executor 1492.
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO TransactionManager: Transaction ending with 
transaction owner Option{val=[==>20250311130120164__commit__INFLIGHT]}
   
   25/03/11 15:04:00 INFO TransactionManager: Transaction ending with 
transaction owner Option{val=[==>20250311130120164__commit__INFLIGHT]}
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO InProcessLockProvider: Base Path s3://<s3 path to 
hoodie table>, Lock Instance 
java.util.concurrent.locks.ReentrantReadWriteLock@1212c905[Write locks = 1, 
Read locks = 0], Thread main, In-process lock state RELEASING
   
   25/03/11 15:04:00 INFO InProcessLockProvider: Base Path s3://<s3 path to 
hoodie table>, Lock Instance 
java.util.concurrent.locks.ReentrantReadWriteLock@1212c905[Write locks = 1, 
Read locks = 0], Thread main, In-process lock state RELEASING
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO InProcessLockProvider: Base Path s3://<s3 path to 
hoodie table>, Lock Instance 
java.util.concurrent.locks.ReentrantReadWriteLock@1212c905[Write locks = 0, 
Read locks = 0], Thread main, In-process lock state RELEASED
   
   25/03/11 15:04:00 INFO InProcessLockProvider: Base Path s3://<s3 path to 
hoodie table>, Lock Instance 
java.util.concurrent.locks.ReentrantReadWriteLock@1212c905[Write locks = 0, 
Read locks = 0], Thread main, In-process lock state RELEASED
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO InProcessLockProvider: Base Path s3://<s3 path to 
hoodie table>, Lock Instance 
java.util.concurrent.locks.ReentrantReadWriteLock@1212c905[Write locks = 0, 
Read locks = 0], Thread main, In-process lock state ALREADY_RELEASED
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO LockManager: Released connection created for 
acquiring lock
   
   25/03/11 15:04:00 INFO LockManager: Released connection created for 
acquiring lock
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO TransactionManager: Transaction ended with 
transaction owner Option{val=[==>20250311130120164__commit__INFLIGHT]}
   
   25/03/11 15:04:00 INFO TransactionManager: Transaction ended with 
transaction owner Option{val=[==>20250311130120164__commit__INFLIGHT]}
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO HoodieSparkSqlWriterInternal: 
Config.inlineCompactionEnabled ? false
   
   25/03/11 15:04:00 INFO HoodieSparkSqlWriterInternal: 
Config.inlineCompactionEnabled ? false
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO HoodieSparkSqlWriterInternal: 
Config.asyncClusteringEnabled ? false
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 WARN HoodieSparkSqlWriterInternal: Closing write client
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO HoodieHeartbeatClient: Stopping heartbeat for instant 
20250311130120164
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO HoodieHeartbeatClient: Stopped heartbeat for instant 
20250311130120164
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO TransactionManager: Transaction manager closed
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO TransactionManager: Transaction manager closed
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 WARN YarnSchedulerBackend$YarnDriverEndpoint: Attempted to 
kill task 1152419 for unknown executor 1492.
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 WARN YarnSchedulerBackend$YarnDriverEndpoint: Attempted to 
kill task 1152417 for unknown executor 1492.
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 WARN YarnSchedulerBackend$YarnDriverEndpoint: Attempted to 
kill task 1152412 for unknown executor 1492.
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 WARN YarnSchedulerBackend$YarnDriverEndpoint: Attempted to 
kill task 1152416 for unknown executor 1492.
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 WARN YarnSchedulerBackend$YarnDriverEndpoint: Attempted to 
kill task 1152415 for unknown executor 1492.
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 WARN YarnSchedulerBackend$YarnDriverEndpoint: Attempted to 
kill task 1152413 for unknown executor 1492.
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 WARN YarnSchedulerBackend$YarnDriverEndpoint: Attempted to 
kill task 1152418 for unknown executor 1492.
   2025-03-11T15:04:00.242Z
   Exception in thread "main" org.apache.hudi.exception.HoodieException: Failed 
to instantiate Metadata table 
        at 
org.apache.hudi.client.SparkRDDWriteClient.initializeMetadataTable(SparkRDDWriteClient.java:326)
        at 
org.apache.hudi.client.SparkRDDWriteClient.initMetadataTable(SparkRDDWriteClient.java:288)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.doInitTable(BaseHoodieWriteClient.java:1244)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.initTable(BaseHoodieWriteClient.java:1284)
        at 
org.apache.hudi.client.SparkRDDWriteClient.upsert(SparkRDDWriteClient.java:154)
        at 
org.apache.hudi.DataSourceUtils.doWriteOperation(DataSourceUtils.java:225)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.liftedTree1$1(HoodieSparkSqlWriter.scala:492)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.writeInternal(HoodieSparkSqlWriter.scala:490)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.write(HoodieSparkSqlWriter.scala:187)
        at 
org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:125)
        at org.apache.hudi.DefaultSource.createRelation(DefaultSource.scala:168)
        at 
org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:47)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:75)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:73)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:84)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:104)
        at 
org.apache.spark.sql.catalyst.QueryPlanningTracker$.withTracker(QueryPlanningTracker.scala:107)
        at 
org.apache.spark.sql.execution.SQLExecution$.withTracker(SQLExecution.scala:250)
        at 
org.apache.spark.sql.execution.SQLExecution$.executeQuery$1(SQLExecution.scala:123)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$9(SQLExecution.scala:160)
        at 
org.apache.spark.sql.catalyst.QueryPlanningTracker$.withTracker(QueryPlanningTracker.scala:107)
        at 
org.apache.spark.sql.execution.SQLExecution$.withTracker(SQLExecution.scala:250)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$8(SQLExecution.scala:160)
        at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:271)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:159)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:827)
        at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:69)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:101)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:97)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:554)
        at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:107)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:554)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:32)
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267)
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:530)
        at 
org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:97)
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:84)
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:82)
        at 
org.apache.spark.sql.execution.QueryExecution.assertCommandExecuted(QueryExecution.scala:142)
        at 
org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:856)
        at 
org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:387)
        at 
org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:360)
        at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:239)
        .
        .
        .
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
        at scala.util.Try$.apply(Try.scala:213)
        .
        .
        .
        .
   
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
        at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:1066)
        at 
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:192)
        at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:215)
        at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:91)
        at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1158)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1167)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   
   Exception in thread "main" org.apache.hudi.exception.HoodieException: Failed 
to instantiate Metadata table at 
org.apache.hudi.client.SparkRDDWriteClient.initializeMetadataTable(SparkRDDWriteClient.java:326)
 at 
org.apache.hudi.client.SparkRDDWriteClient.initMetadataTable(SparkRDDWriteClient.java:288)
 at 
org.apache.hudi.client.BaseHoodieWriteClient.doInitTable(BaseHoodieWriteClient.java:1244)
 at 
org.apache.hudi.client.BaseHoodieWriteClient.initTable(BaseHoodieWriteClient.java:1284)
 at 
org.apache.hudi.client.SparkRDDWriteClient.upsert(SparkRDDWriteClient.java:154) 
at org.apache.hudi.DataSourceUtils.doWriteOperation(DataSourceUtils.java:225) 
at 
org.apache.hudi.HoodieSparkSqlWriterInternal.liftedTree1$1(HoodieSparkSqlWriter.scala:492)
 at 
org.apache.hudi.HoodieSparkSqlWriterInternal.writeInternal(HoodieSparkSqlWriter.scala:490)
 at 
org.apache.hudi.HoodieSparkSqlWriterInternal.write(HoodieSparkSqlWriter.scala:187)
 at org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:1
 25) at org.apache.hudi.DefaultSource.createRelation(DefaultSource.scala:168) 
at 
org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:47)
 at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:75)
 at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:73)
 at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:84)
 at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:104)
 at 
org.apache.spark.sql.catalyst.QueryPlanningTracker$.withTracker(QueryPlanningTracker.scala:107)
 at 
org.apache.spark.sql.execution.SQLExecution$.withTracker(SQLExecution.scala:250)
 at 
org.apache.spark.sql.execution.SQLExecution$.executeQuery$1(SQLExecution.scala:123)
 at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$9(SQLExecution.scala:160)
 a
 t 
org.apache.spark.sql.catalyst.QueryPlanningTracker$.withTracker(QueryPlanningTracker.scala:107)
 at 
org.apache.spark.sql.execution.SQLExecution$.withTracker(SQLExecution.scala:250)
 at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$8(SQLExecution.scala:160)
 at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:271)
 at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:159)
 at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:827) at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:69)
 at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:101)
 at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:97)
 at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:554)
 at org.apa
 che.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:107) at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:554)
 at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:32)
 at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267)
 at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263)
 at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)
 at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)
 at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:530) 
at 
org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:97)
 at org.apache.spark.sql.execution.Query
 Execution.commandExecuted$lzycompute(QueryExecution.scala:84) at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:82)
 at 
org.apache.spark.sql.execution.QueryExecution.assertCommandExecuted(QueryExecution.scala:142)
 at org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:856) 
at 
org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:387) 
at org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:360) 
at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:239) 
   .
   .
   .
   .
   .
   .
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:498) at 
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52) at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:1066)
 at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:192) at 
org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:215) at 
org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:91) at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1158) at 
org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1167) at 
org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   2025-03-11T15:04:00.242Z
   Caused by: org.apache.hudi.exception.HoodieCompactionException: Could not 
compact s3://<s3 path>/.hoodie/metadata
        at 
org.apache.hudi.table.action.compact.RunCompactionActionExecutor.execute(RunCompactionActionExecutor.java:129)
        at 
org.apache.hudi.table.HoodieSparkMergeOnReadTable.compact(HoodieSparkMergeOnReadTable.java:155)
        at 
org.apache.hudi.client.BaseHoodieTableServiceClient.compact(BaseHoodieTableServiceClient.java:297)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.compact(BaseHoodieWriteClient.java:1126)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.compact(BaseHoodieWriteClient.java:1046)
        at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.compactIfNecessary(HoodieBackedTableMetadataWriter.java:1316)
        at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.performTableServices(HoodieBackedTableMetadataWriter.java:1267)
        at 
org.apache.hudi.client.SparkRDDWriteClient.initializeMetadataTable(SparkRDDWriteClient.java:323)
        ... 67 more
   
   Caused by: org.apache.hudi.exception.HoodieCompactionException: Could not 
compact s3://<s3 path>/.hoodie/metadata at 
org.apache.hudi.table.action.compact.RunCompactionActionExecutor.execute(RunCompactionActionExecutor.java:129)
 at 
org.apache.hudi.table.HoodieSparkMergeOnReadTable.compact(HoodieSparkMergeOnReadTable.java:155)
 at 
org.apache.hudi.client.BaseHoodieTableServiceClient.compact(BaseHoodieTableServiceClient.java:297)
 at 
org.apache.hudi.client.BaseHoodieWriteClient.compact(BaseHoodieWriteClient.java:1126)
 at 
org.apache.hudi.client.BaseHoodieWriteClient.compact(BaseHoodieWriteClient.java:1046)
 at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.compactIfNecessary(HoodieBackedTableMetadataWriter.java:1316)
 at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.performTableServices(HoodieBackedTableMetadataWriter.java:1267)
 at 
org.apache.hudi.client.SparkRDDWriteClient.initializeMetadataTable(SparkRDDWriteClient.java:323)
 ... 67 more
   2025-03-11T15:04:00.242Z
   Caused by: org.apache.spark.SparkException: Job aborted due to stage 
failure: Task 2075 in stage 260.0 failed 4 times, most recent failure: Lost 
task 2075.3 in stage 260.0 (TID 1152449) (<masked ***> executor 1493): 
ExecutorLostFailure (executor 1493 exited caused by one of the running tasks) 
Reason: Executor heartbeat timed out after 131337 ms
   
   Caused by: org.apache.spark.SparkException: Job aborted due to stage 
failure: Task 2075 in stage 260.0 failed 4 times, most recent failure: Lost 
task 2075.3 in stage 260.0 (TID 1152449) (<masked ***> executor 1493): 
ExecutorLostFailure (executor 1493 exited caused by one of the running tasks) 
Reason: Executor heartbeat timed out after 131337 ms
   2025-03-11T15:04:00.242Z
   Driver stacktrace:
        at 
org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2974)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2910)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2909)
        at 
scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
        at 
scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
        at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2909)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1263)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1263)
        at scala.Option.foreach(Option.scala:407)
        at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1263)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:3173)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:3112)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:3101)
        at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
        at 
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:1028)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2271)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2292)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2311)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2336)
        at org.apache.spark.rdd.RDD.$anonfun$collect$1(RDD.scala:1019)
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
        at org.apache.spark.rdd.RDD.withScope(RDD.scala:405)
        at org.apache.spark.rdd.RDD.collect(RDD.scala:1018)
        at org.apache.spark.api.java.JavaRDDLike.collect(JavaRDDLike.scala:362)
        at org.apache.spark.api.java.JavaRDDLike.collect$(JavaRDDLike.scala:361)
        at 
org.apache.spark.api.java.AbstractJavaRDDLike.collect(JavaRDDLike.scala:45)
        at 
org.apache.hudi.data.HoodieJavaRDD.collectAsList(HoodieJavaRDD.java:177)
        at 
org.apache.hudi.table.action.compact.RunCompactionActionExecutor.execute(RunCompactionActionExecutor.java:113)
        ... 74 more
   
   Driver stacktrace: at 
org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2974)
 at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2910)
 at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2909)
 at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) at 
scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) at 
scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2909) at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1263)
 at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1263)
 at scala.Option.foreach(Option.scala:407) at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1263)
 at org.apache.spark.scheduler.DAGSchedulerEventProcess
 Loop.doOnReceive(DAGScheduler.scala:3173) at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:3112)
 at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:3101)
 at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49) at 
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:1028) at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2271) at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2292) at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2311) at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:2336) at 
org.apache.spark.rdd.RDD.$anonfun$collect$1(RDD.scala:1019) at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) 
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) 
at org.apache.spark.rdd.RDD.withScope(RDD.scala:405) at 
org.apache.spark.rdd.RDD.collect(RDD.scala:1018) at 
org.apache.spark.api.java.Java
 RDDLike.collect(JavaRDDLike.scala:362) at 
org.apache.spark.api.java.JavaRDDLike.collect$(JavaRDDLike.scala:361) at 
org.apache.spark.api.java.AbstractJavaRDDLike.collect(JavaRDDLike.scala:45) at 
org.apache.hudi.data.HoodieJavaRDD.collectAsList(HoodieJavaRDD.java:177) at 
org.apache.hudi.table.action.compact.RunCompactionActionExecutor.execute(RunCompactionActionExecutor.java:113)
 ... 74 more
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO SparkContext: Invoking stop() from shutdown hook
   2025-03-11T15:04:00.242Z
   25/03/11 15:04:00 INFO SparkContext: SparkContext is stopping with exitCode 
0.
   
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to