sstimmel opened a new issue, #7409:
URL: https://github.com/apache/hudi/issues/7409
See the Following exception in the logs after running deltastreamer. I
believe this is shown when we have metrics enabled.
java.lang.NoSuchMethodError: 'int
org.apache.hudi.avro.model.HoodieRollbackMetadata.getTotalFilesDeleted()'
A clear and concise description of what you expected to happen.
**Environment Description**
* Hudi version : 0.12.1
* Spark version : 3.3.1
* Hive version :
* Hadoop version : 3
* Storage (HDFS/S3/GCS..) : S3
* Running on Docker? (yes/no) : yes, kubernetes
**Additional context**
Add any other context about the problem here.
The following hudi jars are in the classpath
-hudi-aws-bundle.jar
- hudi-cli.jar
-hudi-hive-sync-bundle.jar
-hudi-spark-bundle.jar
-hudi-utilities-slim-bundle.jar
hudi-utilities-slim-bundle does show that class.
10851 Mon Oct 17 22:29:30 UTC 2022
org/apache/hudi/avro/model/HoodieRollbackMetadata.class
**Stacktrace**
```22/12/07 16:27:35 INFO org.apache.spark.internal.Logging: Disabling
executor 1.
22/12/07 16:27:35 DEBUG org.apache.spark.internal.Logging: Executor 1 on
10.22.164.85 lost, but reason not yet known.
22/12/07 16:27:35 DEBUG org.apache.spark.internal.Logging: Cleaning up
TaskScheduler state for tasks [] on failed executor 1
22/12/07 16:27:35 ERROR org.apache.spark.internal.Logging: Uncaught
exception in thread dispatcher-CoarseGrainedScheduler
org.apache.spark.SparkException: Could not find CoarseGrainedScheduler.
at
org.apache.spark.rpc.netty.Dispatcher.postMessage(Dispatcher.scala:178)
at
org.apache.spark.rpc.netty.Dispatcher.postOneWayMessage(Dispatcher.scala:150)
at org.apache.spark.rpc.netty.NettyRpcEnv.send(NettyRpcEnv.scala:193)
at
org.apache.spark.rpc.netty.NettyRpcEndpointRef.send(NettyRpcEnv.scala:563)
at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.$anonfun$reviveOffers$1(CoarseGrainedSchedulerBackend.scala:630)
at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1484)
at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.reviveOffers(CoarseGrainedSchedulerBackend.scala:630)
at
org.apache.spark.scheduler.TaskSchedulerImpl.executorLost(TaskSchedulerImpl.scala:1004)
at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend$DriverEndpoint.disableExecutor(CoarseGrainedSchedulerBackend.scala:482)
at
org.apache.spark.scheduler.cluster.k8s.KubernetesClusterSchedulerBackend$KubernetesDriverEndpoint.$anonfun$onDisconnected$1(KubernetesClusterSchedulerBackend.scala:328)
at
org.apache.spark.scheduler.cluster.k8s.KubernetesClusterSchedulerBackend$KubernetesDriverEndpoint.$anonfun$onDisconnected$1$adapted(KubernetesClusterSchedulerBackend.scala:328)
at scala.Option.foreach(Option.scala:407)
at
org.apache.spark.scheduler.cluster.k8s.KubernetesClusterSchedulerBackend$KubernetesDriverEndpoint.onDisconnected(KubernetesClusterSchedulerBackend.scala:328)
at org.apache.spark.rpc.netty.Inbox.$anonfun$process$1(Inbox.scala:141)
at org.apache.spark.rpc.netty.Inbox.safelyCall(Inbox.scala:213)
at org.apache.spark.rpc.netty.Inbox.process(Inbox.scala:100)
at
org.apache.spark.rpc.netty.MessageLoop.org$apache$spark$rpc$netty$MessageLoop$$receiveLoop(MessageLoop.scala:75)
at
org.apache.spark.rpc.netty.MessageLoop$$anon$1.run(MessageLoop.scala:41)
at
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
22/12/07 16:27:35 WARN org.apache.spark.internal.Logging: Kubernetes client
has been closed.
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging:
MapOutputTrackerMasterEndpoint stopped!
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: MemoryStore cleared
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: BlockManager
stopped
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: BlockManagerMaster
stopped
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging:
OutputCommitCoordinator stopped!
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Successfully
stopped SparkContext
Exception in thread "main" java.lang.NoSuchMethodError: 'int
org.apache.hudi.avro.model.HoodieRollbackMetadata.getTotalFilesDeleted()'
at
org.apache.hudi.client.BaseHoodieWriteClient.rollback(BaseHoodieWriteClient.java:774)
at
org.apache.hudi.client.BaseHoodieWriteClient.rollbackFailedWrites(BaseHoodieWriteClient.java:1194)
at
org.apache.hudi.client.BaseHoodieWriteClient.rollbackFailedWrites(BaseHoodieWriteClient.java:1177)
at
org.apache.hudi.client.BaseHoodieWriteClient.rollbackFailedWrites(BaseHoodieWriteClient.java:1165)
at
org.apache.hudi.client.BaseHoodieWriteClient.lambda$startCommitWithTime$97cdbdca$1(BaseHoodieWriteClient.java:966)
at
org.apache.hudi.common.util.CleanerUtils.rollbackFailedWrites(CleanerUtils.java:153)
at
org.apache.hudi.client.BaseHoodieWriteClient.startCommitWithTime(BaseHoodieWriteClient.java:965)
at
org.apache.hudi.client.BaseHoodieWriteClient.startCommitWithTime(BaseHoodieWriteClient.java:958)
at
org.apache.hudi.utilities.deltastreamer.DeltaSync.startCommit(DeltaSync.java:677)
at
org.apache.hudi.utilities.deltastreamer.DeltaSync.writeToSink(DeltaSync.java:581)
at
org.apache.hudi.utilities.deltastreamer.DeltaSync.syncOnce(DeltaSync.java:336)
at
org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.lambda$sync$2(HoodieDeltaStreamer.java:204)
at org.apache.hudi.common.util.Option.ifPresent(Option.java:97)
at
org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.sync(HoodieDeltaStreamer.java:202)
at
org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer.main(HoodieDeltaStreamer.java:571)
at
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at
org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
at
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:958)
at
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
at
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1046)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1055)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Shutdown hook
called
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Deleting directory
/tmp/spark-7a432e99-1f82-46a4-a555-10bd6090a2ba
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Deleting directory
/tmp/spark-42983bb3-9835-4bf4-9ea5-c68004077e7c
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Deleting directory
/tmp/spark-6098afd8-d431-4b91-a81e-3ebea9419630
22/12/07 16:27:36 INFO org.apache.spark.internal.Logging: Deleting directory
/tmp/spark-f394dce8-567a-461d-b559-c953966adb93
22/12/07 16:27:36 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl:
Stopping s3a-file-system metrics system...
22/12/07 16:27:36 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl:
s3a-file-system metrics system stopped.
22/12/07 16:27:36 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl:
s3a-file-system metrics system shutdown complete.```
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]