Jihwan Lee created HUDI-7763:
--------------------------------

             Summary: Fix that jmx reporter cannot initialized if metadata 
enables
                 Key: HUDI-7763
                 URL: https://issues.apache.org/jira/browse/HUDI-7763
             Project: Apache Hudi
          Issue Type: Bug
         Environment: hudi0.14.1, Spark3.2
            Reporter: Jihwan Lee


If the jmx metric option is activated, port settings can be set to range.

 

Because metadata is also written as hoodie table, requires multiple metric 
instances. (If not, occur exception 'ObjID already in use')

JmxReporterServer can only use one port each.

So, jmx server might be able to be initialized on multiple ports.

 

error log:

 ( jmx reporter for metadata is initialized first, then reporter for data 
occurs exception )
{code:java}
24/05/13 20:28:27 INFO table.HoodieTableMetaClient: Loading 
HoodieTableMetaClient from /data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:27 INFO table.HoodieTableConfig: Loading table properties from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/hoodie.properties
24/05/13 20:28:27 INFO table.HoodieTableMetaClient: Finished Loading Table of 
type MERGE_ON_READ(version=1, baseFileFormat=PARQUET) from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:27 INFO table.HoodieTableMetaClient: Loading 
HoodieTableMetaClient from /data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:27 INFO table.HoodieTableConfig: Loading table properties from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/hoodie.properties
24/05/13 20:28:27 INFO table.HoodieTableMetaClient: Finished Loading Table of 
type MERGE_ON_READ(version=1, baseFileFormat=PARQUET) from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:27 INFO timeline.HoodieActiveTimeline: Loaded instants upto : 
Option{val=[==>20240513195519782__deltacommit__REQUESTED__20240513195521160]}
24/05/13 20:28:28 INFO config.HoodieWriteConfig: Automatically set 
hoodie.cleaner.policy.failed.writes=LAZY since optimistic concurrency control 
is used
24/05/13 20:28:28 INFO metrics.JmxMetricsReporter: Started JMX server on port 
9889.
24/05/13 20:28:28 INFO metrics.JmxMetricsReporter: Configured JMXReporter with 
{port:9889}
24/05/13 20:28:28 INFO embedded.EmbeddedTimelineService: Overriding hostIp to 
(feeder-affiliate-book-svc-sink-09c3c08f71b47a5d-driver-svc.csp.svc) found in 
spark-conf. It was null
24/05/13 20:28:28 INFO view.FileSystemViewManager: Creating View Manager with 
storage type :MEMORY
24/05/13 20:28:28 INFO view.FileSystemViewManager: Creating in-memory based 
Table View
24/05/13 20:28:28 INFO util.log: Logging initialized @53678ms to 
org.apache.hudi.org.apache.jetty.util.log.Slf4jLog
24/05/13 20:28:28 INFO javalin.Javalin:
       __                      __ _            __ __
      / /____ _ _   __ ____ _ / /(_)____      / // /
 __  / // __ `/| | / // __ `// // // __ \    / // /_
/ /_/ // /_/ / | |/ // /_/ // // // / / /   /__  __/
\____/ \__,_/  |___/ \__,_//_//_//_/ /_/      /_/


          https://javalin.io/documentation


24/05/13 20:28:28 INFO javalin.Javalin: Starting Javalin ...
24/05/13 20:28:28 INFO javalin.Javalin: You are running Javalin 4.6.7 (released 
October 24, 2022. Your Javalin version is 567 days old. Consider checking for a 
newer version.).
24/05/13 20:28:28 INFO server.Server: jetty-9.4.48.v20220622; built: 
2022-06-21T20:42:25.880Z; git: 6b67c5719d1f4371b33655ff2d047d24e171e49a; jvm 
11.0.20.1+1
24/05/13 20:28:28 INFO server.Server: Started @54065ms
24/05/13 20:28:28 INFO javalin.Javalin: Listening on http://localhost:35071/
24/05/13 20:28:28 INFO javalin.Javalin: Javalin started in 177ms \o/
24/05/13 20:28:28 INFO service.TimelineService: Starting Timeline server on 
port :35071
24/05/13 20:28:28 INFO embedded.EmbeddedTimelineService: Started embedded 
timeline server at 
feeder-affiliate-book-svc-sink-09c3c08f71b47a5d-driver-svc.csp.svc:35071
24/05/13 20:28:28 INFO client.BaseHoodieClient: Timeline Server already 
running. Not restarting the service
24/05/13 20:28:28 INFO hudi.HoodieSparkSqlWriterInternal: 
Config.inlineCompactionEnabled ? true
24/05/13 20:28:28 INFO hudi.HoodieSparkSqlWriterInternal: 
Config.asyncClusteringEnabled ? false
24/05/13 20:28:28 INFO codegen.CodeGenerator: Code generated in 77.42891 ms
24/05/13 20:28:28 INFO table.HoodieTableMetaClient: Loading 
HoodieTableMetaClient from /data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:28 INFO table.HoodieTableConfig: Loading table properties from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/hoodie.properties
24/05/13 20:28:28 INFO table.HoodieTableMetaClient: Finished Loading Table of 
type MERGE_ON_READ(version=1, baseFileFormat=PARQUET) from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:28 INFO table.HoodieTableMetaClient: Loading Active commit 
timeline for /data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:28 INFO timeline.HoodieActiveTimeline: Loaded instants upto : 
Option{val=[==>20240513195519782__deltacommit__REQUESTED__20240513195521160]}
24/05/13 20:28:28 INFO client.BaseHoodieWriteClient: Generate a new instant 
time: 20240513202827651 action: deltacommit
24/05/13 20:28:28 INFO heartbeat.HoodieHeartbeatClient: Received request to 
start heartbeat for instant time 20240513202827651
24/05/13 20:28:28 INFO timeline.HoodieActiveTimeline: Creating a new instant 
[==>20240513202827651__deltacommit__REQUESTED]
24/05/13 20:28:28 INFO table.HoodieTableMetaClient: Loading 
HoodieTableMetaClient from /data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:28 INFO table.HoodieTableConfig: Loading table properties from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/hoodie.properties
24/05/13 20:28:28 INFO table.HoodieTableMetaClient: Finished Loading Table of 
type MERGE_ON_READ(version=1, baseFileFormat=PARQUET) from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:28 INFO table.HoodieTableMetaClient: Loading Active commit 
timeline for /data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:28 INFO timeline.HoodieActiveTimeline: Loaded instants upto : 
Option{val=[==>20240513202827651__deltacommit__REQUESTED__20240513202828906]}
24/05/13 20:28:28 INFO transaction.TransactionManager: Transaction starting for 
Option{val=[==>20240513202827651__deltacommit__INFLIGHT]} with latest completed 
transaction instant Optional.empty
24/05/13 20:28:28 INFO lock.LockManager: LockProvider 
org.apache.hudi.client.transaction.lock.FileSystemBasedLockProvider
24/05/13 20:28:29 INFO transaction.TransactionManager: Transaction started for 
Option{val=[==>20240513202827651__deltacommit__INFLIGHT]} with latest completed 
transaction instant Optional.empty
24/05/13 20:28:29 INFO table.HoodieTableMetaClient: Loading 
HoodieTableMetaClient from /data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:29 INFO table.HoodieTableConfig: Loading table properties from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/hoodie.properties
24/05/13 20:28:29 INFO table.HoodieTableMetaClient: Finished Loading Table of 
type MERGE_ON_READ(version=1, baseFileFormat=PARQUET) from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:29 INFO table.HoodieTableMetaClient: Loading 
HoodieTableMetaClient from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/metadata
24/05/13 20:28:29 INFO table.HoodieTableConfig: Loading table properties from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/metadata/.hoodie/hoodie.properties
24/05/13 20:28:29 INFO table.HoodieTableMetaClient: Finished Loading Table of 
type MERGE_ON_READ(version=1, baseFileFormat=HFILE) from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/metadata
24/05/13 20:28:29 INFO timeline.HoodieActiveTimeline: Loaded instants upto : 
Option{val=[20240513182607590__deltacommit__COMPLETED__20240513182612582]}
24/05/13 20:28:29 INFO table.HoodieTableMetaClient: Loading 
HoodieTableMetaClient from /data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:29 INFO table.HoodieTableConfig: Loading table properties from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/hoodie.properties
24/05/13 20:28:29 INFO table.HoodieTableMetaClient: Finished Loading Table of 
type MERGE_ON_READ(version=1, baseFileFormat=PARQUET) from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc
24/05/13 20:28:29 INFO table.HoodieTableMetaClient: Loading 
HoodieTableMetaClient from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/metadata
24/05/13 20:28:29 INFO table.HoodieTableConfig: Loading table properties from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/metadata/.hoodie/hoodie.properties
24/05/13 20:28:29 INFO table.HoodieTableMetaClient: Finished Loading Table of 
type MERGE_ON_READ(version=1, baseFileFormat=HFILE) from 
/data/feeder/affiliate/book/affiliate_feeder_book_svc/.hoodie/metadata
24/05/13 20:28:29 INFO timeline.HoodieActiveTimeline: Loaded instants upto : 
Option{val=[20240513182607590__deltacommit__COMPLETED__20240513182612582]}
24/05/13 20:28:29 INFO view.AbstractTableFileSystemView: Took 1 ms to read  0 
instants, 0 replaced file groups
24/05/13 20:28:29 INFO util.ClusteringUtils: Found 0 files in pending 
clustering operations
24/05/13 20:28:29 ERROR metrics.JmxMetricsReporter: Jmx initialize failed:
org.apache.hudi.exception.HoodieException: Jmx service url created 
service:jmx:rmi://localhost:9889/jndi/rmi://localhost:9889/jmxrmi
        at 
org.apache.hudi.metrics.JmxReporterServer.<init>(JmxReporterServer.java:107)
        at 
org.apache.hudi.metrics.JmxReporterServer$Builder.build(JmxReporterServer.java:88)
        at 
org.apache.hudi.metrics.JmxMetricsReporter.createJmxReport(JmxMetricsReporter.java:104)
        at 
org.apache.hudi.metrics.JmxMetricsReporter.<init>(JmxMetricsReporter.java:58)
        at 
org.apache.hudi.metrics.MetricsReporterFactory.createReporter(MetricsReporterFactory.java:75)
        at org.apache.hudi.metrics.Metrics.<init>(Metrics.java:58)
        at org.apache.hudi.metrics.Metrics.getInstance(Metrics.java:82)
        at org.apache.hudi.metrics.HoodieMetrics.<init>(HoodieMetrics.java:95)
        at 
org.apache.hudi.client.BaseHoodieClient.<init>(BaseHoodieClient.java:96)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.<init>(BaseHoodieWriteClient.java:163)
        at 
org.apache.hudi.client.SparkRDDWriteClient.<init>(SparkRDDWriteClient.java:86)
        at 
org.apache.hudi.client.SparkRDDWriteClient.<init>(SparkRDDWriteClient.java:75)
        at 
org.apache.hudi.metadata.SparkHoodieBackedTableMetadataWriter.initializeWriteClient(SparkHoodieBackedTableMetadataWriter.java:146)
        at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.getWriteClient(HoodieBackedTableMetadataWriter.java:1474)
        at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.performTableServices(HoodieBackedTableMetadataWriter.java:1186)
        at 
org.apache.hudi.client.SparkRDDWriteClient.initializeMetadataTable(SparkRDDWriteClient.java:290)
        at 
org.apache.hudi.client.SparkRDDWriteClient.initMetadataTable(SparkRDDWriteClient.java:273)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.doInitTable(BaseHoodieWriteClient.java:1257)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.initTable(BaseHoodieWriteClient.java:1297)
        at 
org.apache.hudi.client.SparkRDDWriteClient.upsert(SparkRDDWriteClient.java:139)
        at 
org.apache.hudi.DataSourceUtils.doWriteOperation(DataSourceUtils.java:224)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.liftedTree1$1(HoodieSparkSqlWriter.scala:504)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.writeInternal(HoodieSparkSqlWriter.scala:502)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.write(HoodieSparkSqlWriter.scala:204)
        at 
org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:121)
        at org.apache.hudi.DefaultSource.createRelation(DefaultSource.scala:150)
        at 
org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:45)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:75)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:73)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:84)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:97)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
        at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)
        at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:97)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:93)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:481)
        at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:82)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:481)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:30)
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267)
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:457)
        at 
org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:93)
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:80)
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:78)
        at 
org.apache.spark.sql.execution.QueryExecution.assertCommandExecuted(QueryExecution.scala:115)
        at 
org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:848)
        at 
org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:382)
        at 
org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:355)
        at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:239)
        at 
com.naver.csp.data.cores.common.spark.datasets.hudi.HudiWriter$DataFrameWriterImplicits.hudi(HudiWriter.scala:35)
        at 
com.naver.csp.data.cores.common.spark.datasets.kafka.KafkaHudiSink.batchAsCdc(KafkaHudiSink.scala:160)
        at 
com.naver.csp.data.cores.common.spark.datasets.kafka.KafkaHudiSink.$anonfun$startCdcStreaming$1(KafkaHudiSink.scala:78)
        at 
com.naver.csp.data.cores.common.spark.datasets.kafka.KafkaHudiSink.$anonfun$startCdcStreaming$1$adapted(KafkaHudiSink.scala:78)
        at 
org.apache.spark.sql.execution.streaming.sources.ForeachBatchSink.addBatch(ForeachBatchSink.scala:35)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runBatch$17(MicroBatchExecution.scala:600)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
        at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)
        at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runBatch$16(MicroBatchExecution.scala:598)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken(ProgressReporter.scala:375)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken$(ProgressReporter.scala:373)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:69)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.runBatch(MicroBatchExecution.scala:598)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runActivatedStream$2(MicroBatchExecution.scala:228)
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken(ProgressReporter.scala:375)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken$(ProgressReporter.scala:373)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:69)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runActivatedStream$1(MicroBatchExecution.scala:193)
        at 
org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:57)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.runActivatedStream(MicroBatchExecution.scala:187)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.$anonfun$runStream$1(StreamExecution.scala:303)
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runStream(StreamExecution.scala:286)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:209)
Caused by: java.rmi.server.ExportException: internal error: ObjID already in use
        at 
java.rmi/sun.rmi.transport.ObjectTable.putTarget(ObjectTable.java:184)
        at java.rmi/sun.rmi.transport.Transport.exportObject(Transport.java:106)
        at 
java.rmi/sun.rmi.transport.tcp.TCPTransport.exportObject(TCPTransport.java:254)
        at 
java.rmi/sun.rmi.transport.tcp.TCPEndpoint.exportObject(TCPEndpoint.java:412)
        at java.rmi/sun.rmi.transport.LiveRef.exportObject(LiveRef.java:147)
        at 
java.rmi/sun.rmi.server.UnicastServerRef.exportObject(UnicastServerRef.java:234)
        at java.rmi/sun.rmi.registry.RegistryImpl.setup(RegistryImpl.java:220)
        at java.rmi/sun.rmi.registry.RegistryImpl.<init>(RegistryImpl.java:205)
        at 
java.rmi/java.rmi.registry.LocateRegistry.createRegistry(LocateRegistry.java:203)
        at 
org.apache.hudi.metrics.JmxReporterServer.<init>(JmxReporterServer.java:104)
        ... 83 more
24/05/13 20:28:29 INFO transaction.TransactionManager: Transaction ending with 
transaction owner Option{val=[==>20240513202827651__deltacommit__INFLIGHT]}
24/05/13 20:28:29 INFO lock.LockManager: Released connection created for 
acquiring lock
24/05/13 20:28:29 INFO transaction.TransactionManager: Transaction ended with 
transaction owner Option{val=[==>20240513202827651__deltacommit__INFLIGHT]}
24/05/13 20:28:29 INFO hudi.HoodieSparkSqlWriterInternal: 
Config.inlineCompactionEnabled ? true
24/05/13 20:28:29 INFO hudi.HoodieSparkSqlWriterInternal: 
Config.asyncClusteringEnabled ? false
24/05/13 20:28:29 WARN hudi.HoodieSparkSqlWriterInternal: Closing write client
24/05/13 20:28:29 INFO client.BaseHoodieClient: Stopping Timeline service !!
24/05/13 20:28:29 INFO embedded.EmbeddedTimelineService: Closing Timeline server
24/05/13 20:28:29 INFO service.TimelineService: Closing Timeline Service
24/05/13 20:28:29 INFO javalin.Javalin: Stopping Javalin ...
24/05/13 20:28:29 INFO javalin.Javalin: Javalin has stopped
24/05/13 20:28:29 INFO service.TimelineService: Closed Timeline Service
24/05/13 20:28:29 INFO embedded.EmbeddedTimelineService: Closed Timeline server
24/05/13 20:28:29 INFO heartbeat.HoodieHeartbeatClient: Stopping heartbeat for 
instant 20240513202827651
24/05/13 20:28:29 INFO heartbeat.HoodieHeartbeatClient: Stopped heartbeat for 
instant 20240513202827651
24/05/13 20:28:29 INFO transaction.TransactionManager: Transaction manager 
closed
24/05/13 20:28:29 INFO transaction.TransactionManager: Transaction manager 
closed
24/05/13 20:28:29 ERROR streaming.MicroBatchExecution: Query [id = 
7b545afa-8957-4e93-8c53-438e7996b501, runId = 
98e99ae3-61b3-4e76-b80b-bbaff8af971e] terminated with error
org.apache.hudi.exception.HoodieException: Failed to instantiate Metadata table
        at 
org.apache.hudi.client.SparkRDDWriteClient.initializeMetadataTable(SparkRDDWriteClient.java:293)
        at 
org.apache.hudi.client.SparkRDDWriteClient.initMetadataTable(SparkRDDWriteClient.java:273)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.doInitTable(BaseHoodieWriteClient.java:1257)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.initTable(BaseHoodieWriteClient.java:1297)
        at 
org.apache.hudi.client.SparkRDDWriteClient.upsert(SparkRDDWriteClient.java:139)
        at 
org.apache.hudi.DataSourceUtils.doWriteOperation(DataSourceUtils.java:224)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.liftedTree1$1(HoodieSparkSqlWriter.scala:504)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.writeInternal(HoodieSparkSqlWriter.scala:502)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.write(HoodieSparkSqlWriter.scala:204)
        at 
org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:121)
        at org.apache.hudi.DefaultSource.createRelation(DefaultSource.scala:150)
        at 
org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:45)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:75)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:73)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:84)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:97)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
        at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)
        at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:97)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:93)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:481)
        at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:82)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:481)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:30)
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267)
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:457)
        at 
org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:93)
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:80)
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:78)
        at 
org.apache.spark.sql.execution.QueryExecution.assertCommandExecuted(QueryExecution.scala:115)
        at 
org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:848)
        at 
org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:382)
        at 
org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:355)
        at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:239)
        at 
com.naver.csp.data.cores.common.spark.datasets.hudi.HudiWriter$DataFrameWriterImplicits.hudi(HudiWriter.scala:35)
        at 
com.naver.csp.data.cores.common.spark.datasets.kafka.KafkaHudiSink.batchAsCdc(KafkaHudiSink.scala:160)
        at 
com.naver.csp.data.cores.common.spark.datasets.kafka.KafkaHudiSink.$anonfun$startCdcStreaming$1(KafkaHudiSink.scala:78)
        at 
com.naver.csp.data.cores.common.spark.datasets.kafka.KafkaHudiSink.$anonfun$startCdcStreaming$1$adapted(KafkaHudiSink.scala:78)
        at 
org.apache.spark.sql.execution.streaming.sources.ForeachBatchSink.addBatch(ForeachBatchSink.scala:35)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runBatch$17(MicroBatchExecution.scala:600)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
        at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)
        at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runBatch$16(MicroBatchExecution.scala:598)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken(ProgressReporter.scala:375)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken$(ProgressReporter.scala:373)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:69)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.runBatch(MicroBatchExecution.scala:598)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runActivatedStream$2(MicroBatchExecution.scala:228)
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken(ProgressReporter.scala:375)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken$(ProgressReporter.scala:373)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:69)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runActivatedStream$1(MicroBatchExecution.scala:193)
        at 
org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:57)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.runActivatedStream(MicroBatchExecution.scala:187)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.$anonfun$runStream$1(StreamExecution.scala:303)
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runStream(StreamExecution.scala:286)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:209)
Caused by: org.apache.hudi.exception.HoodieException: Jmx initialize failed:
        at 
org.apache.hudi.metrics.JmxMetricsReporter.<init>(JmxMetricsReporter.java:71)
        at 
org.apache.hudi.metrics.MetricsReporterFactory.createReporter(MetricsReporterFactory.java:75)
        at org.apache.hudi.metrics.Metrics.<init>(Metrics.java:58)
        at org.apache.hudi.metrics.Metrics.getInstance(Metrics.java:82)
        at org.apache.hudi.metrics.HoodieMetrics.<init>(HoodieMetrics.java:95)
        at 
org.apache.hudi.client.BaseHoodieClient.<init>(BaseHoodieClient.java:96)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.<init>(BaseHoodieWriteClient.java:163)
        at 
org.apache.hudi.client.SparkRDDWriteClient.<init>(SparkRDDWriteClient.java:86)
        at 
org.apache.hudi.client.SparkRDDWriteClient.<init>(SparkRDDWriteClient.java:75)
        at 
org.apache.hudi.metadata.SparkHoodieBackedTableMetadataWriter.initializeWriteClient(SparkHoodieBackedTableMetadataWriter.java:146)
        at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.getWriteClient(HoodieBackedTableMetadataWriter.java:1474)
        at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.performTableServices(HoodieBackedTableMetadataWriter.java:1186)
        at 
org.apache.hudi.client.SparkRDDWriteClient.initializeMetadataTable(SparkRDDWriteClient.java:290)
        ... 68 more
Caused by: org.apache.hudi.exception.HoodieException: Jmx service url created 
service:jmx:rmi://localhost:9889/jndi/rmi://localhost:9889/jmxrmi
        at 
org.apache.hudi.metrics.JmxReporterServer.<init>(JmxReporterServer.java:107)
        at 
org.apache.hudi.metrics.JmxReporterServer$Builder.build(JmxReporterServer.java:88)
        at 
org.apache.hudi.metrics.JmxMetricsReporter.createJmxReport(JmxMetricsReporter.java:104)
        at 
org.apache.hudi.metrics.JmxMetricsReporter.<init>(JmxMetricsReporter.java:58)
        ... 80 more
Caused by: java.rmi.server.ExportException: internal error: ObjID already in use
        at 
java.rmi/sun.rmi.transport.ObjectTable.putTarget(ObjectTable.java:184)
        at java.rmi/sun.rmi.transport.Transport.exportObject(Transport.java:106)
        at 
java.rmi/sun.rmi.transport.tcp.TCPTransport.exportObject(TCPTransport.java:254)
        at 
java.rmi/sun.rmi.transport.tcp.TCPEndpoint.exportObject(TCPEndpoint.java:412)
        at java.rmi/sun.rmi.transport.LiveRef.exportObject(LiveRef.java:147)
        at 
java.rmi/sun.rmi.server.UnicastServerRef.exportObject(UnicastServerRef.java:234)
        at java.rmi/sun.rmi.registry.RegistryImpl.setup(RegistryImpl.java:220)
        at java.rmi/sun.rmi.registry.RegistryImpl.<init>(RegistryImpl.java:205)
        at 
java.rmi/java.rmi.registry.LocateRegistry.createRegistry(LocateRegistry.java:203)
        at 
org.apache.hudi.metrics.JmxReporterServer.<init>(JmxReporterServer.java:104)
        ... 83 more
24/05/13 20:28:29 INFO server.AbstractConnector: Stopped 
Spark@376498da{HTTP/1.1, (http/1.1)}{0.0.0.0:4040}
24/05/13 20:28:29 INFO ui.SparkUI: Stopped Spark web UI at 
http://feeder-affiliate-book-svc-sink-09c3c08f71b47a5d-driver-svc.csp.svc:4040
24/05/13 20:28:29 INFO k8s.KubernetesClusterSchedulerBackend: Shutting down all 
executors
24/05/13 20:28:29 INFO 
k8s.KubernetesClusterSchedulerBackend$KubernetesDriverEndpoint: Asking each 
executor to shut down
24/05/13 20:28:29 WARN k8s.ExecutorPodsWatchSnapshotSource: Kubernetes client 
has been closed.
24/05/13 20:28:29 ERROR util.Utils: Uncaught exception in thread main
io.fabric8.kubernetes.client.KubernetesClientException: Failure executing: GET 
at: 
https://kubernetes.default.svc/api/v1/namespaces/csp/persistentvolumeclaims?labelSelector=spark-app-selector%3Dspark-52b31b8f89564fee94b07a763597d89f.
 Message: Forbidden!Configured service account doesn't have access. Service 
account may have been revoked. persistentvolumeclaims is forbidden: User 
"system:serviceaccount:csp:spark" cannot list resource "persistentvolumeclaims" 
in API group "" in the namespace "csp".
        at 
io.fabric8.kubernetes.client.dsl.base.OperationSupport.requestFailure(OperationSupport.java:639)
        at 
io.fabric8.kubernetes.client.dsl.base.OperationSupport.assertResponseCode(OperationSupport.java:576)
        at 
io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleResponse(OperationSupport.java:543)
        at 
io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleResponse(OperationSupport.java:504)
        at 
io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleResponse(OperationSupport.java:487)
        at 
io.fabric8.kubernetes.client.dsl.base.BaseOperation.listRequestHelper(BaseOperation.java:163)
        at 
io.fabric8.kubernetes.client.dsl.base.BaseOperation.list(BaseOperation.java:672)
        at 
io.fabric8.kubernetes.client.dsl.base.BaseOperation.deleteList(BaseOperation.java:786)
        at 
io.fabric8.kubernetes.client.dsl.base.BaseOperation.delete(BaseOperation.java:704)
        at 
org.apache.spark.scheduler.cluster.k8s.KubernetesClusterSchedulerBackend.$anonfun$stop$6(KubernetesClusterSchedulerBackend.scala:138)
        at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1471)
        at 
org.apache.spark.scheduler.cluster.k8s.KubernetesClusterSchedulerBackend.stop(KubernetesClusterSchedulerBackend.scala:139)
        at 
org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:927)
        at org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:2567)
        at 
org.apache.spark.SparkContext.$anonfun$stop$12(SparkContext.scala:2086)
        at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1471)
        at org.apache.spark.SparkContext.stop(SparkContext.scala:2086)
        at 
org.apache.spark.deploy.SparkSubmit.$anonfun$runMain$13(SparkSubmit.scala:963)
        at 
org.apache.spark.deploy.SparkSubmit.$anonfun$runMain$13$adapted(SparkSubmit.scala:963)
        at scala.Option.foreach(Option.scala:407)
        at 
org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:963)
        at 
org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:180)
        at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:203)
        at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:90)
        at 
org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:1043)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:1052)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
24/05/13 20:28:29 INFO spark.MapOutputTrackerMasterEndpoint: 
MapOutputTrackerMasterEndpoint stopped!
24/05/13 20:28:29 INFO memory.MemoryStore: MemoryStore cleared
24/05/13 20:28:29 INFO storage.BlockManager: BlockManager stopped
24/05/13 20:28:29 INFO storage.BlockManagerMaster: BlockManagerMaster stopped
24/05/13 20:28:29 INFO 
scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: 
OutputCommitCoordinator stopped!
24/05/13 20:28:29 INFO spark.SparkContext: Successfully stopped SparkContext
Exception in thread "main" 
org.apache.spark.sql.streaming.StreamingQueryException: Failed to instantiate 
Metadata table
=== Streaming Query ===
Identifier: [id = 7b545afa-8957-4e93-8c53-438e7996b501, runId = 
98e99ae3-61b3-4e76-b80b-bbaff8af971e]
Current Committed Offsets: {KafkaV2[Subscribe[CDC_CDC_GSHOPAPPO_BOOK]]: 
{"CDC_CDC_GSHOPAPPO_BOOK":{"23":4040666,"8":4020522,"17":4016708,"11":4011011,"2":3988721,"20":4040743,"5":3984751,"14":4013617,"4":3989875,"13":4014484,"22":4049439,"7":4009753,"16":4044345,"10":4015180,"1":4014228,"19":4081496,"9":3999227,"18":4046117,"12":3990938,"3":4016564,"21":4039580,"15":4071459,"6":3986614,"24":4051505,"0":3996595}}}
Current Available Offsets: {KafkaV2[Subscribe[CDC_CDC_GSHOPAPPO_BOOK]]: 
{"CDC_CDC_GSHOPAPPO_BOOK":{"23":4040667,"8":4020522,"17":4016708,"11":4011011,"2":3988721,"20":4040743,"5":3984751,"14":4013617,"4":3989875,"13":4014484,"22":4049439,"7":4009753,"16":4044345,"10":4015180,"1":4014228,"19":4081496,"9":3999227,"18":4046117,"12":3990938,"3":4016564,"21":4039580,"15":4071459,"6":3986614,"24":4051505,"0":3996595}}}


Current State: ACTIVE
Thread State: RUNNABLE


Logical Plan:
StreamingDataSourceV2Relation [key#7, value#8, topic#9, partition#10, 
offset#11L, timestamp#12, timestampType#13], 
org.apache.spark.sql.kafka010.KafkaSourceProvider$KafkaScan@325fe569, 
KafkaV2[Subscribe[CDC_CDC_GSHOPAPPO_BOOK]]


        at 
org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runStream(StreamExecution.scala:325)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:209)
Caused by: org.apache.hudi.exception.HoodieException: Failed to instantiate 
Metadata table
        at 
org.apache.hudi.client.SparkRDDWriteClient.initializeMetadataTable(SparkRDDWriteClient.java:293)
        at 
org.apache.hudi.client.SparkRDDWriteClient.initMetadataTable(SparkRDDWriteClient.java:273)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.doInitTable(BaseHoodieWriteClient.java:1257)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.initTable(BaseHoodieWriteClient.java:1297)
        at 
org.apache.hudi.client.SparkRDDWriteClient.upsert(SparkRDDWriteClient.java:139)
        at 
org.apache.hudi.DataSourceUtils.doWriteOperation(DataSourceUtils.java:224)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.liftedTree1$1(HoodieSparkSqlWriter.scala:504)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.writeInternal(HoodieSparkSqlWriter.scala:502)
        at 
org.apache.hudi.HoodieSparkSqlWriterInternal.write(HoodieSparkSqlWriter.scala:204)
        at 
org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:121)
        at org.apache.hudi.DefaultSource.createRelation(DefaultSource.scala:150)
        at 
org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:45)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:75)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:73)
        at 
org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:84)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.$anonfun$applyOrElse$1(QueryExecution.scala:97)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
        at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)
        at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:97)
        at 
org.apache.spark.sql.execution.QueryExecution$$anonfun$eagerlyExecuteCommands$1.applyOrElse(QueryExecution.scala:93)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:481)
        at 
org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:82)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:481)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:30)
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:267)
        at 
org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:263)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)
        at 
org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:30)
        at 
org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:457)
        at 
org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:93)
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:80)
        at 
org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:78)
        at 
org.apache.spark.sql.execution.QueryExecution.assertCommandExecuted(QueryExecution.scala:115)
        at 
org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:848)
        at 
org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:382)
        at 
org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:355)
        at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:239)
        at 
com.naver.csp.data.cores.common.spark.datasets.hudi.HudiWriter$DataFrameWriterImplicits.hudi(HudiWriter.scala:35)
        at 
com.naver.csp.data.cores.common.spark.datasets.kafka.KafkaHudiSink.batchAsCdc(KafkaHudiSink.scala:160)
        at 
com.naver.csp.data.cores.common.spark.datasets.kafka.KafkaHudiSink.$anonfun$startCdcStreaming$1(KafkaHudiSink.scala:78)
        at 
com.naver.csp.data.cores.common.spark.datasets.kafka.KafkaHudiSink.$anonfun$startCdcStreaming$1$adapted(KafkaHudiSink.scala:78)
        at 
org.apache.spark.sql.execution.streaming.sources.ForeachBatchSink.addBatch(ForeachBatchSink.scala:35)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runBatch$17(MicroBatchExecution.scala:600)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103)
        at 
org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163)
        at 
org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)
        at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runBatch$16(MicroBatchExecution.scala:598)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken(ProgressReporter.scala:375)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken$(ProgressReporter.scala:373)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:69)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.runBatch(MicroBatchExecution.scala:598)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runActivatedStream$2(MicroBatchExecution.scala:228)
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken(ProgressReporter.scala:375)
        at 
org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken$(ProgressReporter.scala:373)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:69)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runActivatedStream$1(MicroBatchExecution.scala:193)
        at 
org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:57)
        at 
org.apache.spark.sql.execution.streaming.MicroBatchExecution.runActivatedStream(MicroBatchExecution.scala:187)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.$anonfun$runStream$1(StreamExecution.scala:303)
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
        at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775)
        at 
org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runStream(StreamExecution.scala:286)
        ... 1 more
Caused by: org.apache.hudi.exception.HoodieException: Jmx initialize failed:
        at 
org.apache.hudi.metrics.JmxMetricsReporter.<init>(JmxMetricsReporter.java:71)
        at 
org.apache.hudi.metrics.MetricsReporterFactory.createReporter(MetricsReporterFactory.java:75)
        at org.apache.hudi.metrics.Metrics.<init>(Metrics.java:58)
        at org.apache.hudi.metrics.Metrics.getInstance(Metrics.java:82)
        at org.apache.hudi.metrics.HoodieMetrics.<init>(HoodieMetrics.java:95)
        at 
org.apache.hudi.client.BaseHoodieClient.<init>(BaseHoodieClient.java:96)
        at 
org.apache.hudi.client.BaseHoodieWriteClient.<init>(BaseHoodieWriteClient.java:163)
        at 
org.apache.hudi.client.SparkRDDWriteClient.<init>(SparkRDDWriteClient.java:86)
        at 
org.apache.hudi.client.SparkRDDWriteClient.<init>(SparkRDDWriteClient.java:75)
        at 
org.apache.hudi.metadata.SparkHoodieBackedTableMetadataWriter.initializeWriteClient(SparkHoodieBackedTableMetadataWriter.java:146)
        at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.getWriteClient(HoodieBackedTableMetadataWriter.java:1474)
        at 
org.apache.hudi.metadata.HoodieBackedTableMetadataWriter.performTableServices(HoodieBackedTableMetadataWriter.java:1186)
        at 
org.apache.hudi.client.SparkRDDWriteClient.initializeMetadataTable(SparkRDDWriteClient.java:290)
        ... 68 more
Caused by: org.apache.hudi.exception.HoodieException: Jmx service url created 
service:jmx:rmi://localhost:9889/jndi/rmi://localhost:9889/jmxrmi
        at 
org.apache.hudi.metrics.JmxReporterServer.<init>(JmxReporterServer.java:107)
        at 
org.apache.hudi.metrics.JmxReporterServer$Builder.build(JmxReporterServer.java:88)
        at 
org.apache.hudi.metrics.JmxMetricsReporter.createJmxReport(JmxMetricsReporter.java:104)
        at 
org.apache.hudi.metrics.JmxMetricsReporter.<init>(JmxMetricsReporter.java:58)
        ... 80 more
Caused by: java.rmi.server.ExportException: internal error: ObjID already in use
        at 
java.rmi/sun.rmi.transport.ObjectTable.putTarget(ObjectTable.java:184)
        at java.rmi/sun.rmi.transport.Transport.exportObject(Transport.java:106)
        at 
java.rmi/sun.rmi.transport.tcp.TCPTransport.exportObject(TCPTransport.java:254)
        at 
java.rmi/sun.rmi.transport.tcp.TCPEndpoint.exportObject(TCPEndpoint.java:412)
        at java.rmi/sun.rmi.transport.LiveRef.exportObject(LiveRef.java:147)
        at 
java.rmi/sun.rmi.server.UnicastServerRef.exportObject(UnicastServerRef.java:234)
        at java.rmi/sun.rmi.registry.RegistryImpl.setup(RegistryImpl.java:220)
        at java.rmi/sun.rmi.registry.RegistryImpl.<init>(RegistryImpl.java:205)
        at 
java.rmi/java.rmi.registry.LocateRegistry.createRegistry(LocateRegistry.java:203)
        at 
org.apache.hudi.metrics.JmxReporterServer.<init>(JmxReporterServer.java:104)
        ... 83 more {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to