[ 
https://issues.apache.org/jira/browse/HUDI-7231?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Y Ethan Guo updated HUDI-7231:
------------------------------
    Fix Version/s: 1.0.2

> OOM from uplading metadata in timeline service
> ----------------------------------------------
>
>                 Key: HUDI-7231
>                 URL: https://issues.apache.org/jira/browse/HUDI-7231
>             Project: Apache Hudi
>          Issue Type: Task
>          Components: timeline-server
>            Reporter: Lin Liu
>            Priority: Major
>             Fix For: 1.0.2
>
>
> {code:java}
>     java.util.concurrent.CompletionException: java.lang.OutOfMemoryError: 
> Java heap space
>         at 
> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:273)
>  ~[?:1.8.0_392]
>         at 
> java.util.concurrent.CompletableFuture.uniComposeStage(CompletableFuture.java:1005)
>  ~[?:1.8.0_392]
>         at 
> java.util.concurrent.CompletableFuture.thenCompose(CompletableFuture.java:2137)
>  ~[?:1.8.0_392]
>         at 
> io.javalin.http.JavalinServletHandler.queueNextTaskOrFinish$javalin(JavalinServletHandler.kt:85)
>  ~[javalin-4.6.7.jar:4.6.7]
>         at io.javalin.http.JavalinServlet.service(JavalinServlet.kt:89) 
> ~[javalin-4.6.7.jar:4.6.7]
>         at javax.servlet.http.HttpServlet.service(HttpServlet.java:790) 
> ~[javax.servlet-api-3.1.0.jar:3.1.0]
>         at 
> io.javalin.jetty.JavalinJettyServlet.service(JavalinJettyServlet.kt:58) 
> ~[javalin-4.6.7.jar:4.6.7]
>         at javax.servlet.http.HttpServlet.service(HttpServlet.java:790) 
> ~[javax.servlet-api-3.1.0.jar:3.1.0]
>         at 
> org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:799) 
> ~[jetty-servlet-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:554) 
> ~[jetty-servlet-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:233)
>  ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:1624)
>  ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:233)
>  ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> io.javalin.jetty.JettyServer$start$wsAndHttpHandler$1.doHandle(JettyServer.kt:52)
>  ~[javalin-4.6.7.jar:4.6.7]
>         at 
> org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:188)
>  ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:505) 
> ~[jetty-servlet-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:1594)
>  ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:186)
>  ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1355)
>  ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) 
> ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127)
>  ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at org.eclipse.jetty.server.Server.handle(Server.java:516) 
> ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:487) 
> ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:732) 
> ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:479) 
> ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:277) 
> ~[jetty-server-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311)
>  ~[jetty-io-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105) 
> ~[jetty-io-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104) 
> ~[jetty-io-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:338)
>  ~[jetty-util-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:315)
>  ~[jetty-util-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173)
>  ~[jetty-util-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:131)
>  ~[jetty-util-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:409)
>  ~[jetty-util-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)
>  ~[jetty-util-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at 
> org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)
>  ~[jetty-util-9.4.48.v20220622.jar:9.4.48.v20220622]
>         at java.lang.Thread.run(Thread.java:750) ~[?:1.8.0_392]
>     Caused by: java.lang.OutOfMemoryError: Java heap space
>         at java.io.BufferedInputStream.<init>(BufferedInputStream.java:203) 
> ~[?:1.8.0_392]
>         at 
> org.apache.hadoop.fs.BufferedFSInputStream.<init>(BufferedFSInputStream.java:50)
>  ~[hadoop-common-2.10.2.jar:?]
>         at 
> org.apache.hadoop.fs.RawLocalFileSystem.open(RawLocalFileSystem.java:212) 
> ~[hadoop-common-2.10.2.jar:?]
>         at 
> org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSInputChecker.<init>(ChecksumFileSystem.java:147)
>  ~[hadoop-common-2.10.2.jar:?]
>         at 
> org.apache.hadoop.fs.ChecksumFileSystem.open(ChecksumFileSystem.java:347) 
> ~[hadoop-common-2.10.2.jar:?]
>         at 
> org.apache.hudi.common.fs.HoodieWrapperFileSystem.open(HoodieWrapperFileSystem.java:203)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.log.HoodieLogFileReader.getFSDataInputStream(HoodieLogFileReader.java:499)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.log.HoodieLogFileReader.<init>(HoodieLogFileReader.java:120)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.log.HoodieLogFormatReader.hasNext(HoodieLogFormatReader.java:110)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.scanInternalV1(AbstractHoodieLogRecordReader.java:247)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.log.AbstractHoodieLogRecordReader.scanInternal(AbstractHoodieLogRecordReader.java:220)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.log.HoodieMergedLogRecordScanner.performScan(HoodieMergedLogRecordScanner.java:201)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.log.HoodieMergedLogRecordScanner.<init>(HoodieMergedLogRecordScanner.java:117)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.log.HoodieMergedLogRecordScanner.<init>(HoodieMergedLogRecordScanner.java:76)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.log.HoodieMergedLogRecordScanner$Builder.build(HoodieMergedLogRecordScanner.java:466)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.metadata.HoodieMetadataLogRecordReader$Builder.build(HoodieMetadataLogRecordReader.java:219)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.metadata.HoodieBackedTableMetadata.getLogRecordScanner(HoodieBackedTableMetadata.java:501)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.metadata.HoodieBackedTableMetadata.openReaders(HoodieBackedTableMetadata.java:432)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.metadata.HoodieBackedTableMetadata.lambda$getOrCreateReaders$10(HoodieBackedTableMetadata.java:415)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.metadata.HoodieBackedTableMetadata$$Lambda$3800/1949887064.apply(Unknown
>  Source) ~[?:?]
>         at 
> java.util.concurrent.ConcurrentHashMap.computeIfAbsent(ConcurrentHashMap.java:1660)
>  ~[?:1.8.0_392]
>         at 
> org.apache.hudi.metadata.HoodieBackedTableMetadata.getOrCreateReaders(HoodieBackedTableMetadata.java:415)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.metadata.HoodieBackedTableMetadata.lookupKeysFromFileSlice(HoodieBackedTableMetadata.java:294)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.metadata.HoodieBackedTableMetadata.getRecordsByKeys(HoodieBackedTableMetadata.java:258)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.metadata.HoodieBackedTableMetadata.getRecordByKey(HoodieBackedTableMetadata.java:148)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.metadata.BaseTableMetadata.fetchAllPartitionPaths(BaseTableMetadata.java:316)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.metadata.BaseTableMetadata.getAllPartitionPaths(BaseTableMetadata.java:125)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.metadata.HoodieMetadataFileSystemView.getAllPartitionPaths(HoodieMetadataFileSystemView.java:72)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.view.AbstractTableFileSystemView.ensureAllPartitionsLoadedCorrectly(AbstractTableFileSystemView.java:336)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.view.AbstractTableFileSystemView.loadAllPartitions(AbstractTableFileSystemView.java:807)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.timeline.service.handlers.FileSliceHandler.loadAllPartitions(FileSliceHandler.java:163)
>  ~[hudi-timeline-service-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.timeline.service.RequestHandler.lambda$registerFileSlicesAPI$59(RequestHandler.java:450)
>  ~[hudi-timeline-service-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>     3385303 [stream execution thread for [id = 
> 0499fd07-0abb-4e74-9d3a-685633ff49bc, runId = 
> 754fdc56-cea1-43f4-9f8d-3d5bfe7bfe5c]] ERROR 
> org.apache.hudi.common.table.view.PriorityBasedFileSystemView [] - Got error 
> running preferred function. Trying secondary
>     org.apache.hudi.exception.HoodieRemoteException: Server Error
>         at 
> org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.loadAllPartitions(RemoteHoodieTableFileSystemView.java:535)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.view.PriorityBasedFileSystemView.execute(PriorityBasedFileSystemView.java:69)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.view.PriorityBasedFileSystemView.loadAllPartitions(PriorityBasedFileSystemView.java:172)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.table.action.clean.CleanPlanner.<init>(CleanPlanner.java:110) 
> ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.table.action.clean.CleanPlanActionExecutor.requestClean(CleanPlanActionExecutor.java:107)
>  ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.table.action.clean.CleanPlanActionExecutor.requestClean(CleanPlanActionExecutor.java:159)
>  ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.table.action.clean.CleanPlanActionExecutor.execute(CleanPlanActionExecutor.java:185)
>  ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.table.HoodieSparkCopyOnWriteTable.scheduleCleaning(HoodieSparkCopyOnWriteTable.java:217)
>  ~[hudi-spark-client-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.client.BaseHoodieTableServiceClient.scheduleTableServiceInternal(BaseHoodieTableServiceClient.java:631)
>  ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.client.BaseHoodieTableServiceClient.clean(BaseHoodieTableServiceClient.java:754)
>  ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.client.BaseHoodieWriteClient.clean(BaseHoodieWriteClient.java:862)
>  ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.client.BaseHoodieWriteClient.clean(BaseHoodieWriteClient.java:835)
>  ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.client.BaseHoodieWriteClient.clean(BaseHoodieWriteClient.java:866)
>  ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.client.BaseHoodieWriteClient.autoCleanOnCommit(BaseHoodieWriteClient.java:600)
>  ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.client.BaseHoodieWriteClient.mayBeCleanAndArchive(BaseHoodieWriteClient.java:579)
>  ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.client.BaseHoodieWriteClient.commitStats(BaseHoodieWriteClient.java:248)
>  ~[hudi-client-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.client.SparkRDDWriteClient.commit(SparkRDDWriteClient.java:104)
>  ~[hudi-spark-client-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.HoodieSparkSqlWriterInternal.commitAndPerformPostOperations(HoodieSparkSqlWriter.scala:965)
>  ~[classes/:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.HoodieSparkSqlWriterInternal.writeInternal(HoodieSparkSqlWriter.scala:530)
>  ~[classes/:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.HoodieSparkSqlWriterInternal.write(HoodieSparkSqlWriter.scala:203)
>  ~[classes/:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.HoodieSparkSqlWriter$.write(HoodieSparkSqlWriter.scala:120) 
> ~[classes/:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.DefaultSource.createRelation(DefaultSource.scala:145) 
> ~[classes/:1.0.0-SNAPSHOT]
>         at 
> org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:46)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:90)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:180)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:218)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>  ~[spark-core_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:215) 
> ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:176) 
> ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:127)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:126) 
> ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.DataFrameWriter.$anonfun$runCommand$1(DataFrameWriter.scala:962)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:764) 
> ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:962) 
> ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:414)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:398) 
> ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:287) 
> ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.hudi.functional.cdc.TestCDCStreamingSuite.$anonfun$cdcStreaming$2(TestCDCStreamingSuite.scala:156)
>  ~[test-classes/:?]
>         at 
> org.apache.hudi.functional.cdc.TestCDCStreamingSuite.$anonfun$cdcStreaming$2$adapted(TestCDCStreamingSuite.scala:123)
>  ~[test-classes/:?]
>         at 
> org.apache.spark.sql.execution.streaming.sources.ForeachBatchSink.addBatch(ForeachBatchSink.scala:36)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runBatch$16(MicroBatchExecution.scala:583)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:100)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:764) 
> ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runBatch$15(MicroBatchExecution.scala:581)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken(ProgressReporter.scala:352)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken$(ProgressReporter.scala:350)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:69)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.MicroBatchExecution.runBatch(MicroBatchExecution.scala:581)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runActivatedStream$2(MicroBatchExecution.scala:223)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) 
> ~[scala-library-2.12.10.jar:?]
>         at 
> org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken(ProgressReporter.scala:352)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.ProgressReporter.reportTimeTaken$(ProgressReporter.scala:350)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.StreamExecution.reportTimeTaken(StreamExecution.scala:69)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$runActivatedStream$1(MicroBatchExecution.scala:191)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:57)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.MicroBatchExecution.runActivatedStream(MicroBatchExecution.scala:185)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runStream(StreamExecution.scala:334)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>         at 
> org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:245)
>  ~[spark-sql_2.12-3.0.2.jar:3.0.2]
>     Caused by: org.apache.http.client.HttpResponseException: Server Error
>         at 
> org.apache.http.impl.client.AbstractResponseHandler.handleResponse(AbstractResponseHandler.java:69)
>  ~[httpclient-4.4.1.jar:4.4.1]
>         at 
> org.apache.http.client.fluent.Response.handleResponse(Response.java:90) 
> ~[fluent-hc-4.4.1.jar:4.4.1]
>         at 
> org.apache.http.client.fluent.Response.returnContent(Response.java:97) 
> ~[fluent-hc-4.4.1.jar:4.4.1]
>         at 
> org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.executeRequest(RemoteHoodieTableFileSystemView.java:208)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         at 
> org.apache.hudi.common.table.view.RemoteHoodieTableFileSystemView.loadAllPartitions(RemoteHoodieTableFileSystemView.java:532)
>  ~[hudi-common-1.0.0-SNAPSHOT.jar:1.0.0-SNAPSHOT]
>         ... 66 more
>     3387347 [stream execution thread for [id = 
> 0499fd07-0abb-4e74-9d3a-685633ff49bc, runId = 
> 754fdc56-cea1-43f4-9f8d-3d5bfe7bfe5c]] WARN  
> org.apache.hudi.HoodieSparkSqlWriterInternal [] - Closing write client
>     3387351 [stream execution thread for [id = 
> 0499fd07-0abb-4e74-9d3a-685633ff49bc, runId = 
> 754fdc56-cea1-43f4-9f8d-3d5bfe7bfe5c]] ERROR 
> org.apache.spark.sql.execution.streaming.MicroBatchExecution [] - Query [id = 
> 0499fd07-0abb-4e74-9d3a-685633ff49bc, runId = 
> 754fdc56-cea1-43f4-9f8d-3d5bfe7bfe5c] terminated with error {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to