[
https://issues.apache.org/jira/browse/HUDI-6696?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
sivabalan narayanan reassigned HUDI-6696:
-----------------------------------------
Assignee: Jonathan Vexler
> Clustering fails w/ spark3.3 bundle 0.12.3
> -------------------------------------------
>
> Key: HUDI-6696
> URL: https://issues.apache.org/jira/browse/HUDI-6696
> Project: Apache Hudi
> Issue Type: Bug
> Components: clustering
> Reporter: sivabalan narayanan
> Assignee: Jonathan Vexler
> Priority: Major
>
> I was trying our spark streaming ingestion to hudi and ran into clustering
> issue w/ 0.12.3 bundle.
> {code:java}
> import org.apache.hudi.QuickstartUtils._
> import scala.collection.JavaConversions._
> import org.apache.spark.sql.SaveMode._
> import org.apache.hudi.DataSourceReadOptions._
> import org.apache.hudi.DataSourceWriteOptions._
> import org.apache.hudi.config.HoodieWriteConfig._
> import org.apache.hudi.common.model.HoodieRecord
> import java.time.LocalDateTime
> import org.apache.spark.sql.streaming.Trigger{code}
> {code:java}
> val sourceBasePath = SOURCE_PATH
> val targetPath = TARGET_PATH{code}
> {code:java}
> val df = spark.readStream.format("hudi").load(sourceBasePath)val query =
> df.writeStream.foreachBatch { (batchDF: org.apache.spark.sql.DataFrame, _:
> Long) => {
> println(LocalDateTime.now() + " Start writing cow table")
>
> batchDF.drop("_hoodie_commit_time").drop("_hoodie_commit_seqno").drop("_hoodie_record_key").drop("_hoodie_partition_path").drop("_hoodie_file_name").
> filter("id%2 ==1").
> withColumn("date_col",substring(col("created_at"),0,10)).
> write.format("hudi").
> option(TABLE_TYPE.key, "COPY_ON_WRITE").
> option(PRECOMBINE_FIELD.key, "created_at").
> option(RECORDKEY_FIELD.key, "id").
> option(PARTITIONPATH_FIELD.key, "date_col").
> option("hoodie.datasource.write.operation","insert").
> option("hoodie.insert.shuffle.parallelism","10").
> option("hoodie.parquet.small.file.limit","0").
> option("hoodie.clustering.inline", "true").
> option("hoodie.clustering.inline.max.commits", "10").
> option("hoodie.clustering.plan.strategy.target.file.max.bytes",
> "104857600").
> option("hoodie.clustering.plan.strategy.small.file.limit",
> "52428800").
> option("hoodie.table.name","hudi_tbl").
>
>
> option("hoodie.write.concurrency.mode","optimistic_concurrency_control").
> option("hoodie.cleaner.policy.failed.writes","LAZY").
>
> option("hoodie.write.lock.provider","org.apache.hudi.client.transaction.lock.ZookeeperBasedLockProvider").
> option("hoodie.write.lock.zookeeper.url","ABC").
> option("hoodie.write.lock.zookeeper.port","2181").
> option("hoodie.write.lock.zookeeper.lock_key","tbl1").
>
> option("hoodie.write.lock.zookeeper.base_path","/tmp/locks").
> mode(Append).
> save(targetPath)
> println(LocalDateTime.now() + " finish")
> }
> }.option("checkpointLocation", TARGET_CHECKPOINT).
> trigger(Trigger.ProcessingTime("10
> minutes")).start()query.awaitTermination()
> {code}
> stacktrace:
> {code:java}
> scala> query.awaitTermination()
> 2023-08-15T03:00:05.583925 Start writing cow table
>
> 23/08/15 03:00:05 WARN DFSPropertiesConfiguration: Cannot find HUDI_CONF_DIR,
> please set it as the dir of hudi-defaults.conf
> 23/08/15 03:00:05 WARN DFSPropertiesConfiguration: Properties file
> file:/etc/hudi/conf/hudi-defaults.conf not found. Ignoring to load props file
> # WARNING: Unable to get Instrumentation. Dynamic Attach failed. You may add
> this JAR as -javaagent manually, or supply -Djdk.attach.allowAttachSelf
> # WARNING: Unable to attach Serviceability Agent. Unable to attach even with
> module exceptions: [org.apache.hudi.org.openjdk.jol.vm.sa.SASupportException:
> Sense failed., org.apache.hudi.org.openjdk.jol.vm.sa.SASupportException:
> Sense failed., org.apache.hudi.org.openjdk.jol.vm.sa.SASupportException:
> Sense failed.]
> 23/08/15 03:01:14 WARN BaseRollbackActionExecutor: Rollback finished without
> deleting inflight instant file.
> Instant=[==>20230814185344428__replacecommit__INFLIGHT]
> 23/08/15 03:01:16 ERROR MicroBatchExecution: Query [id =
> fd43b951-f1fd-479e-bd87-1aae2e9e396d, runId =
> 70797db4-fbcd-40f0-b09d-540013da6717] terminated with error
> java.util.concurrent.CompletionException: java.lang.ClassNotFoundException:
> Failed to find data source: hoodie-parquet. Please find packages at
> https://spark.apache.org/third-party-projects.html
>
> at
> java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
> ~[?:?]
> at
> java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
> ~[?:?]
> at
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1702)
> ~[?:?]
> at
> java.util.concurrent.CompletableFuture$AsyncSupply.exec(CompletableFuture.java:1692)
> ~[?:?]
> at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290)
> ~[?:?]
> at
> java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020)
> ~[?:?]
> at java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656)
> ~[?:?]
> at
> java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594) ~[?:?]
> at
> java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:183)
> ~[?:?]
> Caused by: java.lang.ClassNotFoundException:
> Failed to find data source: hoodie-parquet. Please find packages at
> https://spark.apache.org/third-party-projects.html
>
> at
> org.apache.spark.sql.errors.QueryExecutionErrors$.failedToFindDataSourceError(QueryExecutionErrors.scala:573)
> ~[spark-catalyst_2.12-3.3.0.jar:3.3.0]
> at
> org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:675)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at
> org.apache.spark.sql.execution.datasources.DataSource.providingClass$lzycompute(DataSource.scala:100)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at
> org.apache.spark.sql.execution.datasources.DataSource.providingClass(DataSource.scala:99)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at
> org.apache.spark.sql.execution.datasources.DataSource.providingInstance(DataSource.scala:113)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at
> org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:345)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at
> org.apache.hudi.BaseFileOnlyRelation.toHadoopFsRelation(BaseFileOnlyRelation.scala:203)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> org.apache.hudi.DefaultSource$.resolveBaseFileOnlyRelation(DefaultSource.scala:268)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> org.apache.hudi.DefaultSource$.createRelation(DefaultSource.scala:232)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> org.apache.spark.sql.adapter.BaseSpark3Adapter.createRelation(BaseSpark3Adapter.scala:101)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.readRecordsForGroupAsRow(MultipleSparkJobExecutionStrategy.java:405)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.lambda$runClusteringForGroupAsyncAsRow$6(MultipleSparkJobExecutionStrategy.java:248)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1700)
> ~[?:?]
> ... 6 more
> Caused by: java.lang.ClassNotFoundException: hoodie-parquet.DefaultSource
> at
> jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581)
> ~[?:?]
> at
> jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
> ~[?:?]
> at java.lang.ClassLoader.loadClass(ClassLoader.java:527) ~[?:?]
> at
> org.apache.spark.sql.execution.datasources.DataSource$.$anonfun$lookupDataSource$5(DataSource.scala:661)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at scala.util.Try$.apply(Try.scala:213) ~[scala-library-2.12.14.jar:?]
> at
> org.apache.spark.sql.execution.datasources.DataSource$.$anonfun$lookupDataSource$4(DataSource.scala:661)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at scala.util.Failure.orElse(Try.scala:224)
> ~[scala-library-2.12.14.jar:?]
> at
> org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:661)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at
> org.apache.spark.sql.execution.datasources.DataSource.providingClass$lzycompute(DataSource.scala:100)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at
> org.apache.spark.sql.execution.datasources.DataSource.providingClass(DataSource.scala:99)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at
> org.apache.spark.sql.execution.datasources.DataSource.providingInstance(DataSource.scala:113)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at
> org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:345)
> ~[spark-sql_2.12-3.3.0.jar:3.3.0]
> at
> org.apache.hudi.BaseFileOnlyRelation.toHadoopFsRelation(BaseFileOnlyRelation.scala:203)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> org.apache.hudi.DefaultSource$.resolveBaseFileOnlyRelation(DefaultSource.scala:268)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> org.apache.hudi.DefaultSource$.createRelation(DefaultSource.scala:232)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> org.apache.spark.sql.adapter.BaseSpark3Adapter.createRelation(BaseSpark3Adapter.scala:101)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.readRecordsForGroupAsRow(MultipleSparkJobExecutionStrategy.java:405)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.lambda$runClusteringForGroupAsyncAsRow$6(MultipleSparkJobExecutionStrategy.java:248)
> ~[hudi-spark3.3-bundle_2.12-0.12.3.jar:0.12.3]
> at
> java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1700)
> ~[?:?]
> ... 6 more
> org.apache.spark.sql.streaming.StreamingQueryException: Query [id =
> fd43b951-f1fd-479e-bd87-1aae2e9e396d, runId =
> 70797db4-fbcd-40f0-b09d-540013da6717] terminated with exception:
> java.lang.ClassNotFoundException:
> Failed to find data source: hoodie-parquet. Please find packages at
> https://spark.apache.org/third-party-projects.html at
> org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runStream(StreamExecution.scala:330)
> at
> org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:208)
> Caused by: java.util.concurrent.CompletionException:
> java.lang.ClassNotFoundException:
> Failed to find data source: hoodie-parquet. Please find packages at
> https://spark.apache.org/third-party-projects.html at
> java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
> at
> java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
> at
> java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1702)
> at
> java.base/java.util.concurrent.CompletableFuture$AsyncSupply.exec(CompletableFuture.java:1692)
> at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290)
> at
> java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020)
> at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656)
> at
> java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594)
> at
> java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:183)
> Caused by: java.lang.ClassNotFoundException:
> Failed to find data source: hoodie-parquet. Please find packages at
> https://spark.apache.org/third-party-projects.html at
> org.apache.spark.sql.errors.QueryExecutionErrors$.failedToFindDataSourceError(QueryExecutionErrors.scala:573)
> at
> org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:675)
> at
> org.apache.spark.sql.execution.datasources.DataSource.providingClass$lzycompute(DataSource.scala:100)
> at
> org.apache.spark.sql.execution.datasources.DataSource.providingClass(DataSource.scala:99)
> at
> org.apache.spark.sql.execution.datasources.DataSource.providingInstance(DataSource.scala:113)
> at
> org.apache.spark.sql.execution.datasources.DataSource.resolveRelation(DataSource.scala:345)
> at
> org.apache.hudi.BaseFileOnlyRelation.toHadoopFsRelation(BaseFileOnlyRelation.scala:203)
> at
> org.apache.hudi.DefaultSource$.resolveBaseFileOnlyRelation(DefaultSource.scala:268)
> at org.apache.hudi.DefaultSource$.createRelation(DefaultSource.scala:232)
> at
> org.apache.spark.sql.adapter.BaseSpark3Adapter.createRelation(BaseSpark3Adapter.scala:101)
> at
> org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.readRecordsForGroupAsRow(MultipleSparkJobExecutionStrategy.java:405)
> at
> org.apache.hudi.client.clustering.run.strategy.MultipleSparkJobExecutionStrategy.lambda$runClusteringForGroupAsyncAsRow$6(MultipleSparkJobExecutionStrategy.java:248)
> at
> java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1700)
> ... 6 more
> Caused by: java.lang.ClassNotFoundException: hoodie-parquet.DefaultSource
> at
> java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581)
> at
> java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
> at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:527)
> at
> org.apache.spark.sql.execution.datasources.DataSource$.$anonfun$lookupDataSource$5(DataSource.scala:661)
> at scala.util.Try$.apply(Try.scala:213)
> at
> org.apache.spark.sql.execution.datasources.DataSource$.$anonfun$lookupDataSource$4(DataSource.scala:661)
> at scala.util.Failure.orElse(Try.scala:224)
> at
> org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:661)
> ... 17 morescala>
> {code}
>
>
--
This message was sent by Atlassian Jira
(v8.20.10#820010)