voonhous opened a new issue, #17938:
URL: https://github.com/apache/hudi/issues/17938

   ### Bug Description
   
   **What happened:**
   Running the integration test for 
`org.apache.hudi.integ.ITTestHoodieDemo#testParquetDemo` throws the error below:
   
   **What you expected:**
   
   **Steps to reproduce:**
   1.
   2.
   3.
   
   
   ### Environment
   
   **Hudi version:** 1.2.0-SNAPSHOT
   **Query engine:** (Spark/Flink/Trino etc) Spark
   **Relevant configs:**
   ```java
   /bin/bash -c spark-submit --class 
org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamer 
/var/hoodie/ws/docker/hoodie/hadoop/hive_base/target/hoodie-utilities.jar \ 
       --table-type COPY_ON_WRITE \
       --base-file-format PARQUET \
       --source-class org.apache.hudi.utilities.sources.JsonDFSSource \
       --source-ordering-field ts \
       --target-base-path /user/hive/warehouse/stock_ticks_cow \
       --target-table stock_ticks_cow \
       --props /var/demo/config/dfs-source.properties \
       --schemaprovider-class 
org.apache.hudi.utilities.schema.FilebasedSchemaProvider \
       --enable-hive-sync \
       --hoodie-conf 
hoodie.datasource.hive_sync.jdbcurl=jdbc:hive2://hiveserver:10000/ 
       --hoodie-conf 
hoodie.datasource.hive_sync.partition_extractor_class=org.apache.hudi.hive.SlashEncodedDayPartitionValueExtractor
  \
       --hoodie-conf hoodie.datasource.hive_sync.username=hive  \
       --hoodie-conf hoodie.datasource.hive_sync.password=hive  \
       --hoodie-conf hoodie.datasource.hive_sync.partition_fields=dt  \
       --hoodie-conf hoodie.datasource.hive_sync.database=default  \
       --hoodie-conf hoodie.datasource.hive_sync.table=stock_ticks_cow
   ```
   
   ### Logs and Stack Trace
   
   ```
   java.lang.ClassCastException: class org.apache.avro.util.Utf8 cannot be cast 
to class java.lang.String (org.apache.avro.util.Utf8 is in unnamed module of 
loader 'app'; java.lang.String is in module java.base of loader 'bootstrap')
        at java.lang.String.compareTo(String.java:125) ~[?:?]
        at 
org.apache.hudi.common.table.read.BufferedRecordMergerFactory.shouldKeepNewerRecord(BufferedRecordMergerFactory.java:492)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.BufferedRecordMergerFactory.access$100(BufferedRecordMergerFactory.java:41)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.BufferedRecordMergerFactory$EventTimeRecordMerger.finalMerge(BufferedRecordMergerFactory.java:193)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.buffer.FileGroupRecordBuffer.hasNextBaseRecord(FileGroupRecordBuffer.java:241)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.buffer.KeyBasedFileGroupRecordBuffer.hasNextBaseRecord(KeyBasedFileGroupRecordBuffer.java:138)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.buffer.KeyBasedFileGroupRecordBuffer.doHasNext(KeyBasedFileGroupRecordBuffer.java:147)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.buffer.FileGroupRecordBuffer.hasNext(FileGroupRecordBuffer.java:152)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.HoodieFileGroupReader.hasNext(HoodieFileGroupReader.java:247)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.HoodieFileGroupReader$HoodieFileGroupReaderIterator.hasNext(HoodieFileGroupReader.java:334)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.util.collection.MappingIterator.hasNext(MappingIterator.java:39)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.io.FileGroupReaderBasedMergeHandle.doMerge(FileGroupReaderBasedMergeHandle.java:270)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at org.apache.hudi.io.IOUtils.runMerge(IOUtils.java:120) 
~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.handleUpdate(BaseSparkCommitActionExecutor.java:392)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.handleUpsertPartition(BaseSparkCommitActionExecutor.java:358)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.lambda$mapPartitionsAsRDD$e664f7e$1(BaseSparkCommitActionExecutor.java:298)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.spark.api.java.JavaRDDLike.$anonfun$mapPartitionsWithIndex$1(JavaRDDLike.scala:102)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.api.java.JavaRDDLike.$anonfun$mapPartitionsWithIndex$1$adapted(JavaRDDLike.scala:102)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$2(RDD.scala:910) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$2$adapted(RDD.scala:910)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:331) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.$anonfun$getOrCompute$1(RDD.scala:381) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.$anonfun$getOrElseUpdate$1(BlockManager.scala:1372)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.$anonfun$doPutIterator$1(BlockManager.scala:1614)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.org$apache$spark$storage$BlockManager$$doPut(BlockManager.scala:1524)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:1588) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:1389) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.getOrElseUpdateRDDBlock(BlockManager.scala:1343)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:379) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:329) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:331) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.scheduler.Task.run(Task.scala:141) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
 [spark-common-utils_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
 [spark-common-utils_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94) 
[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623) 
[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) 
[?:?]
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) 
[?:?]
        at java.lang.Thread.run(Thread.java:829) [?:?]
   26/01/18 13:39:49 WARN BlockManager: Putting block rdd_36_0 failed due to 
exception org.apache.hudi.exception.HoodieUpsertException: Error upserting 
bucketType UPDATE for partition :0.
   26/01/18 13:39:49 WARN BlockManager: Block rdd_36_0 could not be removed as 
it was not found on disk or in memory
   26/01/18 13:39:49 ERROR Executor: Exception in task 0.0 in stage 14.0 (TID 
17)
   org.apache.hudi.exception.HoodieUpsertException: Error upserting bucketType 
UPDATE for partition :0
        at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.handleUpsertPartition(BaseSparkCommitActionExecutor.java:365)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.lambda$mapPartitionsAsRDD$e664f7e$1(BaseSparkCommitActionExecutor.java:298)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.spark.api.java.JavaRDDLike.$anonfun$mapPartitionsWithIndex$1(JavaRDDLike.scala:102)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.api.java.JavaRDDLike.$anonfun$mapPartitionsWithIndex$1$adapted(JavaRDDLike.scala:102)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$2(RDD.scala:910) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$2$adapted(RDD.scala:910)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:331) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.$anonfun$getOrCompute$1(RDD.scala:381) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.$anonfun$getOrElseUpdate$1(BlockManager.scala:1372)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.$anonfun$doPutIterator$1(BlockManager.scala:1614)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.org$apache$spark$storage$BlockManager$$doPut(BlockManager.scala:1524)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:1588) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:1389) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.storage.BlockManager.getOrElseUpdateRDDBlock(BlockManager.scala:1343)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:379) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:329) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:331) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.scheduler.Task.run(Task.scala:141) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
 ~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
 ~[spark-common-utils_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
 ~[spark-common-utils_2.12-3.5.3.jar:3.5.3]
        at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94) 
~[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623) 
[spark-core_2.12-3.5.3.jar:3.5.3]
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) 
[?:?]
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) 
[?:?]
        at java.lang.Thread.run(Thread.java:829) [?:?]
   Caused by: java.lang.ClassCastException: class org.apache.avro.util.Utf8 
cannot be cast to class java.lang.String (org.apache.avro.util.Utf8 is in 
unnamed module of loader 'app'; java.lang.String is in module java.base of 
loader 'bootstrap')
        at java.lang.String.compareTo(String.java:125) ~[?:?]
        at 
org.apache.hudi.common.table.read.BufferedRecordMergerFactory.shouldKeepNewerRecord(BufferedRecordMergerFactory.java:492)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.BufferedRecordMergerFactory.access$100(BufferedRecordMergerFactory.java:41)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.BufferedRecordMergerFactory$EventTimeRecordMerger.finalMerge(BufferedRecordMergerFactory.java:193)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.buffer.FileGroupRecordBuffer.hasNextBaseRecord(FileGroupRecordBuffer.java:241)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.buffer.KeyBasedFileGroupRecordBuffer.hasNextBaseRecord(KeyBasedFileGroupRecordBuffer.java:138)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.buffer.KeyBasedFileGroupRecordBuffer.doHasNext(KeyBasedFileGroupRecordBuffer.java:147)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.buffer.FileGroupRecordBuffer.hasNext(FileGroupRecordBuffer.java:152)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.HoodieFileGroupReader.hasNext(HoodieFileGroupReader.java:247)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.table.read.HoodieFileGroupReader$HoodieFileGroupReaderIterator.hasNext(HoodieFileGroupReader.java:334)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.common.util.collection.MappingIterator.hasNext(MappingIterator.java:39)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.io.FileGroupReaderBasedMergeHandle.doMerge(FileGroupReaderBasedMergeHandle.java:270)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at org.apache.hudi.io.IOUtils.runMerge(IOUtils.java:120) 
~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.handleUpdate(BaseSparkCommitActionExecutor.java:392)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.handleUpsertPartition(BaseSparkCommitActionExecutor.java:358)
 ~[hoodie-utilities.jar:1.2.0-SNAPSHOT]
        ... 35 more
   26/01/18 13:39:49 WARN TaskSetManager: Lost task 0.0 in stage 14.0 (TID 17) 
(adhoc-1 executor driver): org.apache.hudi.exception.HoodieUpsertException: 
Error upserting bucketType UPDATE for partition :0
        at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.handleUpsertPartition(BaseSparkCommitActionExecutor.java:365)
        at 
org.apache.hudi.table.action.commit.BaseSparkCommitActionExecutor.lambda$mapPartitionsAsRDD$e664f7e$1(BaseSparkCommitActionExecutor.java:298)
        at 
org.apache.spark.api.java.JavaRDDLike.$anonfun$mapPartitionsWithIndex$1(JavaRDDLike.scala:102)
        at 
org.apache.spark.api.java.JavaRDDLike.$anonfun$mapPartitionsWithIndex$1$adapted(JavaRDDLike.scala:102)
        at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$2(RDD.scala:910)
        at 
org.apache.spark.rdd.RDD.$anonfun$mapPartitionsWithIndex$2$adapted(RDD.scala:910)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
        at org.apache.spark.rdd.RDD.$anonfun$getOrCompute$1(RDD.scala:381)
        at 
org.apache.spark.storage.BlockManager.$anonfun$getOrElseUpdate$1(BlockManager.scala:1372)
        at 
org.apache.spark.storage.BlockManager.$anonfun$doPutIterator$1(BlockManager.scala:1614)
        at 
org.apache.spark.storage.BlockManager.org$apache$spark$storage$BlockManager$$doPut(BlockManager.scala:1524)
        at 
org.apache.spark.storage.BlockManager.doPutIterator(BlockManager.scala:1588)
        at 
org.apache.spark.storage.BlockManager.getOrElseUpdate(BlockManager.scala:1389)
        at 
org.apache.spark.storage.BlockManager.getOrElseUpdateRDDBlock(BlockManager.scala:1343)
        at org.apache.spark.rdd.RDD.getOrCompute(RDD.scala:379)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:329)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:367)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:331)
        at 
org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:104)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:54)
        at 
org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:166)
        at org.apache.spark.scheduler.Task.run(Task.scala:141)
        at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$4(Executor.scala:620)
        at 
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)
        at 
org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)
        at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:94)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:623)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
        at java.base/java.lang.Thread.run(Thread.java:829)
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to