This is an automated email from the ASF dual-hosted git repository.
yihua pushed a change to branch release-1.1.0
in repository https://gitbox.apache.org/repos/asf/hudi.git
from 629e33fae07c chore: Fix integ-test pom and add new bundles to release
validation script
new d537c90467ad chore: Reduce log volume by changing INFO to DEBUG for
table loading messages (#14057)
new ee6cd456d8f8 fix: Fix output type extracting for key selector in flink
stream read (#14065)
new 216701dc4f88 chore: Exclude hudi-trino-plugin from RAT checks (#14067)
new 4882bb255c09 fix: Spark Schema Evolution Fix for nested columns
(#14075)
new fc83da464184 fix: Fixing point lookup in MDT partitions (#14085)
new df7a366bc5e6 fix: disable embedded timeline service for flink upgrade
(#14096)
new 8e2a32e994a4 fix: flink mdt compaction should finish pending
compactions first (#14095)
new 8398e1ff53de fix: exclude unused dependencies from META INF in presto
bundle (#14102)
new 1b2ebfaca4fc fix: Upgrade Parquet Avro and commons lang3 versions in
presto bundle (#14099)
new 65aa9da7afb0 fix: show_index command (#14113)
new 3e8b56694bdb fix: Remove catalog access from SparkSQLWriter (#14083)
new 119ee0cffa36 fix: Ignore field nullability while checking whether
record should be rewritten in COW write path (#14094)
new ca9dea50b8ac fix: disable NBCC with default single writer for MDT for
downgrade from version 9 (#14109)
new eeb07995b3ac fix: Skip payload class validation when merge mode is not
custom with v9 tables (#14116)
The 14 revisions listed above as "new" are entirely new to this
repository and will be described in separate emails. The revisions
listed as "add" were already present in the repository and have only
been added to this reference.
Summary of changes:
.../hudi/client/BaseHoodieTableServiceClient.java | 2 +-
.../HoodieLogCompactionPlanGenerator.java | 1 -
.../hudi/table/upgrade/UpgradeDowngradeUtils.java | 6 +
.../hudi/client/HoodieFlinkTableServiceClient.java | 8 --
.../apache/hudi/client/HoodieFlinkWriteClient.java | 6 -
.../FlinkHoodieBackedTableMetadataWriter.java | 58 +--------
.../table/upgrade/FlinkUpgradeDowngradeHelper.java | 4 +
.../hudi/io/storage/HoodieSparkParquetReader.java | 8 +-
.../parquet/HoodieParquetReadSupport.scala | 98 +++++++++++++++
.../org/apache/spark/sql/hudi/SparkAdapter.scala | 5 +-
.../TestUpgradeDowngradeConcurrencyControl.java | 95 +++++++++++++++
.../parquet/TestHoodieParquetReadSupport.scala | 103 ++++++++++++++++
.../java/org/apache/hudi/avro/AvroSchemaUtils.java | 11 +-
.../hudi/common/table/HoodieTableMetaClient.java | 4 +-
.../table/log/BaseHoodieLogRecordReader.java | 18 ++-
.../table/log/HoodieLogBlockMetadataScanner.java | 5 +-
.../table/log/HoodieMergedLogRecordReader.java | 18 +--
.../common/table/log/block/HoodieDataBlock.java | 22 ++--
.../table/log/block/HoodieHFileDataBlock.java | 27 ++++-
.../table/read/buffer/FileGroupRecordBuffer.java | 20 ++--
.../timeline/versioning/v1/ActiveTimelineV1.java | 2 +-
.../timeline/versioning/v2/ActiveTimelineV2.java | 2 +-
.../io/storage/HoodieNativeAvroHFileReader.java | 20 +++-
.../hudi/io/storage/HoodieSeekingFileReader.java | 8 ++
.../org/apache/hudi/avro/TestAvroSchemaUtils.java | 38 +++++-
.../org/apache/hudi/table/HoodieTableSource.java | 8 +-
.../sink/TestStreamWriteOperatorCoordinator.java | 12 +-
.../source/TestStreamReadMonitoringFunction.java | 2 +-
.../apache/hudi/table/ITTestHoodieDataSource.java | 35 ++++++
.../io/hadoop/TestHoodieHFileReaderWriter.java | 133 +++++++++++++++++++++
.../org/apache/hudi/HoodieSparkSqlWriter.scala | 39 +++---
.../scala/org/apache/hudi/HoodieWriterUtils.scala | 17 ++-
.../spark/sql/catalyst/plans/logical/Index.scala | 4 +-
.../org/apache/hudi/TestHoodieWriterUtils.java | 45 ++++++-
.../TestHoodieClientOnMergeOnReadStorage.java | 2 -
.../hudi/functional/TestSparkDataSource.scala | 22 +++-
.../apache/spark/sql/hudi/ddl/TestSpark3DDL.scala | 2 -
.../command/InsertIntoHoodieTableCommand.scala | 8 +-
.../hudi/command/MergeIntoHoodieTableCommand.scala | 5 +-
.../apache/spark/sql/adapter/Spark3_3Adapter.scala | 6 +-
.../Spark33LegacyHoodieParquetFileFormat.scala | 2 +-
.../datasources/parquet/Spark33ParquetReader.scala | 2 +-
.../apache/spark/sql/adapter/Spark3_4Adapter.scala | 6 +-
.../Spark34LegacyHoodieParquetFileFormat.scala | 2 +-
.../datasources/parquet/Spark34ParquetReader.scala | 2 +-
.../apache/spark/sql/adapter/Spark3_5Adapter.scala | 6 +-
.../Spark35LegacyHoodieParquetFileFormat.scala | 2 +-
.../datasources/parquet/Spark35ParquetReader.scala | 2 +-
.../command/InsertIntoHoodieTableCommand.scala | 8 +-
.../hudi/command/MergeIntoHoodieTableCommand.scala | 5 +-
.../apache/spark/sql/adapter/Spark4_0Adapter.scala | 6 +-
.../Spark40LegacyHoodieParquetFileFormat.scala | 2 +-
.../datasources/parquet/Spark40ParquetReader.scala | 2 +-
packaging/hudi-presto-bundle/pom.xml | 14 +++
pom.xml | 3 +
55 files changed, 806 insertions(+), 187 deletions(-)
create mode 100644
hudi-client/hudi-spark-client/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/HoodieParquetReadSupport.scala
create mode 100644
hudi-client/hudi-spark-client/src/test/java/org/apache/hudi/table/upgrade/TestUpgradeDowngradeConcurrencyControl.java
create mode 100644
hudi-client/hudi-spark-client/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/TestHoodieParquetReadSupport.scala