voonhous commented on code in PR #17573:
URL: https://github.com/apache/hudi/pull/17573#discussion_r2635098331
##########
hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/ColumnStatIndexTestBase.scala:
##########
@@ -36,7 +36,7 @@ import org.apache.hudi.storage.StoragePath
import org.apache.hudi.storage.hadoop.HadoopStorageConfiguration
import org.apache.hudi.testutils.{HoodieSparkClientTestBase,
LogFileColStatsTestUtil}
-import org.apache.avro.Schema
+import org.apache.avro.{Schema}
Review Comment:
Addressed
##########
hudi-spark-datasource/hudi-spark3.5.x/src/main/scala/org/apache/spark/sql/avro/HoodieSpark3_5AvroSerializer.scala:
##########
@@ -17,6 +17,8 @@
package org.apache.spark.sql.avro
+import org.apache.hudi.common.schema.HoodieSchema
Review Comment:
Addressed
##########
hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/cdc/TestCDCDataFrameSuite.scala:
##########
@@ -21,6 +21,7 @@ package org.apache.hudi.functional.cdc
import org.apache.hudi.DataSourceWriteOptions
import org.apache.hudi.DataSourceWriteOptions.{MOR_TABLE_TYPE_OPT_VAL,
PARTITIONPATH_FIELD_OPT_KEY, PRECOMBINE_FIELD_OPT_KEY, RECORDKEY_FIELD_OPT_KEY}
import org.apache.hudi.QuickstartUtils.getQuickstartWriteConfigs
+import org.apache.hudi.common.schema.HoodieSchema
Review Comment:
Addressed
##########
hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/functional/TestBootstrap.java:
##########
@@ -62,7 +62,8 @@
import org.apache.hudi.testutils.HoodieMergeOnReadTestUtils;
import org.apache.hudi.testutils.HoodieSparkClientTestBase;
-import org.apache.avro.Schema;
+import org.apache.hudi.common.schema.HoodieSchema;
+
Review Comment:
Addressed
##########
hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/TestHoodieMergeHandleWithSparkMerger.java:
##########
@@ -171,19 +171,19 @@ public HoodieWriteConfig getWriteConfig(Schema
avroSchema, String recordMergerIm
return getConfigBuilder(true)
.withPath(basePath())
- .withSchema(avroSchema.toString())
+ .withSchema(schema.toString())
.withProps(properties)
.build();
}
- public HoodieWriteConfig buildDefaultWriteConfig(Schema avroSchema) {
- HoodieWriteConfig writeConfig = getWriteConfig(avroSchema,
DefaultMerger.class.getName(),
HoodieRecordMerger.EVENT_TIME_BASED_MERGE_STRATEGY_UUID,
RecordMergeMode.EVENT_TIME_ORDERING);
+ public HoodieWriteConfig buildDefaultWriteConfig(HoodieSchema hoodieSchema) {
+ HoodieWriteConfig writeConfig = getWriteConfig(hoodieSchema,
DefaultMerger.class.getName(),
HoodieRecordMerger.EVENT_TIME_BASED_MERGE_STRATEGY_UUID,
RecordMergeMode.EVENT_TIME_ORDERING);
metaClient = getHoodieMetaClient(storageConf(), basePath(),
HoodieTableType.MERGE_ON_READ, writeConfig.getProps());
return writeConfig;
}
- public HoodieWriteConfig buildCustomWriteConfig(Schema avroSchema) {
- HoodieWriteConfig writeConfig = getWriteConfig(avroSchema,
CustomMerger.class.getName(), HoodieRecordMerger.CUSTOM_MERGE_STRATEGY_UUID,
RecordMergeMode.CUSTOM);
+ public HoodieWriteConfig buildCustomWriteConfig(HoodieSchema hoodieSchema) {
+ HoodieWriteConfig writeConfig = getWriteConfig(hoodieSchema,
CustomMerger.class.getName(), HoodieRecordMerger.CUSTOM_MERGE_STRATEGY_UUID,
RecordMergeMode.CUSTOM);
Review Comment:
Addressed
##########
hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/SparkHelpers.scala:
##########
@@ -51,12 +51,12 @@ object SparkHelpers {
val sourceRecords = HoodieIOFactory.getIOFactory(storage)
.getFileFormatUtils(HoodieFileFormat.PARQUET)
.readAvroRecords(storage, sourceFile).asScala
- val schema: Schema = sourceRecords.head.getSchema
+ val schema: HoodieSchema =
HoodieSchema.fromAvroSchema(sourceRecords.head.getSchema)
val filter: BloomFilter = BloomFilterFactory.createBloomFilter(
BLOOM_FILTER_NUM_ENTRIES_VALUE.defaultValue.toInt,
BLOOM_FILTER_FPP_VALUE.defaultValue.toDouble,
BLOOM_FILTER_DYNAMIC_MAX_ENTRIES.defaultValue.toInt,
BLOOM_FILTER_TYPE.defaultValue);
- val writeSupport: HoodieAvroWriteSupport[_] = new
HoodieAvroWriteSupport(getAvroSchemaConverter(conf.unwrap()).convert(schema),
- schema, Option.of(filter), new Properties())
+ val writeSupport: HoodieAvroWriteSupport[_] = new
HoodieAvroWriteSupport(getAvroSchemaConverter(conf.unwrap()).convert(schema.getAvroSchema),
+ schema.getAvroSchema, Option.of(filter), new Properties())
Review Comment:
Addressed
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]