voonhous commented on code in PR #17573:
URL: https://github.com/apache/hudi/pull/17573#discussion_r2635098331


##########
hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/ColumnStatIndexTestBase.scala:
##########
@@ -36,7 +36,7 @@ import org.apache.hudi.storage.StoragePath
 import org.apache.hudi.storage.hadoop.HadoopStorageConfiguration
 import org.apache.hudi.testutils.{HoodieSparkClientTestBase, 
LogFileColStatsTestUtil}
 
-import org.apache.avro.Schema
+import org.apache.avro.{Schema}

Review Comment:
   Addressed



##########
hudi-spark-datasource/hudi-spark3.5.x/src/main/scala/org/apache/spark/sql/avro/HoodieSpark3_5AvroSerializer.scala:
##########
@@ -17,6 +17,8 @@
 
 package org.apache.spark.sql.avro
 
+import org.apache.hudi.common.schema.HoodieSchema

Review Comment:
   Addressed



##########
hudi-spark-datasource/hudi-spark/src/test/scala/org/apache/hudi/functional/cdc/TestCDCDataFrameSuite.scala:
##########
@@ -21,6 +21,7 @@ package org.apache.hudi.functional.cdc
 import org.apache.hudi.DataSourceWriteOptions
 import org.apache.hudi.DataSourceWriteOptions.{MOR_TABLE_TYPE_OPT_VAL, 
PARTITIONPATH_FIELD_OPT_KEY, PRECOMBINE_FIELD_OPT_KEY, RECORDKEY_FIELD_OPT_KEY}
 import org.apache.hudi.QuickstartUtils.getQuickstartWriteConfigs
+import org.apache.hudi.common.schema.HoodieSchema

Review Comment:
   Addressed



##########
hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/functional/TestBootstrap.java:
##########
@@ -62,7 +62,8 @@
 import org.apache.hudi.testutils.HoodieMergeOnReadTestUtils;
 import org.apache.hudi.testutils.HoodieSparkClientTestBase;
 
-import org.apache.avro.Schema;
+import org.apache.hudi.common.schema.HoodieSchema;
+

Review Comment:
   Addressed



##########
hudi-spark-datasource/hudi-spark/src/test/java/org/apache/hudi/TestHoodieMergeHandleWithSparkMerger.java:
##########
@@ -171,19 +171,19 @@ public HoodieWriteConfig getWriteConfig(Schema 
avroSchema, String recordMergerIm
 
     return getConfigBuilder(true)
         .withPath(basePath())
-        .withSchema(avroSchema.toString())
+        .withSchema(schema.toString())
         .withProps(properties)
         .build();
   }
 
-  public HoodieWriteConfig buildDefaultWriteConfig(Schema avroSchema) {
-    HoodieWriteConfig writeConfig = getWriteConfig(avroSchema, 
DefaultMerger.class.getName(), 
HoodieRecordMerger.EVENT_TIME_BASED_MERGE_STRATEGY_UUID, 
RecordMergeMode.EVENT_TIME_ORDERING);
+  public HoodieWriteConfig buildDefaultWriteConfig(HoodieSchema hoodieSchema) {
+    HoodieWriteConfig writeConfig = getWriteConfig(hoodieSchema, 
DefaultMerger.class.getName(), 
HoodieRecordMerger.EVENT_TIME_BASED_MERGE_STRATEGY_UUID, 
RecordMergeMode.EVENT_TIME_ORDERING);
     metaClient = getHoodieMetaClient(storageConf(), basePath(), 
HoodieTableType.MERGE_ON_READ, writeConfig.getProps());
     return writeConfig;
   }
 
-  public HoodieWriteConfig buildCustomWriteConfig(Schema avroSchema) {
-    HoodieWriteConfig writeConfig = getWriteConfig(avroSchema, 
CustomMerger.class.getName(), HoodieRecordMerger.CUSTOM_MERGE_STRATEGY_UUID, 
RecordMergeMode.CUSTOM);
+  public HoodieWriteConfig buildCustomWriteConfig(HoodieSchema hoodieSchema) {
+    HoodieWriteConfig writeConfig = getWriteConfig(hoodieSchema, 
CustomMerger.class.getName(), HoodieRecordMerger.CUSTOM_MERGE_STRATEGY_UUID, 
RecordMergeMode.CUSTOM);

Review Comment:
   Addressed



##########
hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/SparkHelpers.scala:
##########
@@ -51,12 +51,12 @@ object SparkHelpers {
     val sourceRecords = HoodieIOFactory.getIOFactory(storage)
       .getFileFormatUtils(HoodieFileFormat.PARQUET)
       .readAvroRecords(storage, sourceFile).asScala
-    val schema: Schema = sourceRecords.head.getSchema
+    val schema: HoodieSchema = 
HoodieSchema.fromAvroSchema(sourceRecords.head.getSchema)
     val filter: BloomFilter = BloomFilterFactory.createBloomFilter(
       BLOOM_FILTER_NUM_ENTRIES_VALUE.defaultValue.toInt, 
BLOOM_FILTER_FPP_VALUE.defaultValue.toDouble,
       BLOOM_FILTER_DYNAMIC_MAX_ENTRIES.defaultValue.toInt, 
BLOOM_FILTER_TYPE.defaultValue);
-    val writeSupport: HoodieAvroWriteSupport[_] = new 
HoodieAvroWriteSupport(getAvroSchemaConverter(conf.unwrap()).convert(schema),
-      schema, Option.of(filter), new Properties())
+    val writeSupport: HoodieAvroWriteSupport[_] = new 
HoodieAvroWriteSupport(getAvroSchemaConverter(conf.unwrap()).convert(schema.getAvroSchema),
+      schema.getAvroSchema, Option.of(filter), new Properties())

Review Comment:
   Addressed



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to