yihua commented on code in PR #14344:
URL: https://github.com/apache/hudi/pull/14344#discussion_r2597152073
##########
hudi-sync/hudi-datahub-sync/src/main/java/org/apache/hudi/sync/datahub/DataHubSyncClient.java:
##########
@@ -368,9 +367,9 @@ private Stream<MetadataChangeProposalWrapper>
createDatasetEntity() {
return result;
}
- Schema getAvroSchemaWithoutMetadataFields(HoodieTableMetaClient metaClient) {
+ HoodieSchema getAvroSchemaWithoutMetadataFields(HoodieTableMetaClient
metaClient) {
Review Comment:
```suggestion
HoodieSchema getTableSchemaWithoutMetadataFields(HoodieTableMetaClient
metaClient) {
```
##########
hudi-flink-datasource/hudi-flink/src/main/java/org/apache/hudi/table/catalog/TableOptionProperties.java:
##########
@@ -202,19 +200,17 @@ public static Map<String, String>
getTableOptions(Map<String, String> options) {
public static Map<String, String> translateFlinkTableProperties2Spark(
CatalogTable catalogTable,
- Configuration hadoopConf,
Map<String, String> properties,
List<String> partitionKeys,
boolean withOperationField) {
RowType rowType =
supplementMetaFields(DataTypeUtils.toRowType(catalogTable.getUnresolvedSchema()),
withOperationField);
- Schema schema = AvroSchemaConverter.convertToSchema(rowType);
- MessageType messageType =
ParquetTableSchemaResolver.convertAvroSchemaToParquet(schema, hadoopConf);
Review Comment:
`ParquetTableSchemaResolver` uses `hadoopConf` to control the following
variables `assumeRepeatedIsListElement`, `writeOldListStructure`,
`writeParquetUUID`, `readInt96AsFixed`, `pathsToInt96`. These affect the way
of converting the Avro schema to Parquet schema. As this PR directly uses Avro
to Spark schema conversion to replace Avro -> Parquet -> Spark schema
conversion, are these nuances in the conversion from Avro schema to Parquet
schema no longer relevant?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]