[GitHub] [spark] viirya commented on a change in pull request #27888: [SPARK-31116][SQL] Fix nested schema case-sensitivity in ParquetRowConverter
viirya commented on a change in pull request #27888: [SPARK-31116][SQL] Fix nested schema case-sensitivity in ParquetRowConverter URL: https://github.com/apache/spark/pull/27888#discussion_r393191381 ## File path: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetRowConverter.scala ## @@ -178,8 +179,17 @@ private[parquet] class ParquetRowConverter( // Converters for each field. private[this] val fieldConverters: Array[Converter with HasParentContainerUpdater] = { Review comment: Yea, thanks @dongjoon-hyun. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] [spark] viirya commented on a change in pull request #27888: [SPARK-31116][SQL] Fix nested schema case-sensitivity in ParquetRowConverter
viirya commented on a change in pull request #27888: [SPARK-31116][SQL] Fix nested schema case-sensitivity in ParquetRowConverter URL: https://github.com/apache/spark/pull/27888#discussion_r392838395 ## File path: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetReadSupport.scala ## @@ -122,12 +122,15 @@ class ParquetReadSupport(val convertTz: Option[ZoneId], keyValueMetaData: JMap[String, String], fileSchema: MessageType, readContext: ReadContext): RecordMaterializer[InternalRow] = { +val caseSensitive = conf.getBoolean(SQLConf.CASE_SENSITIVE.key, + SQLConf.CASE_SENSITIVE.defaultValue.get) Review comment: I'm not sure why you need to pass `caseSensitive` across `ParquetRecordMaterializer`, `ParquetRowConverter`. Can't we just get it at `ParquetRowConverter`? This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] [spark] viirya commented on a change in pull request #27888: [SPARK-31116][SQL] Fix nested schema case-sensitivity in ParquetRowConverter
viirya commented on a change in pull request #27888: [SPARK-31116][SQL] Fix nested schema case-sensitivity in ParquetRowConverter URL: https://github.com/apache/spark/pull/27888#discussion_r392786863 ## File path: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetRowConverter.scala ## @@ -176,15 +178,38 @@ private[parquet] class ParquetRowConverter( */ def currentRecord: InternalRow = currentRow + // Converters for each field. private[this] val fieldConverters: Array[Converter with HasParentContainerUpdater] = { -parquetType.getFields.asScala.map { parquetField => - val fieldIndex = catalystType.fieldIndex(parquetField.getName) - val catalystField = catalystType(fieldIndex) - // Converted field value should be set to the `fieldIndex`-th cell of `currentRow` - newConverter(parquetField, catalystField.dataType, new RowUpdater(currentRow, fieldIndex)) -}.toArray - } + +// (SPARK-31116) There is an issue when schema pruning is enabled, so we keep original codes +if (schemaPruning) { + // (SPARK-31116) For letter case issue, create name to field index based on case sensitivity + val catalystFieldNameToIndex = if (caseSensitive) { +catalystType.fieldNames.zipWithIndex.toMap + } else { +CaseInsensitiveMap(catalystType.fieldNames.zipWithIndex.toMap) + } + parquetType.getFields.asScala.map { parquetField => +val fieldIndex = catalystFieldNameToIndex.getOrElse(parquetField.getName, + throw new IllegalArgumentException( +s"${parquetField.getName} does not exist. " + + s"Available: ${catalystType.fieldNames.mkString(", ")}") +) +val catalystField = catalystType(fieldIndex) +// Converted field value should be set to the `fieldIndex`-th cell of `currentRow` +newConverter(parquetField, catalystField.dataType, new RowUpdater(currentRow, fieldIndex)) + }.toArray +} else { + parquetType.getFields.asScala.zip(catalystType).zipWithIndex.map { +case ((parquetFieldType, catalystField), ordinal) => + // Converted field value should be set to the `ordinal`-th cell of `currentRow` + newConverter( +parquetFieldType, catalystField.dataType, new RowUpdater(currentRow, ordinal)) + }.toArray +} Review comment: Why add this part of code? Seems to me the above code inside `if (schemaPruning) { ... }` looks reasonable. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org
[GitHub] [spark] viirya commented on a change in pull request #27888: [SPARK-31116][SQL] Fix nested schema case-sensitivity in ParquetRowConverter
viirya commented on a change in pull request #27888: [SPARK-31116][SQL] Fix nested schema case-sensitivity in ParquetRowConverter URL: https://github.com/apache/spark/pull/27888#discussion_r392785966 ## File path: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetRowConverter.scala ## @@ -176,15 +178,38 @@ private[parquet] class ParquetRowConverter( */ def currentRecord: InternalRow = currentRow + Review comment: nit: remove unnecessary blank link. This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services - To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org For additional commands, e-mail: reviews-h...@spark.apache.org