TheR1sing3un commented on code in PR #13127:
URL: https://github.com/apache/hudi/pull/13127#discussion_r2103715541


##########
hudi-spark-datasource/hudi-spark3.3.x/src/main/scala/org/apache/spark/sql/execution/datasources/parquet/Spark33ParquetReader.scala:
##########
@@ -252,6 +252,17 @@ object Spark33ParquetReader extends 
SparkParquetReaderBuilder {
       sqlConf.getConfString("spark.sql.legacy.parquet.nanosAsLong", 
"false").toBoolean
     )
 
+    // Should always be set by FileSourceScanExec creating this.
+    // Check conf before checking option, to allow working around an issue by 
changing conf.
+    val returningBatch = sqlConf.parquetVectorizedReaderEnabled &&

Review Comment:
   > should we fix the other version of parquet readers?
   
   Only fix spark 3.3 is fine, because the relevant changes of spark were 
integrated in versions after 3.3, only 3.3 needs to be compatible, refer to 
https://github.com/apache/spark/pull/38397



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to