Github user gatorsmile commented on the issue:
https://github.com/apache/spark/pull/14445
Maybe, we can correct the PR description and make it more accurate.
This PR avoids the extra memory copy when the vectorized parquet record
reader is not being used for reading a non-partitioned Parquet table. One of
the typical case is the parquet table with non atomic types, including null,
UDTs, arrays, structs, and maps. Another case is when users set
`spark.sql.parquet.enableVectorizedReader` to `false`.
Is my understanding correct?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]