Does anyone know how and if the LIMIT push down to Parquet file works?
I have a parquet file with 53K records in 1 row group. When I run a SELECT
* from <table> LIMIT 1, I see the Parquet reader operator process 32768
records. I would have expected either 1 or 53K. So questions;
1) Does the Parquet MR library offer the ability to push down LIMITs to
Parquet files? From the above, the answer looks like yes.
2) If so, how does Drill come up with the magic number 32767? Is there a
way I can make it read just 1 row if the query is a LIMIT 1?
Product Specialist - SQL on Hadoop
*MapR Technologies (www.mapr.com <http://www.mapr.com>)*
*(Email) vnaranammalpu...@maprtech.com <naranammalpu...@maprtech.com>*
*(Mobile) 917 683 8116 - can text *
*Timezone: ET (UTC -5:00 / -4:00)*