[
https://issues.apache.org/jira/browse/DRILL-5282?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Paul Rogers reassigned DRILL-5282:
----------------------------------
Assignee: Paul Rogers
> Rationalize record batch sizes in all readers and operators
> -----------------------------------------------------------
>
> Key: DRILL-5282
> URL: https://issues.apache.org/jira/browse/DRILL-5282
> Project: Apache Drill
> Issue Type: Improvement
> Affects Versions: 1.10.0
> Reporter: Paul Rogers
> Assignee: Paul Rogers
>
> Drill uses record batches to process data. A record batch consists of a
> "bundle" of vectors that, combined, hold the data for some number of records.
> The key consideration for a record batch is memory consumed. Various
> operators and readers have vastly different ideas of the size of a batch. The
> text reader can produce batches of 100s of K, while the flatten operator
> produces batches of half a GB. Other operators are randomly in between. Some
> readers produce batches of unlimited size driven by average row width.
> Another key consideration is record count. Batches have a hard physical limit
> of 64K (the number indexed by a two-byte selection vector.) Some operators
> produce this much, others far less. In one case, we saw a reader that
> produced 64K+1 records.
> A final consideration is the size of individual vectors. Drill incurs severe
> memory fragmentation when vectors grow above 16 MB.
> In some cases, operators (such as the Parquet reader) allocate large batches,
> but only partially fill them, creating a large amount of wasted space. That
> space adds up when we must buffer it during a sort.
> This ticket asks to research an optimal batch size. Create a framework to
> build such batches. Retrofit all operators that produce batches to use that
> framework to produce uniform batches.
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)