[
https://issues.apache.org/jira/browse/DRILL-6032?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16354378#comment-16354378
]
ASF GitHub Bot commented on DRILL-6032:
---------------------------------------
Github user ilooner commented on a diff in the pull request:
https://github.com/apache/drill/pull/1101#discussion_r166409878
--- Diff:
exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java
---
@@ -226,7 +221,7 @@ public BatchHolder() {
((FixedWidthVector) vector).allocateNew(HashTable.BATCH_SIZE);
} else if (vector instanceof VariableWidthVector) {
// This case is never used .... a varchar falls under
ObjectVector which is allocated on the heap !
- ((VariableWidthVector) vector).allocateNew(maxColumnWidth,
HashTable.BATCH_SIZE);
+ ((VariableWidthVector) vector).allocateNew(columnSize,
HashTable.BATCH_SIZE);
--- End diff --
Thanks for catching this. It should use stdSize here.
> Use RecordBatchSizer to estimate size of columns in HashAgg
> -----------------------------------------------------------
>
> Key: DRILL-6032
> URL: https://issues.apache.org/jira/browse/DRILL-6032
> Project: Apache Drill
> Issue Type: Improvement
> Reporter: Timothy Farkas
> Assignee: Timothy Farkas
> Priority: Major
> Fix For: 1.13.0
>
>
> We need to use the RecordBatchSize to estimate the size of columns in the
> Partition batches created by HashAgg.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)