[
https://issues.apache.org/jira/browse/DRILL-5516?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16016099#comment-16016099
]
ASF GitHub Bot commented on DRILL-5516:
---------------------------------------
Github user sudheeshkatkam commented on the issue:
https://github.com/apache/drill/pull/839
+1
> Limit memory usage for Hbase reader
> -----------------------------------
>
> Key: DRILL-5516
> URL: https://issues.apache.org/jira/browse/DRILL-5516
> Project: Apache Drill
> Issue Type: Improvement
> Components: Storage - HBase
> Affects Versions: 1.10.0
> Reporter: Arina Ielchiieva
> Assignee: Arina Ielchiieva
> Fix For: 1.11.0
>
>
> If early limit 0 optimization is set to true (alter session set
> `planner.enable_limit0_optimization` = true), when executing limit 0 queries
> Drill will return data type from available metadata if possible.
> When Drill can not determine data types from metadata (or if early limit 0
> optimization is set to false), Drill will read first batch of data and
> determine schema.
> Hbase reader determines max batch size using magic number (4000) which can
> lead to OOM when row size is large. The overall vector/batch size issue will
> be reconsidered in future releases.This is temporary fix to avoid OOM.
> To limit memory usage for Hbase reader we are adding max allowed allocated
> memory contant which will default to 64 mb. Thus batch size will be limited
> to 4000 (as before if memory limit does not exceed) or to number of records
> that are within max allowed memory limit. If first row in batch is larger
> than allowed default, it will be written in batch but batch will contain only
> this row.
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)