[ 
https://issues.apache.org/jira/browse/CARBONDATA-1224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kumar vishal resolved CARBONDATA-1224.
--------------------------------------
       Resolution: Fixed
    Fix Version/s: 1.3.0

> Going out of memory if more segments are compacted at once in V3 format
> -----------------------------------------------------------------------
>
>                 Key: CARBONDATA-1224
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-1224
>             Project: CarbonData
>          Issue Type: Bug
>            Reporter: Ravindra Pesala
>            Assignee: Ravindra Pesala
>            Priority: Major
>             Fix For: 1.3.0
>
>          Time Spent: 5h 50m
>  Remaining Estimate: 0h
>
> In V3 format we read the whole blocklet at once to memory in order save IO 
> time. But it turns out  to be costlier in case of parallel reading of more 
> carbondata files. 
> For example if we need to compact 50 segments then compactor need to open the 
> readers on all the 50 segments to do merge sort. But the memory consumption 
> is too high if each reader reads whole blocklet to the memory and there is 
> high chances of going out of memory.
> Solution:
> In this type of scenarios we can introduce new readers for V3 to read the 
> data page by page instead of reading whole blocklet at once to reduce the 
> memory footprint.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to