[ 
https://issues.apache.org/jira/browse/CASSANDRA-12707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

nicolas ginder updated CASSANDRA-12707:
---------------------------------------
    Fix Version/s:     (was: 2.2.x)
                       (was: 2.1.x)

> JVM out of memory when querying an extra-large partition with lots of 
> tombstones
> --------------------------------------------------------------------------------
>
>                 Key: CASSANDRA-12707
>                 URL: https://issues.apache.org/jira/browse/CASSANDRA-12707
>             Project: Cassandra
>          Issue Type: Bug
>          Components: Core
>            Reporter: nicolas ginder
>
> We have an extra large partition of 40 million cells where most of the cells 
> were deleted. When querying this partition with a slice query, Cassandra runs 
> out of memory as tombstones fill up the JVM heap. After debugging one of the 
> large SSTable we found that this part of the code loads all the tombstones.
> In org.apache.cassandra.db.filter.QueryFilter
> ...
> public static Iterator<Cell> gatherTombstones(final ColumnFamily returnCF, 
> final Iterator<? extends OnDiskAtom> iter)
>     {
> ...
> while (iter.hasNext())
>                 {
>                     OnDiskAtom atom = iter.next();
>                     if (atom instanceof Cell)
>                     {
>                         next = (Cell)atom;
>                         break;
>                     }
>                     else
>                     {
>                         returnCF.addAtom(atom);
>                     }
>                 }
> ...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to