[ https://issues.apache.org/jira/browse/CASSANDRA-12707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
nicolas ginder updated CASSANDRA-12707: --------------------------------------- Fix Version/s: (was: 2.2.x) (was: 2.1.x) > JVM out of memory when querying an extra-large partition with lots of > tombstones > -------------------------------------------------------------------------------- > > Key: CASSANDRA-12707 > URL: https://issues.apache.org/jira/browse/CASSANDRA-12707 > Project: Cassandra > Issue Type: Bug > Components: Core > Reporter: nicolas ginder > > We have an extra large partition of 40 million cells where most of the cells > were deleted. When querying this partition with a slice query, Cassandra runs > out of memory as tombstones fill up the JVM heap. After debugging one of the > large SSTable we found that this part of the code loads all the tombstones. > In org.apache.cassandra.db.filter.QueryFilter > ... > public static Iterator<Cell> gatherTombstones(final ColumnFamily returnCF, > final Iterator<? extends OnDiskAtom> iter) > { > ... > while (iter.hasNext()) > { > OnDiskAtom atom = iter.next(); > if (atom instanceof Cell) > { > next = (Cell)atom; > break; > } > else > { > returnCF.addAtom(atom); > } > } > ... -- This message was sent by Atlassian JIRA (v6.3.4#6332)