[ https://issues.apache.org/jira/browse/CASSANDRA-10680?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14999459#comment-14999459 ]
Jeff Jirsa commented on CASSANDRA-10680: ---------------------------------------- Initial tests suggest this (commit 61d2630e9950e9abc0d8da3939b280ff44b5ddc0) does indeed solve this issue. > Deal with small compression chunk size better during streaming plan setup > ------------------------------------------------------------------------- > > Key: CASSANDRA-10680 > URL: https://issues.apache.org/jira/browse/CASSANDRA-10680 > Project: Cassandra > Issue Type: Bug > Reporter: Jeff Jirsa > Assignee: Yuki Morishita > Fix For: 2.1.x > > > For clusters using small compression chunk size and terabytes of data, the > streaming plan calculations will instantiate hundreds of millions of > compressionmetadata$chunk objects, which will create unreasonable amounts of > heap pressure. Rather than instantiating all of those at once, streaming > should instantiate only as many as needed for a single file per table at a > time. -- This message was sent by Atlassian JIRA (v6.3.4#6332)