[
https://issues.apache.org/jira/browse/CASSANDRA-8177?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14183783#comment-14183783
]
Sean Bridges edited comment on CASSANDRA-8177 at 10/25/14 12:18 AM:
{quote}
My guess for sequential repair generating lots of IO is that, when reading from
snapshot, it is hitting disk for each snapshot SSTable to read its bloom
filters, index files etc
{quote}
When you snapshot you are hardlinking the old and original sstables, they are
the same files, so the os cache shouldn't be the difference
was (Author: sgbridges):
{quote}
My guess for sequential repair generating lots of IO is that, when reading from
snapshot, it is hitting disk for each snapshot SSTable to read its bloom
filters, index files etc
{quote}
When you snapshot you are hardlinking the old and original sstables, they are
the same file, so the os cache shouldn't be the difference
sequential repair is much more expensive than parallel repair
-
Key: CASSANDRA-8177
URL: https://issues.apache.org/jira/browse/CASSANDRA-8177
Project: Cassandra
Issue Type: Bug
Reporter: Sean Bridges
Assignee: Yuki Morishita
Attachments: cassc-week.png, iostats.png
This is with 2.0.10
The attached graph shows io read/write throughput (as measured with iostat)
when doing repairs.
The large hump on the left is a sequential repair of one node. The two much
smaller peaks on the right are parallel repairs.
This is a 3 node cluster using vnodes (I know vnodes on small clusters isn't
recommended). Cassandra reports load of 40 gigs.
We noticed a similar problem with a larger cluster.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)