node- communication network cards on
your C* host machines.
§ If possible, reduce # of vnodes!
From: Chris Stokesmore [mailto:chris.elsm...@demandlogic.co]
Sent: Monday, June 19, 2017 4:50 AM
To: anujw_2...@yahoo.co.in
Cc: user@cassandra.apache.org
Subject: Re: Partition range incremental rep
Anyone have anymore thoughts on this at all? Struggling to understand it..
> On 9 Jun 2017, at 11:32, Chris Stokesmore
> wrote:
>
> Hi Anuj,
>
> Thanks for the reply.
>
> 1). We are using Cassandra 2.2.8, and our repair commands we are comparing
> are
>
>
> I can't recommend *anyone* use incremental repair as there's some pretty
> horrible bugs in it that can cause Merkle trees to wildly mismatch & result
> in massive overstreaming. Check out
> https://issues.apache.org/jira/browse/CASSANDRA-9143
>
Hi Anuj,
Thanks for the reply.
1). We are using Cassandra 2.2.8, and our repair commands we are comparing are
"nodetool repair --in-local-dc --partitioner-range” and
"nodetool repair --in-local-dc”
Since 2.2 I believe inc repairs are the default - that seems to be confirmed in
the logs that
I can't recommend *anyone* use incremental repair as there's some pretty
horrible bugs in it that can cause Merkle trees to wildly mismatch & result
in massive overstreaming. Check out
https://issues.apache.org/jira/browse/CASSANDRA-9143.
TL;DR: Do not use incremental repair before 4.0.
On Tue,
Hi Chris,
Can your share following info:
1. Exact repair commands you use for inc repair and pr repair
2. Repair time should be measured at cluster level for inc repair. So, whats
the total time it takes to run repair on all nodes for incremental vs pr
repairs?
3. You are repairing one dc DC3.
Thank you for the excellent and clear description of the different versions of
repair Anuj, that has cleared up what I expect to be happening.
The problem now is in our cluster, we are running repairs with options
(parallelism: parallel, primary range: false, incremental: true, job threads:
1,
Hi Chris,
Using pr with incremental repairs does not make sense. Primary range repair is
an optimization over full repair. If you run full repair on a n node cluster
with RF=3, you would be repairing each data thrice. E.g. in a 5 node cluster
with RF=3, a range may exist on node A,B and C .
Hi all,
Wondering if anyone had any thoughts on this? At the moment the long running
repairs cause us to be running them on two nodes at once for a bit of time,
which obivould increases the cluster load.
On 2017-05-25 16:18 (+0100), Chris Stokesmore wrote:
> Hi,>
>
>
Hi,
We are running a 7 node Cassandra 2.2.8 cluster, RF=3, and had been running
repairs with the —pr option, via a cron job that runs on each node once per
week.
We changed that as some advice on the Cassandra IRC channel said it would cause
more anticompaction and
10 matches
Mail list logo