Hi,
We are also experiencing the same issue.we have 3 DCs(DC1 RF=3,DC2
RF=3,DC3,RF=1),if we use local_quorum,we are not meant to loss any data,right?
if we use local_one, maybe loss data? then we need to run repair regularly?
Could anyone advise?
Thanks
-- 原始邮件 --
发件人: "Jon Haddad";;
发送时间: 2017年7月28日(星期五) 凌晨1:37
收件人: "user";
主题: Re: Data Loss irreparabley so
We (The Last Pickle) maintain an open source tool to help manage repairs across
your clusters called Reaper. It’s a lot easier to set up and manage than
trying to manage it through cron.
http://thelastpickle.com/reaper.html
On Jul 27, 2017, at 12:38 AM, Daniel Hölbling-Inzko
wrote:
In that vein, Cassandra support Auto compaction and incremental repair.
Does this mean I have to set up cron jobs on each node to do a nodetool repair
or is this taken care of by Cassandra anyways?
How often should I run nodetool repair
Greetings Daniel
Jeff Jirsa schrieb am Do. 27. Juli 2017 um 07:48:
On 2017-07-25 15:49 (-0700), Roger Warner wrote:
> This is a quick informational question. I know that Cassandra can detect
> failures of nodes and repair them given replication and multiple DC.
>
> My question is can Cassandra tell if data was lost after a failure and
> node(s) “fixed” and resumed operation?
>
Sorta concerned by the way you're asking this - Cassandra doesn't "fix" failed
nodes. It can route requests around a down node, but the "fixing" is entirely
manual.
If you have a node go down temporarily, and it comes back up (with it's disk
intact), you can see it "repair" data with a combination of active
(anti-entropy) repair via nodetool repair, or by watching 'nodetool netstats'
and see the read repair counters increase over time (which will happen
naturally as data is requested and mismatches are detected in the data, based
on your consistency level).
-
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org