Re: Cassandra | Cross Data Centre Replication Status

2018-10-31 Thread Alexander Dejanovski
Akshay, avoid running repair in that case, it'll take way longer than rebuild and it will stream data back to your original DC, even between nodes in that original DC, which is not what you're running after, and could lead to all sorts of troubles. Run "nodetool rebuild " as recommended by Jon

Re: Cassandra | Cross Data Centre Replication Status

2018-10-31 Thread Surbhi Gupta
Repair will take way more time then rebuild. On Wed, Oct 31, 2018 at 6:45 AM Kiran mk wrote: > Run the repair with -pr option on each node which will repair only the > > parition range. > > > > nodetool repair -pr > > On Wed, Oct 31, 2018 at 7:04 PM Surbhi Gupta > wrote: > > > > > > Nodetool

Re: Cassandra | Cross Data Centre Replication Status

2018-10-31 Thread Kiran mk
Run the repair with -pr option on each node which will repair only the parition range. nodetool repair -pr On Wed, Oct 31, 2018 at 7:04 PM Surbhi Gupta wrote: > > Nodetool repair will take way more time than nodetool rebuild. > How much data u have in your original data center? > Repair should

Re: Cassandra | Cross Data Centre Replication Status

2018-10-31 Thread Surbhi Gupta
Nodetool repair will take way more time than nodetool rebuild. How much data u have in your original data center? Repair should be run to make the data consistent in case of node down more than hintedhandoff period and dropped mutations. But as a thumb rule ,generally we run repair using opscenter

Re: Cassandra | Cross Data Centre Replication Status

2018-10-31 Thread Akshay Bhardwaj
Hi Jonathan, That makes sense. Thank you for the explanation. Another quick question, as the cluster is still operative and the data for the past 2 weeks (since updating replication factor) is present in both the data centres, should I run "nodetool rebuild" or "nodetool repair"? I read that

Re: Cassandra | Cross Data Centre Replication Status

2018-10-30 Thread Jonathan Haddad
You need to run "nodetool rebuild -- " on each node in the new DC to get the old data to replicate. It doesn't do it automatically because Cassandra has no way of knowing if you're done adding nodes and if it were to migrate automatically, it could cause a lot of problems. Imagine streaming 100

Cassandra | Cross Data Centre Replication Status

2018-10-30 Thread Akshay Bhardwaj
Hi Experts, I previously had 1 Cassandra data centre in AWS Singapore region with 5 nodes, with my keyspace's replication factor as 3 in Network topology. After this cluster has been running smoothly for 4 months (500 GB of data on each node's disk), I added 2nd data centre in AWS Mumbai region