Repair will take way more time then rebuild. On Wed, Oct 31, 2018 at 6:45 AM Kiran mk <coolkiran2...@gmail.com> wrote:
> Run the repair with -pr option on each node which will repair only the > > parition range. > > > > nodetool repair -pr > > On Wed, Oct 31, 2018 at 7:04 PM Surbhi Gupta <surbhi.gupt...@gmail.com> > wrote: > > > > > > Nodetool repair will take way more time than nodetool rebuild. > > > How much data u have in your original data center? > > > Repair should be run to make the data consistent in case of node down > more than hintedhandoff period and dropped mutations. > > > But as a thumb rule ,generally we run repair using opscenter (if using > Datastax) most of the times. > > > > > > So in your case run “nodetool rebuild <original data enter>” on all the > nodes in new data center. > > > For making the rebuild process fast, increase three parameters, > compaction throughput , stream throughput and interdcstream throughput. > > > > > > Thanks > > > Surbhi > > > On Tue, Oct 30, 2018 at 11:29 PM Akshay Bhardwaj < > akshay.bhardwaj1...@gmail.com> wrote: > > >> > > >> Hi Jonathan, > > >> > > >> That makes sense. Thank you for the explanation. > > >> > > >> Another quick question, as the cluster is still operative and the data > for the past 2 weeks (since updating replication factor) is present in both > the data centres, should I run "nodetool rebuild" or "nodetool repair"? > > >> > > >> I read that nodetool rebuild is faster and is useful till the new data > centre is empty and no partition keys are present. So when is the good time > to use either of the commands and what impact can it have on the data > centre operations? > > >> > > >> Thanks and Regards > > >> > > >> Akshay Bhardwaj > > >> +91-97111-33849 > > >> > > >> > > >> On Wed, Oct 31, 2018 at 2:34 AM Jonathan Haddad <j...@jonhaddad.com> > wrote: > > >>> > > >>> You need to run "nodetool rebuild -- <existing-dc-name>" on each node > in the new DC to get the old data to replicate. It doesn't do it > automatically because Cassandra has no way of knowing if you're done adding > nodes and if it were to migrate automatically, it could cause a lot of > problems. Imagine streaming 100 nodes data to 3 nodes in the new DC, not > fun. > > >>> > > >>> On Tue, Oct 30, 2018 at 1:59 PM Akshay Bhardwaj < > akshay.bhardwaj1...@gmail.com> wrote: > > >>>> > > >>>> Hi Experts, > > >>>> > > >>>> I previously had 1 Cassandra data centre in AWS Singapore region with > 5 nodes, with my keyspace's replication factor as 3 in Network topology. > > >>>> > > >>>> After this cluster has been running smoothly for 4 months (500 GB of > data on each node's disk), I added 2nd data centre in AWS Mumbai region > with yet again 5 nodes in Network topology. > > >>>> > > >>>> After updating my keyspace's replication factor to > {"AWS_Sgp":3,"AWS_Mum":3}, my expectation was that the data present in Sgp > region will immediately start replicating on the Mum region's nodes. > However even after 2 weeks I do not see historical data to be replicated, > but new data being written on Sgp region is present in Mum region as well. > > >>>> > > >>>> Any help or suggestions to debug this issue will be highly > appreciated. > > >>>> > > >>>> Regards > > >>>> Akshay Bhardwaj > > >>>> +91-97111-33849 > > >>>> > > >>>> > > >>> > > >>> > > >>> -- > > >>> Jon Haddad > > >>> http://www.rustyrazorblade.com > > >>> twitter: rustyrazorblade > > >>> > > >>> > > >> > > >> > > > > > > -- > > Best Regards, > > Kiran.M.K. > > > > --------------------------------------------------------------------- > > To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org > > For additional commands, e-mail: user-h...@cassandra.apache.org > > > >