Hello, Basically, every node has to have the same token range. So yes you have to play with initial_token having the same numbers of tokens per node like the cluster source. To save time and if you dont have any constraints about the name of the cluster etc. you can just copy and paste the complete keyspace system node by node.
So you will have the same cluster( cluster name, confs, etc) Saludos Jean Carlo "The best way to predict the future is to invent it" Alan Kay On Thu, Jan 11, 2018 at 10:28 AM, Pradeep Chhetri <prad...@stashaway.com> wrote: > Hello Jean, > > I am running cassandra 3.11.1. > > Since i dont have much cassandra operations experience yet, I have a > follow-up question - how can i ensure the same token ranges distribution ? > Do i need to set initial_token configuration for each cassandra node ? > > Thank you for the quick response. > > > > > > On Thu, Jan 11, 2018 at 3:04 PM, Jean Carlo <jean.jeancar...@gmail.com> > wrote: > >> Hello Pradeep, >> >> Actually the key here is to know if your cluster has the same token >> ranges distribution. So it is not only the same size but also the same >> tokens match node by node, from cluster source to cluster destination. In >> that case, you can use nodetool refresh.So after copy all your sstable node >> by node, it would be enough to make nodetool refresh in every node to >> restore your data. You can also restart casandra instead of doing nodetool >> refresh. It will help you to avoid the compactions after refreshing. >> >> >> Saludos >> >> Jean Carlo >> >> "The best way to predict the future is to invent it" Alan Kay >> >> On Thu, Jan 11, 2018 at 9:58 AM, Pradeep Chhetri <prad...@stashaway.com> >> wrote: >> >>> Hello everyone, >>> >>> We are running cassandra cluster inside containers over Kubernetes. We >>> have a requirement where we need to restore a fresh new cluster with >>> existing snapshot on weekly basis. >>> >>> Currently, while doing it manually. i need to copy the snapshot folder >>> inside container and then run sstableloader utility to load those tables. >>> >>> Since the source and destination cluster size is equal, I was thinking >>> if there are some easy way to just copy and paste the complete data >>> directory by mapping the nodes one to one. >>> >>> Since i wasn't able to find documentation around other backup >>> restoration methods apart from nodetool snapshot and sstableloader, I >>> haven't explored much. I recently came across this project - >>> https://github.com/Netflix/Priam but tried it yet. >>> >>> Would be very happy if i can get some ideas around various ways of >>> backup/restoration while running inside containers. >>> >>> Thank you >>> >> >> >