Hello Jean, I am running cassandra 3.11.1.
Since i dont have much cassandra operations experience yet, I have a follow-up question - how can i ensure the same token ranges distribution ? Do i need to set initial_token configuration for each cassandra node ? Thank you for the quick response. On Thu, Jan 11, 2018 at 3:04 PM, Jean Carlo <jean.jeancar...@gmail.com> wrote: > Hello Pradeep, > > Actually the key here is to know if your cluster has the same token ranges > distribution. So it is not only the same size but also the same tokens > match node by node, from cluster source to cluster destination. In that > case, you can use nodetool refresh.So after copy all your sstable node by > node, it would be enough to make nodetool refresh in every node to restore > your data. You can also restart casandra instead of doing nodetool refresh. > It will help you to avoid the compactions after refreshing. > > > Saludos > > Jean Carlo > > "The best way to predict the future is to invent it" Alan Kay > > On Thu, Jan 11, 2018 at 9:58 AM, Pradeep Chhetri <prad...@stashaway.com> > wrote: > >> Hello everyone, >> >> We are running cassandra cluster inside containers over Kubernetes. We >> have a requirement where we need to restore a fresh new cluster with >> existing snapshot on weekly basis. >> >> Currently, while doing it manually. i need to copy the snapshot folder >> inside container and then run sstableloader utility to load those tables. >> >> Since the source and destination cluster size is equal, I was thinking if >> there are some easy way to just copy and paste the complete data directory >> by mapping the nodes one to one. >> >> Since i wasn't able to find documentation around other backup >> restoration methods apart from nodetool snapshot and sstableloader, I >> haven't explored much. I recently came across this project - >> https://github.com/Netflix/Priam but tried it yet. >> >> Would be very happy if i can get some ideas around various ways of >> backup/restoration while running inside containers. >> >> Thank you >> > >