Hello Jean,

I am running cassandra 3.11.1.

Since i dont have much cassandra operations experience yet, I have a
follow-up question - how can i ensure the same token ranges distribution ?
Do i need to set initial_token configuration for each cassandra node ?

Thank you for the quick response.





On Thu, Jan 11, 2018 at 3:04 PM, Jean Carlo <jean.jeancar...@gmail.com>
wrote:

> Hello Pradeep,
>
> Actually the key here is to know if your cluster has the same token ranges
> distribution. So it is not only the same size but also the same tokens
> match node by node, from cluster source to cluster destination. In that
> case, you can use nodetool refresh.So after copy all your sstable node by
> node, it would be enough to make nodetool refresh in every node to restore
> your data. You can also restart casandra instead of doing nodetool refresh.
> It will help you to avoid the compactions after refreshing.
>
>
> Saludos
>
> Jean Carlo
>
> "The best way to predict the future is to invent it" Alan Kay
>
> On Thu, Jan 11, 2018 at 9:58 AM, Pradeep Chhetri <prad...@stashaway.com>
> wrote:
>
>> Hello everyone,
>>
>> We are running cassandra cluster inside containers over Kubernetes. We
>> have a requirement where we need to restore a fresh new cluster with
>> existing snapshot on weekly basis.
>>
>> Currently, while doing it manually. i need to copy the snapshot folder
>> inside container and then run sstableloader utility to load those tables.
>>
>> Since the source and destination cluster size is equal, I was thinking if
>> there are some easy way to just copy and paste the complete data directory
>> by mapping the nodes one to one.
>>
>> Since i wasn't able to find documentation around other  backup
>> restoration methods apart from nodetool snapshot and sstableloader, I
>> haven't explored much. I recently came across this project -
>> https://github.com/Netflix/Priam but tried it yet.
>>
>> Would be very happy if i can get some ideas around various ways of
>> backup/restoration while running inside containers.
>>
>> Thank you
>>
>
>

Reply via email to