Re: Restoring fresh new cluster from existing snapshot

2018-01-11 Thread Pradeep Chhetri
Hello Jeff,

Thank you for the reply.

One doubt - If i copy the /var/lib/cassandra one to one from source cluster
to destination cluster nodes and change the cluster name in configuration
and delete system.peers table and restart each cassandra node, do you think
the cluster will come up properly. Although the clusters are in different
VPC in AWS but still i want to make sure that they get different
clustername.

Regards.

On Thu, Jan 11, 2018 at 9:37 PM, Jeff Jirsa  wrote:

> Make sure the new cluster has a different cluster name, and avoid copying
> system.peers if you can avoid it. Doing so risks merging your new cluster
> and old cluster if they’re able to reach each other.
>
> --
> Jeff Jirsa
>
>
> On Jan 11, 2018, at 1:41 AM, Pradeep Chhetri 
> wrote:
>
> Thank you very much Jean. Since i don't have any constraints, as you said,
> i will try copying the complete keyspace system node by node first and will
> do nodetool refresh and see if it works.
>
>
>
> On Thu, Jan 11, 2018 at 3:21 PM, Jean Carlo 
> wrote:
>
>> Hello,
>>
>> Basically, every node has to have the same token range. So yes you have
>> to play with initial_token having the same numbers of tokens per node like
>> the cluster source. To save time and if you dont have any constraints about
>> the name of the cluster etc. you can just copy and paste the complete
>> keyspace system node by node.
>>
>> So you will have the same cluster( cluster name, confs, etc)
>>
>>
>> Saludos
>>
>> Jean Carlo
>>
>> "The best way to predict the future is to invent it" Alan Kay
>>
>> On Thu, Jan 11, 2018 at 10:28 AM, Pradeep Chhetri 
>> wrote:
>>
>>> Hello Jean,
>>>
>>> I am running cassandra 3.11.1.
>>>
>>> Since i dont have much cassandra operations experience yet, I have a
>>> follow-up question - how can i ensure the same token ranges distribution ?
>>> Do i need to set initial_token configuration for each cassandra node ?
>>>
>>> Thank you for the quick response.
>>>
>>>
>>>
>>>
>>>
>>> On Thu, Jan 11, 2018 at 3:04 PM, Jean Carlo 
>>> wrote:
>>>
 Hello Pradeep,

 Actually the key here is to know if your cluster has the same token
 ranges distribution. So it is not only the same size but also the same
 tokens match node by node, from cluster source to cluster destination. In
 that case, you can use nodetool refresh.So after copy all your sstable node
 by node, it would be enough to make nodetool refresh in every node to
 restore your data. You can also restart casandra instead of doing nodetool
 refresh. It will help you to avoid the compactions after refreshing.


 Saludos

 Jean Carlo

 "The best way to predict the future is to invent it" Alan Kay

 On Thu, Jan 11, 2018 at 9:58 AM, Pradeep Chhetri  wrote:

> Hello everyone,
>
> We are running cassandra cluster inside containers over Kubernetes. We
> have a requirement where we need to restore a fresh new cluster with
> existing snapshot on weekly basis.
>
> Currently, while doing it manually. i need to copy the snapshot folder
> inside container and then run sstableloader utility to load those tables.
>
> Since the source and destination cluster size is equal, I was thinking
> if there are some easy way to just copy and paste the complete data
> directory by mapping the nodes one to one.
>
> Since i wasn't able to find documentation around other  backup
> restoration methods apart from nodetool snapshot and sstableloader, I
> haven't explored much. I recently came across this project -
> https://github.com/Netflix/Priam but tried it yet.
>
> Would be very happy if i can get some ideas around various ways of
> backup/restoration while running inside containers.
>
> Thank you
>


>>>
>>
>


Re: Restoring fresh new cluster from existing snapshot

2018-01-11 Thread Jeff Jirsa
Make sure the new cluster has a different cluster name, and avoid copying 
system.peers if you can avoid it. Doing so risks merging your new cluster and 
old cluster if they’re able to reach each other.

-- 
Jeff Jirsa


> On Jan 11, 2018, at 1:41 AM, Pradeep Chhetri  wrote:
> 
> Thank you very much Jean. Since i don't have any constraints, as you said, i 
> will try copying the complete keyspace system node by node first and will do 
> nodetool refresh and see if it works. 
> 
> 
> 
>> On Thu, Jan 11, 2018 at 3:21 PM, Jean Carlo  
>> wrote:
>> Hello,
>> 
>> Basically, every node has to have the same token range. So yes you have to 
>> play with initial_token having the same numbers of tokens per node like the 
>> cluster source. To save time and if you dont have any constraints about the 
>> name of the cluster etc. you can just copy and paste the complete keyspace 
>> system node by node. 
>> 
>> So you will have the same cluster( cluster name, confs, etc) 
>> 
>> 
>> Saludos
>> 
>> Jean Carlo
>> 
>> "The best way to predict the future is to invent it" Alan Kay
>> 
>>> On Thu, Jan 11, 2018 at 10:28 AM, Pradeep Chhetri  
>>> wrote:
>>> Hello Jean,
>>> 
>>> I am running cassandra 3.11.1.
>>> 
>>> Since i dont have much cassandra operations experience yet, I have a 
>>> follow-up question - how can i ensure the same token ranges distribution ? 
>>> Do i need to set initial_token configuration for each cassandra node ?
>>> 
>>> Thank you for the quick response.
>>> 
>>> 
>>> 
>>> 
>>> 
 On Thu, Jan 11, 2018 at 3:04 PM, Jean Carlo  
 wrote:
 Hello Pradeep,
 
 Actually the key here is to know if your cluster has the same token ranges 
 distribution. So it is not only the same size but also the same tokens 
 match node by node, from cluster source to cluster destination. In that 
 case, you can use nodetool refresh.So after copy all your sstable node by 
 node, it would be enough to make nodetool refresh in every node to restore 
 your data. You can also restart casandra instead of doing nodetool 
 refresh. It will help you to avoid the compactions after refreshing.
 
 
 Saludos
 
 Jean Carlo
 
 "The best way to predict the future is to invent it" Alan Kay
 
> On Thu, Jan 11, 2018 at 9:58 AM, Pradeep Chhetri  
> wrote:
> Hello everyone,
> 
> We are running cassandra cluster inside containers over Kubernetes. We 
> have a requirement where we need to restore a fresh new cluster with 
> existing snapshot on weekly basis.
> 
> Currently, while doing it manually. i need to copy the snapshot folder 
> inside container and then run sstableloader utility to load those tables.
> 
> Since the source and destination cluster size is equal, I was thinking if 
> there are some easy way to just copy and paste the complete data 
> directory by mapping the nodes one to one.
> 
> Since i wasn't able to find documentation around other  backup 
> restoration methods apart from nodetool snapshot and sstableloader, I 
> haven't explored much. I recently came across this project - 
> https://github.com/Netflix/Priam but tried it yet.
> 
> Would be very happy if i can get some ideas around various ways of 
> backup/restoration while running inside containers.
> 
> Thank you
 
>>> 
>> 
> 


Re: Restoring fresh new cluster from existing snapshot

2018-01-11 Thread Pradeep Chhetri
Thank you very much Jean. Since i don't have any constraints, as you said,
i will try copying the complete keyspace system node by node first and will
do nodetool refresh and see if it works.



On Thu, Jan 11, 2018 at 3:21 PM, Jean Carlo 
wrote:

> Hello,
>
> Basically, every node has to have the same token range. So yes you have to
> play with initial_token having the same numbers of tokens per node like the
> cluster source. To save time and if you dont have any constraints about the
> name of the cluster etc. you can just copy and paste the complete keyspace
> system node by node.
>
> So you will have the same cluster( cluster name, confs, etc)
>
>
> Saludos
>
> Jean Carlo
>
> "The best way to predict the future is to invent it" Alan Kay
>
> On Thu, Jan 11, 2018 at 10:28 AM, Pradeep Chhetri 
> wrote:
>
>> Hello Jean,
>>
>> I am running cassandra 3.11.1.
>>
>> Since i dont have much cassandra operations experience yet, I have a
>> follow-up question - how can i ensure the same token ranges distribution ?
>> Do i need to set initial_token configuration for each cassandra node ?
>>
>> Thank you for the quick response.
>>
>>
>>
>>
>>
>> On Thu, Jan 11, 2018 at 3:04 PM, Jean Carlo 
>> wrote:
>>
>>> Hello Pradeep,
>>>
>>> Actually the key here is to know if your cluster has the same token
>>> ranges distribution. So it is not only the same size but also the same
>>> tokens match node by node, from cluster source to cluster destination. In
>>> that case, you can use nodetool refresh.So after copy all your sstable node
>>> by node, it would be enough to make nodetool refresh in every node to
>>> restore your data. You can also restart casandra instead of doing nodetool
>>> refresh. It will help you to avoid the compactions after refreshing.
>>>
>>>
>>> Saludos
>>>
>>> Jean Carlo
>>>
>>> "The best way to predict the future is to invent it" Alan Kay
>>>
>>> On Thu, Jan 11, 2018 at 9:58 AM, Pradeep Chhetri 
>>> wrote:
>>>
 Hello everyone,

 We are running cassandra cluster inside containers over Kubernetes. We
 have a requirement where we need to restore a fresh new cluster with
 existing snapshot on weekly basis.

 Currently, while doing it manually. i need to copy the snapshot folder
 inside container and then run sstableloader utility to load those tables.

 Since the source and destination cluster size is equal, I was thinking
 if there are some easy way to just copy and paste the complete data
 directory by mapping the nodes one to one.

 Since i wasn't able to find documentation around other  backup
 restoration methods apart from nodetool snapshot and sstableloader, I
 haven't explored much. I recently came across this project -
 https://github.com/Netflix/Priam but tried it yet.

 Would be very happy if i can get some ideas around various ways of
 backup/restoration while running inside containers.

 Thank you

>>>
>>>
>>
>


Re: Restoring fresh new cluster from existing snapshot

2018-01-11 Thread Jean Carlo
Hello,

Basically, every node has to have the same token range. So yes you have to
play with initial_token having the same numbers of tokens per node like the
cluster source. To save time and if you dont have any constraints about the
name of the cluster etc. you can just copy and paste the complete keyspace
system node by node.

So you will have the same cluster( cluster name, confs, etc)


Saludos

Jean Carlo

"The best way to predict the future is to invent it" Alan Kay

On Thu, Jan 11, 2018 at 10:28 AM, Pradeep Chhetri 
wrote:

> Hello Jean,
>
> I am running cassandra 3.11.1.
>
> Since i dont have much cassandra operations experience yet, I have a
> follow-up question - how can i ensure the same token ranges distribution ?
> Do i need to set initial_token configuration for each cassandra node ?
>
> Thank you for the quick response.
>
>
>
>
>
> On Thu, Jan 11, 2018 at 3:04 PM, Jean Carlo 
> wrote:
>
>> Hello Pradeep,
>>
>> Actually the key here is to know if your cluster has the same token
>> ranges distribution. So it is not only the same size but also the same
>> tokens match node by node, from cluster source to cluster destination. In
>> that case, you can use nodetool refresh.So after copy all your sstable node
>> by node, it would be enough to make nodetool refresh in every node to
>> restore your data. You can also restart casandra instead of doing nodetool
>> refresh. It will help you to avoid the compactions after refreshing.
>>
>>
>> Saludos
>>
>> Jean Carlo
>>
>> "The best way to predict the future is to invent it" Alan Kay
>>
>> On Thu, Jan 11, 2018 at 9:58 AM, Pradeep Chhetri 
>> wrote:
>>
>>> Hello everyone,
>>>
>>> We are running cassandra cluster inside containers over Kubernetes. We
>>> have a requirement where we need to restore a fresh new cluster with
>>> existing snapshot on weekly basis.
>>>
>>> Currently, while doing it manually. i need to copy the snapshot folder
>>> inside container and then run sstableloader utility to load those tables.
>>>
>>> Since the source and destination cluster size is equal, I was thinking
>>> if there are some easy way to just copy and paste the complete data
>>> directory by mapping the nodes one to one.
>>>
>>> Since i wasn't able to find documentation around other  backup
>>> restoration methods apart from nodetool snapshot and sstableloader, I
>>> haven't explored much. I recently came across this project -
>>> https://github.com/Netflix/Priam but tried it yet.
>>>
>>> Would be very happy if i can get some ideas around various ways of
>>> backup/restoration while running inside containers.
>>>
>>> Thank you
>>>
>>
>>
>


Re: Restoring fresh new cluster from existing snapshot

2018-01-11 Thread Pradeep Chhetri
Hello Jean,

I am running cassandra 3.11.1.

Since i dont have much cassandra operations experience yet, I have a
follow-up question - how can i ensure the same token ranges distribution ?
Do i need to set initial_token configuration for each cassandra node ?

Thank you for the quick response.





On Thu, Jan 11, 2018 at 3:04 PM, Jean Carlo 
wrote:

> Hello Pradeep,
>
> Actually the key here is to know if your cluster has the same token ranges
> distribution. So it is not only the same size but also the same tokens
> match node by node, from cluster source to cluster destination. In that
> case, you can use nodetool refresh.So after copy all your sstable node by
> node, it would be enough to make nodetool refresh in every node to restore
> your data. You can also restart casandra instead of doing nodetool refresh.
> It will help you to avoid the compactions after refreshing.
>
>
> Saludos
>
> Jean Carlo
>
> "The best way to predict the future is to invent it" Alan Kay
>
> On Thu, Jan 11, 2018 at 9:58 AM, Pradeep Chhetri 
> wrote:
>
>> Hello everyone,
>>
>> We are running cassandra cluster inside containers over Kubernetes. We
>> have a requirement where we need to restore a fresh new cluster with
>> existing snapshot on weekly basis.
>>
>> Currently, while doing it manually. i need to copy the snapshot folder
>> inside container and then run sstableloader utility to load those tables.
>>
>> Since the source and destination cluster size is equal, I was thinking if
>> there are some easy way to just copy and paste the complete data directory
>> by mapping the nodes one to one.
>>
>> Since i wasn't able to find documentation around other  backup
>> restoration methods apart from nodetool snapshot and sstableloader, I
>> haven't explored much. I recently came across this project -
>> https://github.com/Netflix/Priam but tried it yet.
>>
>> Would be very happy if i can get some ideas around various ways of
>> backup/restoration while running inside containers.
>>
>> Thank you
>>
>
>


Re: Restoring fresh new cluster from existing snapshot

2018-01-11 Thread Jean Carlo
Hello Pradeep,

Actually the key here is to know if your cluster has the same token ranges
distribution. So it is not only the same size but also the same tokens
match node by node, from cluster source to cluster destination. In that
case, you can use nodetool refresh.So after copy all your sstable node by
node, it would be enough to make nodetool refresh in every node to restore
your data. You can also restart casandra instead of doing nodetool refresh.
It will help you to avoid the compactions after refreshing.


Saludos

Jean Carlo

"The best way to predict the future is to invent it" Alan Kay

On Thu, Jan 11, 2018 at 9:58 AM, Pradeep Chhetri 
wrote:

> Hello everyone,
>
> We are running cassandra cluster inside containers over Kubernetes. We
> have a requirement where we need to restore a fresh new cluster with
> existing snapshot on weekly basis.
>
> Currently, while doing it manually. i need to copy the snapshot folder
> inside container and then run sstableloader utility to load those tables.
>
> Since the source and destination cluster size is equal, I was thinking if
> there are some easy way to just copy and paste the complete data directory
> by mapping the nodes one to one.
>
> Since i wasn't able to find documentation around other  backup restoration
> methods apart from nodetool snapshot and sstableloader, I haven't explored
> much. I recently came across this project - https://github.com/Netflix/
> Priam but tried it yet.
>
> Would be very happy if i can get some ideas around various ways of
> backup/restoration while running inside containers.
>
> Thank you
>