So we have two data centers already running.. AP-SYDNEY, and US-EAST.. I'm using Ec2Snitch over a site-to-site tunnel.. I'm wanting to move the current US-EAST from AZ 1a to 1e.. I know all docs say use ec2multiregion for multi-DC.
I like the GPFS idea. would that work with the multi-DC too? What's the downside? status would report rack of 1a, even though in 1e? Thanks in advance for the help/thoughts!! On Thu, Jun 28, 2018 at 6:20 PM, kurt greaves <k...@instaclustr.com> wrote: > There is a need for a repair with both DCs as rebuild will not stream all > replicas, so unless you can guarantee you were perfectly consistent at time > of rebuild you'll want to do a repair after rebuild. > > On another note you could just replace the nodes but use GPFS instead of > EC2 snitch, using the same rack name. > > On Fri., 29 Jun. 2018, 00:19 Rahul Singh, <rahul.xavier.si...@gmail.com> > wrote: > >> Parallel load is the best approach and then switch your Data access code >> to only access the new hardware. After you verify that there are no local >> read / writes on the OLD dc and that the updates are only via Gossip, then >> go ahead and change the replication factor on the key space to have zero >> replicas in the old DC. Then you can decommissioned. >> >> This way you are hundred percent sure that you aren’t missing any new >> data. No need for a DC to DC repair but a repair is always healthy. >> >> Rahul >> On Jun 28, 2018, 9:15 AM -0500, Randy Lynn <rl...@getavail.com>, wrote: >> >> Already running with Ec2. >> >> My original thought was a new DC parallel to the current, and then >> decommission the other DC. >> >> Also my data load is small right now.. I know small is relative term.. >> each node is carrying about 6GB.. >> >> So given the data size, would you go with parallel DC or let the new AZ >> carry a heavy load until the others are migrated over? >> and then I think "repair" to cleanup the replications? >> >> >> On Thu, Jun 28, 2018 at 10:09 AM, Rahul Singh < >> rahul.xavier.si...@gmail.com> wrote: >> >>> You don’t have to use EC2 snitch on AWS but if you have already started >>> with it , it may put a node in a different DC. >>> >>> If your data density won’t be ridiculous You could add 3 to different >>> DC/ Region and then sync up. After the new DC is operational you can remove >>> one at a time on the old DC and at the same time add to the new one. >>> >>> Rahul >>> On Jun 28, 2018, 9:03 AM -0500, Randy Lynn <rl...@getavail.com>, wrote: >>> >>> I have a 6-node cluster I'm migrating to the new i3 types. >>> But at the same time I want to migrate to a different AZ. >>> >>> What happens if I do the "running node replace method" with 1 node at a >>> time moving to the new AZ. Meaning, I'll have temporarily; >>> >>> 5 nodes in AZ 1c >>> 1 new node in AZ 1e. >>> >>> I'll wash-rinse-repeat till all 6 are on the new machine type and in the >>> new AZ. >>> >>> Any thoughts about whether this gets weird with the Ec2Snitch and a RF 3? >>> >>> -- >>> Randy Lynn >>> rl...@getavail.com >>> >>> office: >>> 859.963.1616 <+1-859-963-1616> ext 202 >>> 163 East Main Street - Lexington, KY 40507 - USA >>> <https://maps.google.com/?q=163+East+Main+Street+-+Lexington,+KY+40507+-+USA&entry=gmail&source=g> >>> >>> <https://www.getavail.com/> getavail.com <https://www.getavail.com/> >>> >>> >> >> >> -- >> Randy Lynn >> rl...@getavail.com >> >> office: >> 859.963.1616 <+1-859-963-1616> ext 202 >> 163 East Main Street - Lexington, KY 40507 - USA >> <https://maps.google.com/?q=163+East+Main+Street+-+Lexington,+KY+40507+-+USA&entry=gmail&source=g> >> >> <https://www.getavail.com/> getavail.com <https://www.getavail.com/> >> >> -- Randy Lynn rl...@getavail.com office: 859.963.1616 <+1-859-963-1616> ext 202 163 East Main Street - Lexington, KY 40507 - USA <https://www.getavail.com/> getavail.com <https://www.getavail.com/>