>
> I just asked that question on this list and the answer was that adding the
> new nodes as rack4, rack5 and rack6 is fine. They are all on
> separate physical racks. Is that ok?
>

Yes, Jeff is right, all 6 nodes each on their own rack will work just fine.

Should I do a full repair first or is the remove node operation basically
> doing that?
>

I don't think you'll need a full repair.  Removenode should be taking care
of streaming that node's data to where it needs to go.

On Mon, Apr 3, 2023 at 10:26 AM David Tinker <david.tin...@gmail.com> wrote:

> Thanks. Yes my big screwup here was to make the new node a seed node so it
> didn't get any data. I am going to add 3 more nodes, one at a time when the
> cluster has finished with the remove and everything seems stable. Should I
> do a full repair first or is the remove node operation basically doing that?
>
> Re the racks. I just asked that question on this list and the answer was
> that adding the new nodes as rack4, rack5 and rack6 is fine. They are all
> on separate physical racks. Is that ok?
>
> On Mon, Apr 3, 2023 at 5:16 PM Aaron Ploetz <aaronplo...@gmail.com> wrote:
>
>> The time it takes to stream data off of a node varies by network, cloud
>> region, and other factors.  So it's not unheard of for it to take a bit to
>> finish.
>>
>> Just thought I'd mention that auto_bootstrap is true by default.  So if
>> you're not setting it, the node should bootstrap as long as it's not a seed
>> node.
>>
>> As for the rack issue, yes, it's a good idea to keep your racks in
>> multiples of your RF.  When performing token ownership calculations,
>> Cassandra takes rack designation into consideration.  It tries to ensure
>> that multiple replicas for a row are not placed in the same rack.  TBH -
>> I'd build out two more nodes to have 6 nodes across 3 racks (2 in each),
>> just to ensure even distribution.  Otherwise, you might notice that the
>> nodes sharing a rack will consume disk at a different rate than the nodes
>> which have their own rack.
>>
>> On Mon, Apr 3, 2023 at 8:57 AM David Tinker <david.tin...@gmail.com>
>> wrote:
>>
>>> Thanks. Hmm, the remove has been busy for hours but seems to be
>>> progressing.
>>>
>>> I have been running this on the nodes to monitor progress:
>>> # nodetool netstats | grep Already
>>>         Receiving 92 files, 843934103369 bytes total. Already received
>>> 82 files (89.13%), 590204687299 bytes total (69.93%)
>>>         Sending 84 files, 860198753783 bytes total. Already sent 56
>>> files (66.67%), 307038785732 bytes total (35.69%)
>>>         Sending 78 files, 815573435637 bytes total. Already sent 56
>>> files (71.79%), 313079823738 bytes total (38.39%)
>>>
>>> The percentages are ticking up.
>>>
>>> # nodetool ring | head -20
>>> Datacenter: dc1
>>> ==========
>>> Address               Rack        Status State   Load            Owns
>>>              Token
>>>
>>>              9189523899826545641
>>> xxx.xxx.xxx..24        rack4       Down   Leaving 26.62 GiB       79.95%
>>>              -9194674091837769168
>>> xxx.xxx.xxx.107       rack1       Up     Normal  2.68 TiB        73.25%
>>>              -9168781258594813088
>>> xxx.xxx.xxx.253       rack2       Up     Normal  2.63 TiB        73.92%
>>>              -9163037340977721917
>>> xxx.xxx.xxx.105       rack3       Up     Normal  2.68 TiB        72.88%
>>>              -9148860739730046229
>>>
>>>
>>> On Mon, Apr 3, 2023 at 3:46 PM Bowen Song via user <
>>> user@cassandra.apache.org> wrote:
>>>
>>>> Use nodetool removenode is strongly preferred in most circumstances,
>>>> and only resort to assassinate if you do not care about data
>>>> consistency or you know there won't be any consistency issue (e.g. no new
>>>> writes and did not run nodetool cleanup).
>>>>
>>>> Since the size of data on the new node is small, nodetool removenode
>>>> should finish fairly quickly and bring your cluster back.
>>>>
>>>> Next time when you are doing something like this again, please test it
>>>> out on a non-production environment, make sure everything works as expected
>>>> before moving onto the production.
>>>>
>>>>
>>>> On 03/04/2023 06:28, David Tinker wrote:
>>>>
>>>> Should I use assassinate or removenode? Given that there is some data
>>>> on the node. Or will that be found on the other nodes? Sorry for all the
>>>> questions but I really don't want to mess up.
>>>>
>>>> On Mon, Apr 3, 2023 at 7:21 AM Carlos Diaz <crdiaz...@gmail.com> wrote:
>>>>
>>>>> That's what nodetool assassinte will do.
>>>>>
>>>>> On Sun, Apr 2, 2023 at 10:19 PM David Tinker <david.tin...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Is it possible for me to remove the node from the cluster i.e. to
>>>>>> undo this mess and get the cluster operating again?
>>>>>>
>>>>>> On Mon, Apr 3, 2023 at 7:13 AM Carlos Diaz <crdiaz...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> You can leave it in the seed list of the other nodes, just make sure
>>>>>>> it's not included in this node's seed list.  However, if you do decide 
>>>>>>> to
>>>>>>> fix the issue with the racks first assassinate this node (nodetool
>>>>>>> assassinate <ip>), and update the rack name before you restart.
>>>>>>>
>>>>>>> On Sun, Apr 2, 2023 at 10:06 PM David Tinker <david.tin...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> It is also in the seeds list for the other nodes. Should I remove
>>>>>>>> it from those, restart them one at a time, then restart it?
>>>>>>>>
>>>>>>>> /etc/cassandra # grep -i bootstrap *
>>>>>>>> doesn't show anything so I don't think I have auto_bootstrap false.
>>>>>>>>
>>>>>>>> Thanks very much for the help.
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Apr 3, 2023 at 7:01 AM Carlos Diaz <crdiaz...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Just remove it from the seed list in the cassandra.yaml file and
>>>>>>>>> restart the node.  Make sure that auto_bootstrap is set to true first
>>>>>>>>> though.
>>>>>>>>>
>>>>>>>>> On Sun, Apr 2, 2023 at 9:59 PM David Tinker <
>>>>>>>>> david.tin...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> So likely because I made it a seed node when I added it to the
>>>>>>>>>> cluster it didn't do the bootstrap process. How can I recover this?
>>>>>>>>>>
>>>>>>>>>> On Mon, Apr 3, 2023 at 6:41 AM David Tinker <
>>>>>>>>>> david.tin...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Yes replication factor is 3.
>>>>>>>>>>>
>>>>>>>>>>> I ran nodetool repair -pr on all the nodes (one at a time) and
>>>>>>>>>>> am still having issues getting data back from queries.
>>>>>>>>>>>
>>>>>>>>>>> I did make the new node a seed node.
>>>>>>>>>>>
>>>>>>>>>>> Re "rack4": I assumed that was just an indication as to the
>>>>>>>>>>> physical location of the server for redundancy. This one is 
>>>>>>>>>>> separate from
>>>>>>>>>>> the others so I used rack4.
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Apr 3, 2023 at 6:30 AM Carlos Diaz <crdiaz...@gmail.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> I'm assuming that your replication factor is 3.  If that's the
>>>>>>>>>>>> case, did you intentionally put this node in rack 4?  Typically, 
>>>>>>>>>>>> you want
>>>>>>>>>>>> to add nodes in multiples of your replication factor in order to 
>>>>>>>>>>>> keep the
>>>>>>>>>>>> "racks" balanced.  In other words, this node should have been 
>>>>>>>>>>>> added to rack
>>>>>>>>>>>> 1, 2 or 3.
>>>>>>>>>>>>
>>>>>>>>>>>> Having said that, you should be able to easily fix your problem
>>>>>>>>>>>> by running a nodetool repair -pr on the new node.
>>>>>>>>>>>>
>>>>>>>>>>>> On Sun, Apr 2, 2023 at 8:16 PM David Tinker <
>>>>>>>>>>>> david.tin...@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi All
>>>>>>>>>>>>>
>>>>>>>>>>>>> I recently added a node to my 3 node Cassandra 4.0.5 cluster
>>>>>>>>>>>>> and now many reads are not returning rows! What do I need to do 
>>>>>>>>>>>>> to fix
>>>>>>>>>>>>> this? There weren't any errors in the logs or other problems that 
>>>>>>>>>>>>> I could
>>>>>>>>>>>>> see. I expected the cluster to balance itself but this hasn't 
>>>>>>>>>>>>> happened
>>>>>>>>>>>>> (yet?). The nodes are similar so I have num_tokens=256 for each. 
>>>>>>>>>>>>> I am using
>>>>>>>>>>>>> the Murmur3Partitioner.
>>>>>>>>>>>>>
>>>>>>>>>>>>> # nodetool status
>>>>>>>>>>>>> Datacenter: dc1
>>>>>>>>>>>>> ===============
>>>>>>>>>>>>> Status=Up/Down
>>>>>>>>>>>>> |/ State=Normal/Leaving/Joining/Moving
>>>>>>>>>>>>> --  Address          Load       Tokens  Owns (effective)  Host
>>>>>>>>>>>>> ID                               Rack
>>>>>>>>>>>>> UN  xxx.xxx.xxx.105  2.65 TiB   256     72.9%
>>>>>>>>>>>>> afd02287-3f88-4c6f-8b27-06f7a8192402  rack3
>>>>>>>>>>>>> UN  xxx.xxx.xxx.253  2.6 TiB    256     73.9%
>>>>>>>>>>>>> e1af72be-e5df-4c6b-a124-c7bc48c6602a  rack2
>>>>>>>>>>>>> UN  xxx.xxx.xxx.24   93.82 KiB  256     80.0%
>>>>>>>>>>>>> c4e8b4a0-f014-45e6-afb4-648aad4f8500  rack4
>>>>>>>>>>>>> UN  xxx.xxx.xxx.107  2.65 TiB   256     73.2%
>>>>>>>>>>>>> ab72f017-be96-41d2-9bef-a551dec2c7b5  rack1
>>>>>>>>>>>>>
>>>>>>>>>>>>> # nodetool netstats
>>>>>>>>>>>>> Mode: NORMAL
>>>>>>>>>>>>> Not sending any streams.
>>>>>>>>>>>>> Read Repair Statistics:
>>>>>>>>>>>>> Attempted: 0
>>>>>>>>>>>>> Mismatch (Blocking): 0
>>>>>>>>>>>>> Mismatch (Background): 0
>>>>>>>>>>>>> Pool Name                    Active   Pending      Completed
>>>>>>>>>>>>> Dropped
>>>>>>>>>>>>> Large messages                  n/a         0          71754
>>>>>>>>>>>>>       0
>>>>>>>>>>>>> Small messages                  n/a         0        8398184
>>>>>>>>>>>>>      14
>>>>>>>>>>>>> Gossip messages                 n/a         0        1303634
>>>>>>>>>>>>>       0
>>>>>>>>>>>>>
>>>>>>>>>>>>> # nodetool ring
>>>>>>>>>>>>> Datacenter: dc1
>>>>>>>>>>>>> ==========
>>>>>>>>>>>>> Address               Rack        Status State   Load
>>>>>>>>>>>>>    Owns                Token
>>>>>>>>>>>>>
>>>>>>>>>>>>>                        9189523899826545641
>>>>>>>>>>>>> xxx.xxx.xxx.24        rack4       Up     Normal  93.82 KiB
>>>>>>>>>>>>>   79.95%              -9194674091837769168
>>>>>>>>>>>>> xxx.xxx.xxx.107       rack1       Up     Normal  2.65 TiB
>>>>>>>>>>>>>    73.25%              -9168781258594813088
>>>>>>>>>>>>> xxx.xxx.xxx.253       rack2       Up     Normal  2.6 TiB
>>>>>>>>>>>>>   73.92%              -9163037340977721917
>>>>>>>>>>>>> xxx.xxx.xxx.105       rack3       Up     Normal  2.65 TiB
>>>>>>>>>>>>>    72.88%              -9148860739730046229
>>>>>>>>>>>>> xxx.xxx.xxx.107       rack1       Up     Normal  2.65 TiB
>>>>>>>>>>>>>    73.25%              -9125240034139323535
>>>>>>>>>>>>> xxx.xxx.xxx.253       rack2       Up     Normal  2.6 TiB
>>>>>>>>>>>>>   73.92%              -9112518853051755414
>>>>>>>>>>>>> xxx.xxx.xxx.105       rack3       Up     Normal  2.65 TiB
>>>>>>>>>>>>>    72.88%              -9100516173422432134
>>>>>>>>>>>>> ...
>>>>>>>>>>>>>
>>>>>>>>>>>>> This is causing a serious production issue. Please help if you
>>>>>>>>>>>>> can.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks
>>>>>>>>>>>>> David
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>

Reply via email to