Is the OP expecting a perfect 50%/50% split? That, to my experience, is not
going to happen, it is almost always shifted from a fraction of a percent
to a couple percent.

Datacenter: eu-west
===================
Status=Up/Down
|/ State=Normal/Leaving/Joining/Moving
--  Address        Load       Tokens       Owns (effective)  Host ID
                        Rack
UN  XX.XX.XX.XX    22.71 GiB  256          47.6%
57dafdde-2f62-467c-a8ff-c91e712f89c9  1c
UN  XX.XX.XX.XX  17.17 GiB  256          51.3%
d2a65c51-087d-48de-ae1f-a41142eb148d  1b
UN  XX.XX.XX.XX  26.15 GiB  256          52.4%
acf5dd34-5b81-4e5b-b7be-85a7fccd8e1c  1c
UN  XX.XX.XX.XX   16.64 GiB  256          50.2%
6c8842dd-a966-467c-a7bc-bd6269ce3e7e  1a
UN  XX.XX.XX.XX  24.39 GiB  256          49.8%
fd92525d-edf2-4974-8bc5-a350a8831dfa  1a
UN  XX.XX.XX.XX   23.8 GiB   256          48.7%
bdc597c0-718c-4ef6-b3ef-7785110a9923  1b

Though maybe part of what you are experiencing can be cleared up by
repair/compaction/cleanup. Also, what are your outputs when you call out
specific keyspaces? Do the numbers get more even?

Cheers,

On Mon, Jun 12, 2017 at 5:22 AM Akhil Mehra <akhilme...@gmail.com> wrote:

> auto_bootstrap is true by default. Ensure its set to true. On startup look
> at your logs for your auto_bootstrap value.  Look at the node configuration
> line in your log file.
>
> Akhil
>
> On Mon, Jun 12, 2017 at 6:18 PM, Junaid Nasir <jna...@an10.io> wrote:
>
>> No, I didn't set it (left it at default value)
>>
>> On Fri, Jun 9, 2017 at 3:18 AM, ZAIDI, ASAD A <az1...@att.com> wrote:
>>
>>> Did you make sure auto_bootstrap property is indeed set to [true] when
>>> you added the node?
>>>
>>>
>>>
>>> *From:* Junaid Nasir [mailto:jna...@an10.io]
>>> *Sent:* Monday, June 05, 2017 6:29 AM
>>> *To:* Akhil Mehra <akhilme...@gmail.com>
>>> *Cc:* Vladimir Yudovin <vla...@winguzone.com>; user@cassandra.apache.org
>>> *Subject:* Re: Convert single node C* to cluster (rebalancing problem)
>>>
>>>
>>>
>>> not evenly, i have setup a new cluster with subset of data (around 5gb).
>>> using the configuration above I am getting these results
>>>
>>>
>>>
>>> Datacenter: datacenter1
>>>
>>> =======================
>>>
>>> Status=Up/Down
>>>
>>> |/ State=Normal/Leaving/Joining/Moving
>>>
>>> --  Address      Load       Tokens       Owns (effective)  Host ID     Rack
>>>
>>> UN  10.128.2.1   4.86 GiB   256          44.9%             
>>> e4427611-c247-42ee-9404-371e177f5f17  rack1
>>>
>>> UN  10.128.2.10  725.03 MiB  256         55.1%             
>>> 690d5620-99d3-4ae3-aebe-8f33af54a08b  rack1
>>>
>>> is there anything else I can tweak/check to make the distribution even?
>>>
>>>
>>>
>>> On Sat, Jun 3, 2017 at 3:30 AM, Akhil Mehra <akhilme...@gmail.com>
>>> wrote:
>>>
>>> So now the data is evenly balanced in both nodes?
>>>
>>>
>>>
>>> Refer to the following documentation to get a better understanding of
>>> the roc_address and the broadcast_rpc_address
>>> https://www.instaclustr.com/demystifying-cassandras-broadcast_address/
>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__www.instaclustr.com_demystifying-2Dcassandras-2Dbroadcast-5Faddress_&d=DwMFaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=57WqcUduTb1GA2Ij5E1fXgw3Cf21HYBK_4l2HVryPrk&s=MaTA43pugg78xQNfaOQElhyvd8k7CjVqZPr3IWALdWI&e=>.
>>> I am surprised that your node started up with rpc_broadcast_address set as
>>> this is an unsupported property. I am assuming you are using Cassandra
>>> version 3.10.
>>>
>>>
>>>
>>>
>>>
>>> Regards,
>>>
>>> Akhil
>>>
>>>
>>>
>>> On 2/06/2017, at 11:06 PM, Junaid Nasir <jna...@an10.io> wrote:
>>>
>>>
>>>
>>> I am able to get it working. I added a new node with following changes
>>>
>>> #rpc_address:0.0.0.0
>>>
>>> rpc_address: 10.128.1.11
>>>
>>> #rpc_broadcast_address:10.128.1.11
>>>
>>> rpc_address was set to 0.0.0.0, (I ran into a problem previously
>>> regarding remote connection and made these changes
>>> https://stackoverflow.com/questions/12236898/apache-cassandra-remote-access
>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__stackoverflow.com_questions_12236898_apache-2Dcassandra-2Dremote-2Daccess&d=DwMFaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=57WqcUduTb1GA2Ij5E1fXgw3Cf21HYBK_4l2HVryPrk&s=oj8BCLiyBDqqVQNqfGohFeujtqlzAkd-uwS878d4mg4&e=>
>>> )
>>>
>>>
>>>
>>> should it be happening?
>>>
>>>
>>>
>>> On Thu, Jun 1, 2017 at 6:31 PM, Vladimir Yudovin <vla...@winguzone.com>
>>> wrote:
>>>
>>> Did you run "nodetool cleanup" on first node after second was
>>> bootstrapped? It should clean rows not belonging to node after tokens
>>> changed.
>>>
>>>
>>>
>>> Best regards, Vladimir Yudovin,
>>>
>>> *Winguzone
>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__winguzone.com_-3Ffrom-3Dlist&d=DwMFaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=57WqcUduTb1GA2Ij5E1fXgw3Cf21HYBK_4l2HVryPrk&s=Q1M5YRAsw0iUQKOIulEmO72RhdENQCRhpqZSjgxxHos&e=>
>>> - Cloud Cassandra Hosting*
>>>
>>>
>>>
>>>
>>>
>>> ---- On Wed, 31 May 2017 03:55:54 -0400 *Junaid Nasir <jna...@an10.io
>>> <jna...@an10.io>>* wrote ----
>>>
>>>
>>>
>>> Cassandra ensure that adding or removing nodes are very easy and that
>>> load is balanced between nodes when a change is made. but it's not working
>>> in my case.
>>>
>>> I have a single node C* deployment (with 270 GB of data) and want to
>>> load balance the data on multiple nodes, I followed this guide
>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__docs.datastax.com_en_cassandra_2.1_cassandra_operations_ops-5Fadd-5Fnode-5Fto-5Fcluster-5Ft.html&d=DwMFaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=57WqcUduTb1GA2Ij5E1fXgw3Cf21HYBK_4l2HVryPrk&s=xnLuX4qqSZxVuY6Gz2NXRdc6TG8J7at9kDhkhBnWpnM&e=>
>>>
>>>
>>> `nodetool status` shows 2 nodes but load is not balanced between them
>>>
>>> Datacenter: dc1
>>>
>>> ===============
>>>
>>> Status=Up/Down
>>>
>>> |/ State=Normal/Leaving/Joining/Moving
>>>
>>> --  Address      Load       Tokens       Owns (effective)  Host ID    Rack
>>>
>>> UN  10.128.0.7   270.75 GiB  256          48.6%        
>>> 1a3f6faa-4376-45a8-9c20-11480ae5664c  rack1
>>>
>>> UN  10.128.0.14  414.36 KiB  256          51.4%        
>>> 66a89fbf-08ba-4b5d-9f10-55d52a199b41  rack1
>>>
>>> I also ran 'nodetool repair' on new node but result is same. any
>>> pointers would be appreciated :)
>>>
>>>
>>>
>>> conf file of new node
>>>
>>> cluster_name: 'cluster1'
>>>
>>>  - seeds: "10.128.0.7"
>>> num_tokens: 256
>>>
>>> endpoint_snitch: GossipingPropertyFileSnitch
>>>
>>> Thanks,
>>>
>>> Junaid
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>
>>
>

Reply via email to