Hey Godefroy,

Do you have Riak Search enabled?

Mark

On Wed, Mar 20, 2013 at 11:41 AM, Godefroy de Compreignac
<[email protected]> wrote:
> Any news? I still have the same problem, same data repartition...
>
>
> 2013/3/13 Godefroy de Compreignac <[email protected]>
>>
>> I'm running Riak 1.2.1
>> I installed it with riak_1.2.1-1_amd64.deb
>>
>> Godefroy
>>
>>
>> 2013/3/13 Tom Santero <[email protected]>
>>>
>>> Godefroy,
>>>
>>> Which version of Riak are you running?
>>>
>>> Tom
>>>
>>>
>>> On Wed, Mar 13, 2013 at 5:51 AM, Godefroy de Compreignac
>>> <[email protected]> wrote:
>>>>
>>>> Hi Mark,
>>>>
>>>> Thanks for your email.
>>>> The rebalancing doesn't seem to be working really good...
>>>> I still have approximately the same repartition as last week :
>>>>
>>>> # riak-admin member-status
>>>> Attempting to restart script through sudo -H -u riak
>>>> ================================= Membership
>>>> ==================================
>>>> Status     Ring    Pending    Node
>>>>
>>>> -------------------------------------------------------------------------------
>>>> valid      18.0%     25.0%    '[email protected]'
>>>> valid       0.0%      0.0%    '[email protected]'
>>>> valid      18.8%     25.0%    '[email protected]'
>>>> valid      29.7%     25.0%    '[email protected]'
>>>> valid      33.6%     25.0%    '[email protected]'
>>>>
>>>> -------------------------------------------------------------------------------
>>>> Valid:5 / Leaving:0 / Exiting:0 / Joining:0 / Down:0
>>>>
>>>> (the second node is waiting for the end of the rebalancing to join the
>>>> cluster and to begin a second rebalancing)
>>>> My servers are in the public network of OVH. No Vlan, but good iptables
>>>> rules. 1 Gbps network interface.
>>>>
>>>> If it can be useful:
>>>>
>>>> # riak-admin ring-status
>>>> Attempting to restart script through sudo -H -u riak
>>>> ================================== Claimant
>>>> ===================================
>>>> Claimant:  '[email protected]'
>>>> Status:     up
>>>> Ring Ready: true
>>>>
>>>> ============================== Ownership Handoff
>>>> ==============================
>>>> Owner:      [email protected]
>>>> Next Owner: [email protected]
>>>>
>>>> Index: 662242929415565384811044689824565743281594433536
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 799258707915337533392640142891717276374338109440
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 844930634081928249586505293914101120738586001408
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 890602560248518965780370444936484965102833893376
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>>
>>>> -------------------------------------------------------------------------------
>>>> Owner:      [email protected]
>>>> Next Owner: [email protected]
>>>>
>>>> Index: 1004782375664995756265033322492444576013453623296
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1050454301831586472458898473514828420377701515264
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1096126227998177188652763624537212264741949407232
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1141798154164767904846628775559596109106197299200
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1187470080331358621040493926581979953470445191168
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1233142006497949337234359077604363797834693083136
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1278813932664540053428224228626747642198940975104
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1324485858831130769622089379649131486563188867072
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1415829711164312202009819681693899175291684651008
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>>
>>>> -------------------------------------------------------------------------------
>>>> Owner:      [email protected]
>>>> Next Owner: [email protected]
>>>>
>>>> Index: 924856504873462002925769308203272848376019812352
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 970528431040052719119634459225656692740267704320
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1016200357206643435313499610248040537104515596288
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1061872283373234151507364761270424381468763488256
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1107544209539824867701229912292808225833011380224
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1153216135706415583895095063315192070197259272192
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1198888061873006300088960214337575914561507164160
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1244559988039597016282825365359959758925755056128
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1290231914206187732476690516382343603290002948096
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1335903840372778448670555667404727447654250840064
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1381575766539369164864420818427111292018498732032
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1427247692705959881058285969449495136382746624000
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>>
>>>> -------------------------------------------------------------------------------
>>>> Owner:      [email protected]
>>>> Next Owner: [email protected]
>>>>
>>>> Index: 936274486415109681974235595958868809467081785344
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 981946412581700398168100746981252653831329677312
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1027618338748291114361965898003636498195577569280
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1073290264914881830555831049026020342559825461248
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1118962191081472546749696200048404186924073353216
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1210306043414653979137426502093171875652569137152
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1255977969581244695331291653115555720016817029120
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1301649895747835411525156804137939564381064921088
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1347321821914426127719021955160323408745312813056
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1392993748081016843912887106182707253109560705024
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>> Index: 1438665674247607560106752257205091097473808596992
>>>>   Waiting on: [riak_kv_vnode]
>>>>   Complete:   [riak_pipe_vnode]
>>>>
>>>>
>>>> -------------------------------------------------------------------------------
>>>>
>>>> ============================== Unreachable Nodes
>>>> ==============================
>>>> All nodes are up and reachable
>>>>
>>>>
>>>> Godefroy
>>>>
>>>>
>>>> 2013/3/13 Mark Phillips <[email protected]>
>>>>>
>>>>> Hi Godefroy,
>>>>>
>>>>> Good to hear you managed to get the node running again. How is the
>>>>> rebalancing going? Also, what's the network setup for the cluster?
>>>>>
>>>>> Mark
>>>>>
>>>>> On Fri, Mar 8, 2013 at 4:24 AM, Godefroy de Compreignac
>>>>> <[email protected]> wrote:
>>>>> > Hi Mark,
>>>>> >
>>>>> > Thanks for your answer.
>>>>> > To have the node running back, I moved a 39GB bitcask dir to another
>>>>> > disk
>>>>> > and made a symlink (ln -s). Rebalancing seems to be running, but
>>>>> > inequalities in data repartition stay huge and the node I added
>>>>> > yesterday
>>>>> > still has 0% of cluster data.
>>>>> >
>>>>> > Godefroy
>>>>> >
>>>>> >
>>>>> > 2013/3/7 Mark Phillips <[email protected]>
>>>>> >>
>>>>> >> Salut Godefroy
>>>>> >>
>>>>> >> On Thu, Mar 7, 2013 at 6:50 AM, Godefroy de Compreignac
>>>>> >> <[email protected]> wrote:
>>>>> >> > Hello,
>>>>> >> >
>>>>> >> > I'm running a cluster of 4 nodes (1,8 TB on each) and I have a
>>>>> >> > problem
>>>>> >> > of
>>>>> >> > balancing. Current data repartition is 18%, 19%, 30%, 34%. The
>>>>> >> > node with
>>>>> >> > 34%
>>>>> >> > of cluster data is completely full and doesn't want to start
>>>>> >> > anymoe.
>>>>> >> > I don't know what to do. Do you have a solution for such a
>>>>> >> > problem?
>>>>> >> >
>>>>> >>
>>>>> >>
>>>>> >> It looks like you need to increase storage capacity for the entire
>>>>> >> cluster so you can move some data off of the full node. Do you have
>>>>> >> the ability to another machine (or two) to the cluster?
>>>>> >>
>>>>> >> The issue, of course, is that you'll need to get that Riak node
>>>>> >> running before it can hand off a subset of its data to a new member.
>>>>> >> I
>>>>> >> assume the disk being full is the primary reason its failing to
>>>>> >> start?
>>>>> >>
>>>>> >> Mark
>>>>> >>
>>>>> >> > Thank you in advance!
>>>>> >> >
>>>>> >> > Godefroy
>>>>> >> >
>>>>> >> > _______________________________________________
>>>>> >> > riak-users mailing list
>>>>> >> > [email protected]
>>>>> >> > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>>>>> >> >
>>>>> >
>>>>> >
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> riak-users mailing list
>>>> [email protected]
>>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>>>>
>>>
>>
>

_______________________________________________
riak-users mailing list
[email protected]
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com

Reply via email to