Hi tison, > When a new node join an existing cluster in this case, how it exactly applies logs? Will it applies raft logs from the beginning one by one?
The leader takes a snapshot from time to time and then purges the old log entries. When a new server joins the group, the leader will first send a snapshot, if there is any, and then send the remaining log entries. Tsz-Wo On Sun, Jul 3, 2022 at 9:26 AM tison <[email protected]> wrote: > Hi Tsz-Wo, > > I have one question that may be a bit out of topic: > > When a new node join an existing cluster in this case, how it exactly > applies logs? Will it applies raft logs from the beginning one by one? TiKV > will sends a snapshot of state machine from leader to the new node in this > case to speed up the catch up progress and I'd like to know how Ratis > implements catch-up. > > Best, > tison. > > > Tsz Wo Sze <[email protected]> 于2022年7月2日周六 01:31写道: > >> > * It's possible to update the old configuration first by using >> client.admin().setConfiguration(), let's say set N=11 first, then start new >> nodes. However, since 5 < 11/2, the cluster won't be able to elect leader >> until at least 1 new node join. >> Yes, you are right. Also, even if one node has joined, the group has to >> wait for it to catch up with the previous log entries in order to obtain a >> majority for committing new entries. >> >> > * Or may be we should limit the count when scaling? From N=5 -> N=7 -> >> N=9 -> N=11. >> >> We may start the 6 new nodes as listeners first. Listeners receive log >> entries but they are not voting members and they won't be counted for >> majority. When the listeners catch up, we may change them to normal nodes >> so that they become voting members. >> >> Tsz-Wo >> >> >> On Fri, Jul 1, 2022 at 10:23 AM Tsz Wo Sze <[email protected]> wrote: >> >>> Hi Riguz, >>> >>> > Start 6 new nodes with new configuration N=11, while keeping the >>> previous nodes running >>> >>> This step probably won't work as expected since it will create a new >>> group but not adding nodes to the original group. We must use the >>> setConfiguration API to change configuration (add/remove nodes); see >>> https://github.com/apache/ratis/blob/bd83e7d7fd41540c8bda6bd92a52ac99ccec2076/ratis-client/src/main/java/org/apache/ratis/client/api/AdminApi.java#L35 >>> >>> Hope it helps. Thanks a lot for trying Ratis! >>> >>> Tsz-Wo >>> >>> On Fri, Jul 1, 2022 at 12:30 AM Riguz Lee <[email protected]> wrote: >>> >>>> >>>> >>>> Hi, >>>> >>>> >>>> I'm testing scaling up/down the raft cluster, but ratis is not working >>>> as expected in new cluster. My steps are: >>>> >>>> >>>> * Initialize a cluster with 5 nodes, the size and peers of the cluster >>>> is configured in a configuration file, let's say N=5. The cluster works >>>> perfectly, raft logs are synchronized across the cluster. >>>> >>>> * Start 6 new nodes with new configuration N=11, while keeping the >>>> previous nodes running >>>> >>>> * Recreate the previous nodes with N=11 one by one >>>> >>>> >>>> According the raft paper, raft should be able to handle configuration >>>> change by design, but after the above steps, what I've found is that: >>>> >>>> >>>> - New nodes not able to join the cluster >>>> >>>> - Old nodes still has a size of 5(by >>>> *client.getGroupManagementApi(peerId).info(groupId)*) >>>> >>>> >>>> So how should I scale the cluster correctly? A few thoughts of mine: >>>> >>>> >>>> * Definitely the old cluster should not be stopped while starting new >>>> nodes, otherwise new nodes might be able to elect new leader(eg. N=11 with >>>> 6 new nodes) and raft logs in old nodes will be overriden. >>>> >>>> * It's possible to update the old configuration first by using >>>> client.admin().setConfiguration(), let's say set N=11 first, then start new >>>> nodes. However, since 5 < 11/2, the cluster won't be able to elect leader >>>> until at least 1 new node join. >>>> >>>> * Or may be we should limit the count when scaling? From N=5 -> N=7 -> >>>> N=9 -> N=11. >>>> >>>> >>>> Thanks, >>>> >>>> Riguz Lee >>>> >>>> >>>> >>>>
