Il mer 21 ago 2019, 17:22 Alexander Shraer <shra...@gmail.com> ha scritto:

> That's great! Thanks for sharing.
>
> > Added benefit is that we can also control which data center gets the
> quorum
> > in case of a network outage between the two.
>
> Can you explain how this works? In case of a network outage between two
> DCs, one of them has a quorum of participants and the other doesn't.
> The participants in the smaller set should not be operational at this time,
> since they can't get quorum. no ?
>

I have recently talked about a similar problem with Norbert K. on gitter
chat.
We got to the conclusion that you need 3 datacenters

Enrico



> Thanks,
> Alex
>
> On Wed, Aug 21, 2019 at 7:55 AM Cee Tee <c.turks...@gmail.com> wrote:
>
> > We have solved this by implementing a 'zookeeper cluster balancer', it
> > calls the admin server api of each zookeeper to get the current status
> and
> > will issue dynamic reconfigure commands to change dead servers into
> > observers so the quorum is not in danger. Once the dead servers
> reconnect,
> > they take the observer role and are then reconfigured into participants
> > again.
> >
> > Added benefit is that we can also control which data center gets the
> > quorum
> > in case of a network outage between the two.
> > Regards
> > Chris
> >
> > On 21 August 2019 16:42:37 Alexander Shraer <shra...@gmail.com> wrote:
> >
> > > Hi,
> > >
> > > Reconfiguration, as implemented, is not automatic. In your case, when
> > > failures happen, this doesn't change the ensemble membership.
> > > When 2 of 5 fail, this is still a minority, so everything should work
> > > normally, you just won't be able to handle an additional failure. If
> > you'd
> > > like
> > > to remove them from the ensemble, you need to issue an explicit
> > > reconfiguration command to do so.
> > >
> > > Please see details in the manual:
> > > https://zookeeper.apache.org/doc/r3.5.5/zookeeperReconfig.html
> > >
> > > Alex
> > >
> > > On Wed, Aug 21, 2019 at 7:29 AM Gao,Wei <wei....@arcserve.com> wrote:
> > >
> > >> Hi
> > >>    I encounter a problem which blocks my development of load balance
> > using
> > >> ZooKeeper 3.5.5.
> > >>    Actually, I have a ZooKeeper cluster which comprises of five zk
> > >> servers. And the dynamic configuration file is as follows:
> > >>
> > >>   server.1=zk1:2888:3888:participant;0.0.0.0:2181
> > >>   server.2=zk2:2888:3888:participant;0.0.0.0:2181
> > >>   server.3=zk3:2888:3888:participant;0.0.0.0:2181
> > >>   server.4=zk4:2888:3888:participant;0.0.0.0:2181
> > >>   server.5=zk5:2888:3888:participant;0.0.0.0:2181
> > >>
> > >>   The zk cluster can work fine if every member works normally.
> However,
> > if
> > >> say two of them are suddenly down without previously being notified,
> > >> the dynamic configuration file shown above will not be synchronized
> > >> dynamically, which leads to the zk cluster fail to work normally.
> > >>   I think this is a very common case which may happen at any time. If
> > so,
> > >> how can we resolve it?
> > >>   Really look forward to hearing from you!
> > >> Thanks
> > >>
> >
> >
> >
> >
>

Reply via email to