Being one of our largest and unfortunately heaviest multi-tenant clusters,
and our last 2.1 prod cluster, we are encountering not enough replica
errors (need 2, only found 1) after only bringing down 1 node. 90 node
cluster, 30/dc, dcs are in europe, asia, and us. AWS.

Are there bugs for erroneous gossip state in 2.1.9? I know system.peers and
other issues can make gossip state detection a bit iffy, and AWS also
introduces uncertainty.

Java-driver is v3.7. It is primarily one app throwing the errors, but this
is the app without caching but with substantive query volume. It is RF3
also, while many of the other apps are RF5, which may also be contributing.

Reply via email to