Benjamin Roth created CASSANDRA-12405: -----------------------------------------
Summary: node health status inconsistent Key: CASSANDRA-12405 URL: https://issues.apache.org/jira/browse/CASSANDRA-12405 Project: Cassandra Issue Type: Bug Environment: Cassandra 3.9, Linux Xenial Reporter: Benjamin Roth ATM we run a 4 node cluster with cassandra 3.9 Due to another issue (hanging repairs) I am forced to restart nodes from time to time. Before I restart a node, all nodes are listed as UP from any other node. When I restart one node in the cluster, the health statuses of other nodes are affected as well. After having restarted node "cas", the "nodetool status" output on all nodes looks like this during startup phase of cas1: https://gist.github.com/brstgt/9be77470814d2fd160617a1c06579804 After cas1 is up again, I restart cas2. During the startup phase of cas2 the status looks like this: https://gist.github.com/brstgt/d27ef540b2389b3a7d2d015ab83af547 The nodetool output goes along with log messages like this: 2016-08-08T07:30:06+00:00 cas1 [GossipTasks: 1] org.apache.cassandra.gms.Gossiper Convicting /10.23.71.3 with status NORMAL - alive false 2016-08-08T07:30:06+00:00 cas1 [GossipTasks: 1] org.apache.cassandra.gms.Gossiper Convicting /10.23.71.2 with status NORMAL - alive false In extreme cases, nodes didn't even come up again after a restart with an error that there are no seed hosts (sorry, don't have the error message in current logs), but the seed host(s) were definitively up and running. A reboot fixed that issue, starting the node again and again did not help. -- This message was sent by Atlassian JIRA (v6.3.4#6332)