There is a bug in Jira related to this, it is not a driver issue, is a Cassandra issue. It is solved on 2.0.14 I think. I will post the ticket once I find it.
Regards, Carlos Juzarte Rolo Cassandra Consultant Pythian - Love your data rolo@pythian | Twitter: cjrolo | Linkedin: *linkedin.com/in/carlosjuzarterolo <http://linkedin.com/in/carlosjuzarterolo>* Mobile: +31 6 159 61 814 | Tel: +1 613 565 8696 x1649 www.pythian.com On Mon, Jul 6, 2015 at 10:50 PM, Jeff Williams <je...@wherethebitsroam.com> wrote: > Anton, > > I have also seen this issue with decommissioned nodes remaining in the > system.peers table. > > On the bright side, they can be safely removed from the system.peers table > without issue. You will have to check every node in the cluster since this > is a local setting per node. > > Jeff > > On 6 July 2015 at 22:45, nowarry <nowa...@gmail.com> wrote: > >> Hey guys, >> >> I'm using Ruby driver( http://datastax.github.io/ruby-driver/ ) for >> backup scripts. I tried to discover all peers and got wrong peers that are >> different with nodetool status. >> >> ================= >> Status=Up/Down >> |/ State=Normal/Leaving/Joining/Moving >> -- Address Load Tokens Owns Host ID >> Rack >> UN 10.40.231.53 1.18 TB 256 ? >> b2d877d7-f031-4190-8569-976bb0ce034f RACK01 >> UN 10.40.231.11 1.24 TB 256 ? >> e15cda1c-65cc-40cb-b85c-c4bd665d02d7 RACK01 >> >> cqlsh> use system; >> cqlsh:system> select peer from system.peers; >> >> peer >> -------------- >> 10.40.231.31 >> 10.40.231.53 >> >> (2 rows) >> >> What to do with these old peers, whether they can be removed without >> consequences since they are not in production cluster? And how to keep up >> to date the peers? >> >> -------------------------------------- >> Anton Koshevoy >> >> > -- --