[ 
https://issues.apache.org/jira/browse/CASSANDRA-5780?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14933745#comment-14933745
 ] 

John Sumsion commented on CASSANDRA-5780:
-----------------------------------------

The only thing I wouldn't want to have happen is to accidentally issue some 
kind of truncate that in a race condition inadvertently gets replicated to the 
entire cluster.  I don't know the cassandra codebase enough to understand 
whether that risk exists when calling {{ColumnFamilyStore.truncateBlocking()}}. 
 From what I can tell, I think it's likely pretty safe because once you get 
down to StorageService, there is no cross-cluster effect of actions taken at 
that level.

Can anyone reply who knows better what cross-cluster effects 
{{truncateBlocking()}} might have?

The reason I don't have that concern with the 'system' keyspace is that it is 
never replicated.

Actually, looking into  {{ColumnFamilyStore.truncateBlocking()}} makes me think 
that my proposed changes will blow up half-way through because a side-effect of 
truncating a table is writing back a "truncated at" record to 'system.local' 
table (which we just truncated).  I guess I need to run ccm with a local-built 
cassandra and try decomissioning to see what happens (not sure how to do that).

> nodetool status and ring report incorrect/stale information after decommission
> ------------------------------------------------------------------------------
>
>                 Key: CASSANDRA-5780
>                 URL: https://issues.apache.org/jira/browse/CASSANDRA-5780
>             Project: Cassandra
>          Issue Type: Bug
>          Components: Tools
>            Reporter: Peter Haggerty
>            Priority: Trivial
>              Labels: lhf, ponies, qa-resolved
>             Fix For: 2.1.x
>
>
> Cassandra 1.2.6 ring of 12 instances, each with 256 tokens.
> Decommission 3 of the 12 nodes, one after another resulting a 9 instance ring.
> The 9 instances of cassandra that are in the ring all correctly report 
> nodetool status information for the ring and have the same data.
> After the first node is decommissioned:
> "nodetool status" on "decommissioned-1st" reports 11 nodes
> After the second node is decommissioned:
> "nodetool status" on "decommissioned-1st" reports 11 nodes
> "nodetool status" on "decommissioned-2nd" reports 10 nodes
> After the second node is decommissioned:
> "nodetool status" on "decommissioned-1st" reports 11 nodes
> "nodetool status" on "decommissioned-2nd" reports 10 nodes
> "nodetool status" on "decommissioned-3rd" reports 9 nodes
> The storage load information is similarly stale on the various decommissioned 
> nodes. The nodetool status and ring commands continue to return information 
> as if they were part of a cluster and they appear to return the last 
> information that they saw.
> In contrast the nodetool info command fails with an exception, which isn't 
> ideal but at least indicates that there was a failure rather than returning 
> stale information.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to