Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Alexander Dejanovski
Atul, our fork has been tested on 2.1 and 3.0.x clusters. I've just tested with a CCM 3.6 cluster and it worked with no issue. With Reaper, if you set incremental to false, it'll perform a full subrange repair with no anticompaction. You'll see this message in the logs : INFO

Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Atul Saroha
Hi Alexander, There is compatibility issue raised with spotify/cassandra-reaper for cassandra version 3.x. Is it comaptible with 3.6 in fork thelastpickle/cassandra-reaper ? There are some suggestions mentioned by *brstgt* which we can try on our side. On Thu, Sep 29, 2016 at 5:42 PM, Atul

Re: Optimising the data model for reads

2016-09-29 Thread Romain Hardouin
Hi Julian, The problem with any deletes here is that you can *read* potentially many tombstones. I mean you have two concerns: 1. Avoid to read tombstones during a query 2. How to evict tombstones as quickly as possible to reclaim disk space     The first point is a data model consideration.

Re: TRUNCATE throws OperationTimedOut randomly

2016-09-29 Thread Romain Hardouin
Hi, @Edward > In older versions you can not control when this call will timeout,truncate_request_timeout_in_ms is available for many years, starting from 1.2. Maybe you have another setting parameter in mind? @GeorgeTry to put cassandra logs in debug Best, Romain Le Mercredi 28 septembre

Re: High load on few nodes in a DC.

2016-09-29 Thread Pranay akula
Yes we are using token aware but not shuffling replicas. On Wed, Sep 21, 2016 at 10:04 AM, Romain Hardouin wrote: > Hi, > > Do you shuffle the replicas with TokenAwarePolicy? > TokenAwarePolicy(LoadBalancingPolicy childPolicy, boolean > shuffleReplicas) > > Best, > > Romain

when taking backups using snapshot if the sstable gets compacted will nodetool snapshot hung ??

2016-09-29 Thread James Joseph
Hai we are taking backups using nodetool snapshots but i occasionally see that my script pauses while taking a snapshot of a CF, is this because when it is taking snapshot does the sstables got compacted to a different one so it couldn't find that particular sstable on which it is taking

How to find the reason for mutation drops ??

2016-09-29 Thread James Joseph
I am seeing mutation drops on one of my nodes in the cluster, the load is low no Gc pauses no wide partitions either, so can i debug what is the reason for mutation drops ?? i ran nodetool tpstats only one node out of 9 is dropping rest 8 nodes in the cluster are having 0 mutation drops. How

Re: Nodetool repair

2016-09-29 Thread Li, Guangxing
Romain, I was trying what you mentioned as below: a. nodetool stop VALIDATION b. echo run -b org.apache.cassandra.db:type=StorageService forceTerminateAllRepairSessions | java -jar /tmp/jmxterm/jmxterm-1.0-alpha-4-uber.jar -l 127.0.0.1:7199 to stop a seemingly forever-going repair but seeing

[cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Atul Saroha
Hi, We have seen a weird behaviour in cassandra 3.6. Once our node was went down more than 10 hrs. After that, we had ran Nodetool repair multiple times. But tombstone are not getting sync properly over the cluster. On day- today basis, on expiry of every grace period, deleted records start

Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Alexander Dejanovski
Hi Atul, could you be more specific on how you are running repair ? What's the precise command line for that, does it run on several nodes at the same time, etc... What is your gc_grace_seconds ? Do you see errors in your logs that would be linked to repairs (Validation failure or failure to

Optimising the data model for reads

2016-09-29 Thread Thomas Julian
Hello, I have created a column family for User File Management. CREATE TABLE "UserFile" ("USERID" bigint,"FILEID" text,"FILETYPE" int,"FOLDER_UID" text,"FILEPATHINFO" text,"JSONCOLUMN" text,PRIMARY KEY ("USERID","FILEID")); Sample Entry (4*003,

Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Atul Saroha
Hi, We are not sure whether this issue is linked to that node or not. Our application does frequent delete and insert. May be our approach is not correct for nodetool repair. Yes, we generally fire repair on all boxes at same time. Till now, it was manual with default configuration ( command:

Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Alexander Dejanovski
Atul, since you're using 3.6, by default you're running incremental repair, which doesn't like concurrency very much. Validation errors are not occurring on a partition or partition range base, but if you're trying to run both anticompaction and validation compaction on the same SSTable. Like

Re: WARN Writing large partition for materialized views

2016-09-29 Thread Alexander Dejanovski
Hi Robert, Materialized Views are regular C* tables underneath, so based on their PK they can generate big partitions. It is often advised to keep partition size under 100MB because larger partitions are hard to read and compact. They usually put pressure on the heap and lead to long GC pauses +

RE: Repairs at scale in Cassandra 2.1.13

2016-09-29 Thread Anubhav Kale
Thanks ! For subrange repairs I have seen two approaches. For our specific requirement, we want to do repairs on a small set of keyspaces. 1. Thrift describe_local_ring(keyspace), parse and get token ranges for a given node, split token ranges for given keyspace + table using

Re: Way to write to dc1 but keep data only in dc2

2016-09-29 Thread Edward Capriolo
You can do something like this, though your use of terminology like "queue" really do not apply. You can setup your keyspace with replication in only one data center. CREATE KEYSPACE NTSkeyspace WITH REPLICATION = { 'class' : 'NetworkTopologyStrategy', 'dc2' : 3 }; This will make the

Docs Contribution (was: Re: [RELEASE] Apache Cassandra 3.9 released)

2016-09-29 Thread Michael Shuler
On 09/29/2016 04:08 PM, Dorian Hoxha wrote: > So how does documentation work? Example: I'm interested in Change Data > Capture. The documentation is in-tree, under doc/source, so create a patch and upload it to a JIRA, just as any source change. :) The docs on patches do have testing details, so

Re: How to get rid of "Cannot start multiple repair sessions over the same sstables" exception

2016-09-29 Thread Robert Sicoie
Thanks Alexander, After roll restart the blocked repair job stopped and I was able to run repair again. Regards, Robert Robert Sicoie On Wed, Sep 28, 2016 at 6:46 PM, Alexander Dejanovski < a...@thelastpickle.com> wrote: > Robert, > > You can restart them in any order, that doesn't make a

Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Atul Saroha
Thanks Alexander. Will look into all these. On Thu, Sep 29, 2016 at 4:39 PM, Alexander Dejanovski < a...@thelastpickle.com> wrote: > Atul, > > since you're using 3.6, by default you're running incremental repair, > which doesn't like concurrency very much. > Validation errors are not occurring

Re: WARN Writing large partition for materialized views

2016-09-29 Thread Robert Sicoie
Thanks! Robert Sicoie On Thu, Sep 29, 2016 at 12:49 PM, Alexander Dejanovski < a...@thelastpickle.com> wrote: > Hi Robert, > > Materialized Views are regular C* tables underneath, so based on their PK > they can generate big partitions. > It is often advised to keep partition size under 100MB

[RELEASE] Apache Cassandra 3.9 released

2016-09-29 Thread Michael Shuler
The Cassandra team is pleased to announce the release of Apache Cassandra version 3.9. Apache Cassandra is a fully distributed database. It is the right choice when you need scalability and high availability without compromising performance. http://cassandra.apache.org/ Downloads of source and

Way to write to dc1 but keep data only in dc2

2016-09-29 Thread Dorian Hoxha
I have dc1 and dc2. I want to keep a keyspace only on dc2. But I only have my app on dc1. And I want to write to dc1 (lower latency) which will not keep data locally but just push it to dc2. While reading will only work for dc2. Since my app is mostly write, my app ~will be faster while not having

[RELEASE] Apache Cassandra 3.8 released

2016-09-29 Thread Michael Shuler
The Cassandra team is pleased to announce the release of Apache Cassandra version 3.8. Apache Cassandra is a fully distributed database. It is the right choice when you need scalability and high availability without compromising performance. http://cassandra.apache.org/ Downloads of source and

Re: [RELEASE] Apache Cassandra 3.9 released

2016-09-29 Thread Dorian Hoxha
So how does documentation work? Example: I'm interested in Change Data Capture. *I do appreciate the work done. On Thu, Sep 29, 2016 at 11:02 PM, Michael Shuler wrote: > The Cassandra team is pleased to announce the release of Apache > Cassandra version 3.9. > > Apache