Re: Cassandra Repair question

2019-10-18 Thread Krish Donald
Thanks Manish, What is the best and fastest way to repair a table using nodetool repair ? We are using 256 vnodes . On Fri, Oct 18, 2019 at 10:05 PM manish khandelwal < manishkhandelwa...@gmail.com> wrote: > No it will only cover primary ranges of nodes on single rac. Repair with > -pr option

Re: Cassandra Repair question

2019-10-18 Thread manish khandelwal
No it will only cover primary ranges of nodes on single rac. Repair with -pr option is to be run on all nodes in a rolling manner. Regards Manish On 19 Oct 2019 10:03, "Krish Donald" wrote: > Hi Cassandra experts, > > > We are on Cassandra 3.11.1. > > We have to run repairs for a big cluster.

Cassandra Repair question

2019-10-18 Thread Krish Donald
Hi Cassandra experts, We are on Cassandra 3.11.1. We have to run repairs for a big cluster. We have 2 DCs. 3 RACs in each DC. Replication factor is 3 for each datacenter . So if I run repair on all nodes of a single RAC with "pr" option then ideally it will cover all the ranges. Please

GC Tuning https://thelastpickle.com/blog/2018/04/11/gc-tuning.html

2019-10-18 Thread Sergio Bilello
Hello! Is it still better to use ParNew + CMS Is it still better than G1GC these days? Any recommendation for i3.xlarge nodes read-heavy workload? Thanks, Sergio - To unsubscribe, e-mail:

Cassandra Recommended System Settings

2019-10-18 Thread Sergio Bilello
Hello everyone! Do you have any setting that you would change or tweak from the below list? sudo cat /proc/4379/limits Limit Soft Limit Hard Limit Units Max cpu time unlimitedunlimitedseconds Max file size

Re: loosing data during saving data from java

2019-10-18 Thread Jeff Jirsa
There is no buffer in cassandra that is known to (or suspected to) lose acknowledged writes if it's overwhelmed. There may be a client bug where you send so many async writes that they overwhelm a bounded queue, or otherwise get dropped or timeout, but those would be client bugs, and I'm not sure

loosing data during saving data from java

2019-10-18 Thread adrien ruffie
Hello all, I have a table cassandra where I insert quickly several java entity about 15.000 entries by minutes. But at the process ending, I only have for exemple 199.921 entries instead 312.212 If I truncate the table and relaunch the process, several time I get 199.354 or 189.012 entries ...

Re: TWCS and gc_grace_seconds

2019-10-18 Thread Paul Chandler
Hi Adarsh, You will have problems if you manually delete data when using TWCS. To fully understand why, I recommend reading this The Last Pickle post: https://thelastpickle.com/blog/2016/12/08/TWCS-part1.html And this post I wrote that dives deeper into the problems with deletes:

Re: TWCS and gc_grace_seconds

2019-10-18 Thread Adarsh Kumar
Thanks Jeff, I just checked with business and we have differences in having TTL. So it will be manula purging always. We do not want to use LCS due to high IOs. So: 1. As the use case is of time series data model, TWCS will be give some benefit (without TTL) and with frequent deleted data

Re: TWCS and gc_grace_seconds

2019-10-18 Thread Jeff Jirsa
Is everything in the table TTL’d? Do you do explicit deletes before the data is expected to expire ? Generally speaking, gcgs exists to prevent data resurrection. But ttl’d data can’t be resurrected once it expires, so gcgs has no purpose unless you’re deleting it before the ttl expires. If