Re: Pluggable throttling of read and write queries

2017-02-20 Thread Oleksandr Shulgin
On Sat, Feb 18, 2017 at 3:12 AM, Abhishek Verma wrote: > Cassandra is being used on a large scale at Uber. We usually create > dedicated clusters for each of our internal use cases, however that is > difficult to scale and manage. > > We are investigating the approach of using a

Re: Error when running nodetool cleanup after adding a new node to a cluster

2017-02-09 Thread Oleksandr Shulgin
On Thu, Feb 9, 2017 at 6:13 AM, Srinath Reddy wrote: > Hi, > > Trying to re-balacne a Cassandra cluster after adding a new node and I'm > getting this error when running nodetool cleanup. The Cassandra cluster > is running in a Kubernetes cluster. > > Cassandra version is

Re: Java GC pauses, reality check

2016-11-25 Thread Oleksandr Shulgin
On Nov 25, 2016 23:47, "Graham Sanderson" wrote: If you are seeing 25-30 second GC pauses then (unless you are so badly configured) seeing full GC under CMS (though G1 may have similar problems). With CMS eventual fragmentation causing promotion failure is inevitable (unless

Re: data not replicated on new node

2016-11-23 Thread Oleksandr Shulgin
On Tue, Nov 22, 2016 at 5:23 PM, Bertrand Brelier < bertrand.brel...@gmail.com> wrote: > Hello Shalom. > > No I really went from 3.1.1 to 3.0.9 . > So you've just installed the 3.0.9 version and re-started with it? I wonder if it's really supported? Regards, -- Alex

Re: Java GC pauses, reality check

2016-11-26 Thread Oleksandr Shulgin
Note in fact ironically, running 2.1.x with off heap memtables, we had some of our JVMs running for over a year which made us hit https://issues.apache.org/jira/browse/CASSANDRA-10969 when we restarted some nodes for other reasons. On Nov 26, 2016, at 12:07 AM, Oleksandr Shulgin < oleksandr.shul...@

Re: Java GC pauses, reality check

2016-11-26 Thread Oleksandr Shulgin
had some of our JVMs running for over a year which made us hit https://issues.apache.org/jira/browse/CASSANDRA-10969 when we restarted some nodes for other reasons. On Nov 26, 2016, at 12:07 AM, Oleksandr Shulgin < oleksandr.shul...@zalando.de> wrote: On Nov 25, 2016 23:47, "Graham

Re: [RELEASE] Apache Cassandra 3.0.10 released

2016-11-17 Thread Oleksandr Shulgin
On Wed, Nov 16, 2016 at 9:17 PM, Michael Shuler wrote: > > The Cassandra team is pleased to announce the release of Apache > Cassandra version 3.0.10. > > Apache Cassandra is a fully distributed database. It is the right choice > when you need scalability and high

Re: [RELEASE] Apache Cassandra 3.0.10 released

2016-11-17 Thread Oleksandr Shulgin
On Thu, Nov 17, 2016 at 1:05 PM, Vladimir Yudovin wrote: > Hi, > >Does this mean that offheap_objects is still available or that there is > no longer support for offheap memtables in version 3.0? > > If you set *offheap_buffers* in cassandra.yaml in 3.0.10, you'll get >

Re: failing bootstraps with OOM

2016-11-02 Thread Oleksandr Shulgin
On Wed, Nov 2, 2016 at 3:35 PM, Mike Torra wrote: > > Hi All - > > I am trying to bootstrap a replacement node in a cluster, but it consistently fails to bootstrap because of OOM exceptions. For almost a week I've been going through cycles of bootstrapping, finding errors,

Re: Rebuilding with vnodes

2016-11-03 Thread Oleksandr Shulgin
On Wed, Nov 2, 2016 at 8:59 PM, Anubhav Kale wrote: > Hello, > > > > I am trying to rebuild a new Data Center with 50 Nodes, and expect 1 TB / > node. Nodes are backed by SSDs, and the rebuild is happening from another > DC in same physical region. This is with

Re: failing bootstraps with OOM

2016-11-03 Thread Oleksandr Shulgin
On Thu, Nov 3, 2016 at 2:32 PM, Mike Torra wrote: > Hi Alex - I do monitor sstable counts and pending compactions, but > probably not closely enough. In 3/4 regions the cluster is running in, both > counts are very high - ~30-40k sstables for one particular CF, and on many

Re: Why compacting process uses more data that is expected

2017-01-04 Thread Oleksandr Shulgin
On Jan 4, 2017 17:58, "Jean Carlo" wrote: Hello guys I have a table with 34Gb of data in sstables (including tmp). And I can see cassandra is doing some compactions on it. What surprissed me is that nodetool compactionstats says he is compacting 138.66GB

Re: Bootstrapping a new Node with Consistency=ONE

2017-08-03 Thread Oleksandr Shulgin
On Thu, Aug 3, 2017 at 9:33 AM, Daniel Hölbling-Inzko < daniel.hoelbling-in...@bitmovin.com> wrote: > No I set Auto bootstrap to true and the node was UN in nodetool status but > when doing a select on the node with ONE I got incomplete data. > What I think is happening here is not related to

Re: Bootstrapping a new Node with Consistency=ONE

2017-08-02 Thread Oleksandr Shulgin
On Wed, Aug 2, 2017 at 10:53 AM, Daniel Hölbling-Inzko < daniel.hoelbling-in...@bitmovin.com> wrote: > > Any advice on how to avoid this in the future? Is there a way to start up > a node that does not serve client requests but does replicate data? > Would it not work if you first increase the

Re: Replacing a Seed Node

2017-08-03 Thread Oleksandr Shulgin
On Thu, Aug 3, 2017 at 3:00 PM, Fd Habash wrote: > Hi all … > > I know there is plenty of docs on how to replace a seed node, but some are > steps are contradictory e.g. need to remote the node from seed list for > entire cluster. > > > > My cluster has 6 nodes with 3 seeds

Re: Cannot achieve consistency level LOCAL_ONE

2017-07-07 Thread Oleksandr Shulgin
On Thu, Jul 6, 2017 at 6:58 PM, Charulata Sharma (charshar) < chars...@cisco.com> wrote: > Hi, > > I am facing similar issues with SYSTEM_AUTH keyspace and wanted to know > the implication of disabling the "*cassandra*" superuser. > Unless you have scheduled any tasks that require the user with

Re: recovering from failed repair , cassandra 3.10

2017-05-31 Thread Oleksandr Shulgin
On Wed, May 31, 2017 at 9:11 AM, Micha wrote: > Hi, > > after failed repair on a three node cluster all nodes were down. > To clarify, was it failed repair that brought the nodes down so that you had to start them back? Do you see any error messages or stack trace in

Re: Convert single node C* to cluster (rebalancing problem)

2017-05-31 Thread Oleksandr Shulgin
On Wed, May 31, 2017 at 9:55 AM, Junaid Nasir wrote: > Cassandra ensure that adding or removing nodes are very easy and that load > is balanced between nodes when a change is made. but it's not working in my > case. > I have a single node C* deployment (with 270 GB of data) and

Re: Cannot achieve consistency level LOCAL_ONE

2017-06-14 Thread Oleksandr Shulgin
On Wed, Jun 14, 2017 at 10:46 AM, wxn...@zjqunshuo.com wrote: > Thanks for the reply. > My system_auth settings is as below and what should I do with it? And I'm > interested why the newly added node is responsible for the user > authentication? > > CREATE KEYSPACE

Re: Cannot achieve consistency level LOCAL_ONE

2017-06-14 Thread Oleksandr Shulgin
On Wed, Jun 14, 2017 at 9:11 AM, wxn...@zjqunshuo.com wrote: > Hi, > Cluster set up: > 1 DC with 5 nodes (each node having 700GB data) > 1 kespace with RF of 2 > write CL is LOCAL_ONE > read CL is LOCAL_QUORUM > > One node was down for about 1 hour because of OOM issue.

Re:

2017-10-04 Thread Oleksandr Shulgin
On Sun, Oct 1, 2017 at 6:14 PM, Bill Walters wrote: > Hi All, > > I need some help with deploying a monitoring and alerting system for our > new Cassandra 3.0.4 cluster that we are setting up in AWS East region. > I have a good experience with Cassandra as we are running

Re: Alter table gc_grace_seconds

2017-10-04 Thread Oleksandr Shulgin
On Sun, Oct 1, 2017 at 4:14 PM, Gábor Auth wrote: > Hi, > > On Sun, Oct 1, 2017 at 3:44 PM Varun Barala > wrote: > >> This is the property of table and It's not written in sstables. If you >> change gc_grace, It'll get applied for all the data. >>

Re: Elastic IP for Cassandra in AWS

2017-10-17 Thread Oleksandr Shulgin
On Tue, Oct 17, 2017 at 7:40 AM, Jeff Jirsa wrote: > Just to expand on this: > > In an asg environment, you could have an autoscaling event to expand or > shrink the cluster and multiple nodes terminate at the same time. Your > userdata or your AMI would have to know how to

Re: Working With Prepared Statements

2017-08-30 Thread Oleksandr Shulgin
On Tue, Aug 29, 2017 at 12:33 PM, Shalom Sagges wrote: > Insights, anyone? > There were reports of Cassandra failing to start due to trying to load the prepared statements from a cached table. This can only affect you if you have a lot (tens of thousands, IIRC) of

Re: Cassandra All host(s) tried for query failed (no host was tried)

2017-08-30 Thread Oleksandr Shulgin
On Wed, Aug 30, 2017 at 5:10 PM, Ivan Iliev wrote: > Hello everyone, > > We are using Cassandra 3.9 for storing quite a lot of data produced from > our tester machines. > > Occasionally, we are seeing issues with apps not being able to communicate > with Cassandra

Re: system_auth replication factor in Cassandra 2.1

2017-08-30 Thread Oleksandr Shulgin
On Wed, Aug 30, 2017 at 6:20 PM, Chuck Reynolds wrote: > So I tried to run a repair with the following on one of the server. > > nodetool repair system_auth -pr –local > > > > After two hours it hadn’t finished. I had to kill the repair because of > another issue and

Re: system_auth replication factor in Cassandra 2.1

2017-08-30 Thread Oleksandr Shulgin
On Wed, Aug 30, 2017 at 5:50 PM, Chuck Reynolds wrote: > So I’ve read that if your using authentication in Cassandra 2.1 that your > replication factor should match the number of nodes in your datacenter. > > > > *Is that true?* > > > > I have two datacenter cluster, 135

Re: system_auth replication factor in Cassandra 2.1

2017-08-30 Thread Oleksandr Shulgin
On Wed, Aug 30, 2017 at 6:40 PM, Chuck Reynolds wrote: > How many users do you have (or expect to be found in system_auth.users)? > > 5 users. > > What are the current RF for system_auth and consistency level you are > using in cqlsh? > > 135 in one DC and 227 in the

Re: Cassandra snapshot restore with VNODES missing some data

2017-08-31 Thread Oleksandr Shulgin
On Thu, Aug 31, 2017 at 10:14 AM, Lutaya Shafiq Holmes < lutayasha...@gmail.com> wrote: > SOME ONE HELP ME GET STARTED WITH CASSANDRA IN WINDOWS Given your user profile picture I've freaked out for a second thinking it's the 45th president of US is shouting at us to get started with

Re: Could not connect to localhost:9160 when installing Cassandra on AWS

2017-10-11 Thread Oleksandr Shulgin
On Tue, Oct 10, 2017 at 7:43 PM, Lutaya Shafiq Holmes < lutayasha...@gmail.com> wrote: > Hello Cassandra Gurus, > > After I installed Cassandra on AWS- This error comes up when I try to > Start CQLSH > > Could not connect to localhost:9160 > Wait, does it try to talk Thrift or what? I thought

Re: gc causes C* node hang

2017-11-29 Thread Oleksandr Shulgin
On Thu, Nov 30, 2017 at 1:38 AM, Peng Xiao <2535...@qq.com> wrote: > looks we are not able to enable –XX:PrintSafepointStatisticsCount=1 > in cassandra-env.sh > Could anyone please advise? > > ... > Error: Could not find or load main class –XX: > PrintSafepointStatisticsCount=1 > Hm, not sure

Re: about write performance

2017-12-07 Thread Oleksandr Shulgin
On Fri, Dec 8, 2017 at 3:05 AM, Eunsu Kim wrote: > There is a table with a timestamp as a cluster key and sorted by ASC for > the column. > > Is it better to insert by the time order when inserting data into this > table for insertion performance? Or does it matter? >

Re: update a record which does not exists

2017-12-04 Thread Oleksandr Shulgin
On Mon, Dec 4, 2017 at 4:13 AM, Peng Xiao <2535...@qq.com> wrote: > Dear All, > If we update a record which actually does not exist in Cassandra,will it > generate a new record or exit? > > UPDATE columnfamily SET data = 'test data' WHERE key = 'row1'; > as in CQL Update and insert are

Re: New token allocation and adding a new DC

2018-01-17 Thread Oleksandr Shulgin
On Wed, Jan 17, 2018 at 4:21 AM, kurt greaves wrote: > I believe you are able to get away with just altering the keyspace to > include both DC's even before the DC exists, and then adding your nodes to > that new DC using the algorithm. Note you'll probably want to take the

New token allocation and adding a new DC

2018-01-16 Thread Oleksandr Shulgin
Hello, We want to add a new rack to an existing cluster (a new Availability Zone on AWS). Currently we have 12 nodes in 2 racks with ~4 TB data per node. We also want to have bigger number of smaller nodes. In order to minimize the streaming we want to add a new DC which will span 3 racks and

Re: New token allocation and adding a new DC

2018-01-16 Thread Oleksandr Shulgin
On Tue, Jan 16, 2018 at 4:16 PM, Alexander Dejanovski < a...@thelastpickle.com> wrote: > Hi Oleksandr, > > if bootstrap is disabled, it will only skip the streaming phase but will > still go through token allocation and thus should use the new algorithm. > The algorithm won't try to spread data

Re: Decommissioned nodes and FailureDetector

2018-01-19 Thread Oleksandr Shulgin
On Fri, Jan 19, 2018 at 11:17 AM, Nicolas Guyomar wrote: > Hi, > > Not sure if StorageService should be accessed, but you can check node > movement here : > 'org.apache.cassandra.db:type=StorageService/LeavingNodes', >

Decommissioned nodes and FailureDetector

2018-01-19 Thread Oleksandr Shulgin
Hello, Is there a better way to monitor for Cassandra nodes going Down than querying via JMX for a condition like FailureDetector.DownEndpointCount > 0? The problem for us is when any node is decommissioned, it affects the DownEndpointCount for another ~3 days (the famous 72 hours of gossip).

Re: Decommissioned nodes and FailureDetector

2018-01-26 Thread Oleksandr Shulgin
On Fri, Jan 19, 2018 at 6:53 PM, Tom van der Woerdt < tom.vanderwoe...@booking.com> wrote: > > Here's the code I use, hope it helps: > ... > Thanks Tom, that really does the trick! -- Alex

Re: Nodes show different number of tokens than initially

2018-01-26 Thread Oleksandr Shulgin
On Fri, Jan 26, 2018 at 3:08 PM, Kenneth Brotman < kenbrot...@yahoo.com.invalid> wrote: > > > Could it be that after distributing the data, some of the nodes did not > need to have a fourth token? > I'm not sure, but that would be definitely against my understanding of how token assignment

Nodes show different number of tokens than initially

2018-01-25 Thread Oleksandr Shulgin
Hello, While testing token allocation with version 3.0.15 we are experiencing some quite unexpected result. We have deployed a secondary virtual DC with 6 nodes, 4 tokens per node. Then we were adding the 7th node to the new DC in order to observe the effect of ownership re-distribution. To set

Re: New token allocation and adding a new DC

2018-01-24 Thread Oleksandr Shulgin
On Thu, Jan 18, 2018 at 5:19 AM, kurt greaves wrote: > Didn't know that about auto_bootstrap and the algorithm. We should > probably fix that. Can you create a JIRA for that issue? > Will do. > Workaround for #2 would be to truncate system.available_ranges after >

Re: Cassandra 2.1: replace running node without streaming

2018-02-02 Thread Oleksandr Shulgin
On Fri, Feb 2, 2018 at 3:15 PM, Kyrylo Lebediev wrote: > Hello All! > > I've got a pretty standard task - to replace a running C* node [version > 2.1.15, vnodes=256, Ec2Snitch] (IP address will change after replacement, > have no control over it). > > There are 2 ways

Re: Upgrading sstables not using all available compaction slots on version 2.2

2018-02-01 Thread Oleksandr Shulgin
On Thu, Feb 1, 2018 at 9:23 AM, Oleksandr Shulgin < oleksandr.shul...@zalando.de> wrote: > On 1 Feb 2018 06:51, "kurt greaves" <k...@instaclustr.com> wrote: > > Would you be able to create a JIRA ticket for this? Not sure if this is > still a problem i

Re: Cassandra 2.1: replace running node without streaming

2018-02-02 Thread Oleksandr Shulgin
On 3 Feb 2018 02:42, "Kyrylo Lebediev" wrote: Thanks, Oleksandr, In my case I'll need to replace all nodes in the cluster (one-by-one), so streaming will introduce perceptible overhead. My question is not about data movement/copy itself, but more about all this token

Re: Cassandra 2.1: replace running node without streaming

2018-02-03 Thread Oleksandr Shulgin
tion if you want to be sure, though. I don't think this is documented anywhere, however, since I had the same doubts before seeing it worked for the first time. -- Alex Am 03.02.2018 um 08:19 schrieb Oleksandr Shulgin < oleksandr.shul...@zalando.de>: On 3 Feb 2018 02:42, "Kyr

Re: Nodes show different number of tokens than initially

2018-02-01 Thread Oleksandr Shulgin
On Thu, Feb 1, 2018 at 5:19 AM, Jeff Jirsa wrote: > >> The reason I find it surprising, is that it makes very little *sense* to >> put a token belonging to a mode from one DC between tokens of nodes from >> another one. >> > > I don't want to really turn this into an argument

Re: Upgrading sstables not using all available compaction slots on version 2.2

2018-01-31 Thread Oleksandr Shulgin
On Wed, Jan 24, 2018 at 10:40 AM, Oleksandr Shulgin < oleksandr.shul...@zalando.de> wrote: > Hello, > > In the process of upgrading our cluster from 2.1 to 2.2 we have triggered > the SSTable rewriting process like this: > > $ nodetool upgradesstables -j 4 #

Re: Nodes show different number of tokens than initially

2018-01-31 Thread Oleksandr Shulgin
On 31 Jan 2018 17:18, "Jeff Jirsa" wrote: I don’t know why this is a surprise (maybe because people like to talk about multiple rings, but the fact that replication strategy is set per keyspace and that you could use SimpleStrategy in a multiple dc cluster demonstrates this),

Re: Nodes show different number of tokens than initially

2018-02-01 Thread Oleksandr Shulgin
On Fri, Feb 2, 2018 at 2:37 AM, kurt greaves wrote: > So one time I tried to understand why only a single node could have a > token, and it appeared that it came over the fence from facebook and has > been kept ever since. Personally I don't think it's necessary, and agree

Re: Cassandra 2.1: replace running node without streaming

2018-02-05 Thread Oleksandr Shulgin
On Sat, Feb 3, 2018 at 11:23 AM, Kyrylo Lebediev wrote: > Just tested on 3.11.1 and it worked for me (you may see the logs below). > > Just comprehended that there is one important prerequisite this method to > work: new node MUST be located in the same rack (in terms

Re: Nodes show different number of tokens than initially

2018-01-31 Thread Oleksandr Shulgin
On Tue, Jan 30, 2018 at 5:44 PM, Jeff Jirsa wrote: > All DCs in a cluster use the same token space in the DHT, > I can't believe my bloody eyes, but this seems to be true... so token conflicts across datacenters are invalid config > If this is deemed invalid config why does

Re: Nodes show different number of tokens than initially

2018-01-31 Thread Oleksandr Shulgin
On Wed, Jan 31, 2018 at 5:06 AM, Dikang Gu wrote: > What's the partitioner you use? We have logic to prevent duplicate tokens. > We are using the default Murmur3Partitioner. The problem arises from the fact that we manually allocating the tokens as described earlier. --

Re: Upgrading sstables not using all available compaction slots on version 2.2

2018-02-01 Thread Oleksandr Shulgin
On 1 Feb 2018 06:51, "kurt greaves" wrote: Would you be able to create a JIRA ticket for this? Not sure if this is still a problem in 3.0+ but worth creating a ticket to investigate. It'd be really helpful if you could try and reproduce on 3.0.15 or 3.11.1 to see if it's an

Re: Nodes show different number of tokens than initially

2018-01-29 Thread Oleksandr Shulgin
On Tue, Jan 30, 2018 at 5:13 AM, kurt greaves wrote: > Shouldn't happen. Can you send through nodetool ring output from one of > those nodes? Also, did the logs have anything to say about tokens when you > started the 3 seed nodes?​ > Hi Kurt, I cannot run nodetool ring

Re: storing indexes on ssd

2018-02-13 Thread Oleksandr Shulgin
On Tue, Feb 13, 2018 at 1:30 AM, Dan Kinder wrote: > Created https://issues.apache.org/jira/browse/CASSANDRA-14229 > This is confusing. You've already started the conversation here... How big are your index files in the end? Even if Cassandra doesn't cache them in or

Re: storing indexes on ssd

2018-02-13 Thread Oleksandr Shulgin
On Tue, Feb 13, 2018 at 10:46 PM, Dan Kinder wrote: > On a single node that's a bit less than half full, the index files are 87G. > That's not small, true. Out of curiosity: how much data per node do you have in total? How will OS disk cache know to keep the index file

Re: What kind of Automation you have for Cassandra related operations on AWS ?

2018-02-09 Thread Oleksandr Shulgin
On Fri, Feb 9, 2018 at 12:46 AM, Krish Donald wrote: > Hi All, > > What kind of Automation you have for Cassandra related operations on AWS > like restacking, restart of the cluster , changing cassandra.yaml > parameters etc ? > We wrote some scripts customized for

Re: What kind of Automation you have for Cassandra related operations on AWS ?

2018-02-09 Thread Oleksandr Shulgin
On Fri, Feb 9, 2018 at 1:01 PM, vincent gromakowski < vincent.gromakow...@gmail.com> wrote: > Working on a Kubernetes operator for Cassandra (Alpha stage...) > I would love to learn more about your approach. Do you have anything to show already? Design docs / prototype? -- Alex

Re: Cassandra Needs to Grow Up by Version Five!

2018-02-21 Thread Oleksandr Shulgin
On Wed, Feb 21, 2018 at 7:54 PM, Durity, Sean R wrote: > > > However, I think the shots at Cassandra are generally unfair. When I > started working with it, the DataStax documentation was some of the best > documentation I had seen on any project, especially an open

Re: Cassandra Needs to Grow Up by Version Five!

2018-02-22 Thread Oleksandr Shulgin
On Thu, Feb 22, 2018 at 9:50 AM, Eric Plowe wrote: > Cassandra, hard to use? I disagree completely. With that said, there are > definitely deficiencies in certain parts of the documentation, but nothing > that is a show stopper. True, there are no show-stoppers from the

Re: Initializing a multiple node cluster (multiple datacenters)

2018-02-22 Thread Oleksandr Shulgin
On Thu, Feb 22, 2018 at 5:36 PM, Jean Carlo wrote: > Hello > > I would like to clarify this, > > In order to initialize a cassandra multi dc cluster, without data. If I > follow the documentation datastax > >

Re: Initializing a multiple node cluster (multiple datacenters)

2018-02-22 Thread Oleksandr Shulgin
On Thu, Feb 22, 2018 at 8:11 PM, Jon Haddad wrote: > Great question. Unfortunately, our OSS docs lack a step by step process > on how to add a DC, I’ve created a JIRA to do that: > https://issues.apache.org/jira/browse/CASSANDRA-14254 > Thanks. I'd love to contribute as

Re: Initializing a multiple node cluster (multiple datacenters)

2018-02-23 Thread Oleksandr Shulgin
On Fri, Feb 23, 2018 at 8:32 AM, Oleksandr Shulgin < oleksandr.shul...@zalando.de> wrote: > > This is important, because if you would like to use it on 3.0, it will not > work unless you make sure that auto_boostrap is *not* set to false. This > is not critical when creatin

Re: Initializing a multiple node cluster (multiple datacenters)

2018-02-22 Thread Oleksandr Shulgin
On Thu, Feb 22, 2018 at 5:42 PM, Jonathan Haddad wrote: > If it's a new cluster, there's no need to disable auto_bootstrap. > True. > That setting prevents the first node in the second DC from being a replica > for all the data in the first DC. > Not sure where did you

Re: Initializing a multiple node cluster (multiple datacenters)

2018-02-23 Thread Oleksandr Shulgin
On Fri, Feb 23, 2018 at 7:02 PM, Jeff Jirsa wrote: > > Yes, seeds don't bootstrap. But why? I don't think I ever seen a >> comprehensive explanation of this. >> >> The meaning of seed in the most common sense is "connect to this host, > and use it as the starting point for

Re: Cassandra Needs to Grow Up by Version Five!

2018-02-21 Thread Oleksandr Shulgin
On Mon, Feb 19, 2018 at 10:01 AM, Kenneth Brotman < kenbrot...@yahoo.com.invalid> wrote: > > >> Cluster wide management should be a big theme in any next major release. > >> > >Na. Stability and testing should be a big theme in the next major release. > > > > Double Na on that one Jeff. I think

Re: Cassandra Hints monitoring

2018-02-26 Thread Oleksandr Shulgin
On Mon, Feb 26, 2018 at 8:21 AM, Jai Bheemsen Rao Dhanwada < jaibheem...@gmail.com> wrote: > > I am trying to evaluate the JMX hints metrics for cassandra(TotalHints). > However, when I am testing the metrics, the total hints accumulated count > shows perfectly fine, but when the hints are

Seed nodes and bootstrap (was: Re: Initializing a multiple node cluster (multiple datacenters))

2018-02-26 Thread Oleksandr Shulgin
On Fri, Feb 23, 2018 at 7:35 PM, Jeff Jirsa wrote: > It comes up from time to time. Rob Coli spent years arguing that this > behavior was confusing ( https://issues.apache.org/ > jira/browse/CASSANDRA-5836 ) , especially in the "I'm replacing a failed > seed" sense. It also

Re: Cassandra Hints monitoring

2018-02-26 Thread Oleksandr Shulgin
On Mon, Feb 26, 2018 at 10:02 AM, Jai Bheemsen Rao Dhanwada < jaibheem...@gmail.com> wrote: > Thank you Alex, > > I tried "TotalHintsInProgress" already, and I don't see it sending the > correct metrics. I used mean("TotalHintsInProgress") and I see 0 always > on grafana. > Do you know what is

Re: Data Deleted After a few days of being off

2018-02-26 Thread Oleksandr Shulgin
On Tue, Feb 27, 2018 at 7:37 AM, A wrote: > > I started going through the logs and haven't noticed anything yet... Very > unexpected behavior. > Maybe I'm asking the obvious, but were your inserts *without* a TTL? -- Alex

Re: Seed nodes and bootstrap (was: Re: Initializing a multiple node cluster (multiple datacenters))

2018-02-26 Thread Oleksandr Shulgin
On Mon, Feb 26, 2018 at 7:05 PM, Jeff Jirsa wrote: > > I'll happily click the re-open button (you could have, too), but I'm not > sure what the 'right' fix is. Feel free to move discussion to 5836. > Thanks, Jeff. Somehow, I don't see any control elements to change issue

Re: [Cassandra] nodetool compactionstats not showing pending task.

2018-08-22 Thread Oleksandr Shulgin
On Fri, May 5, 2017 at 1:20 PM Alain RODRIGUEZ wrote: > Sorry to hear the restart did not help. > Hi, We are hitting the same issue since a few weeks on version 3.0.16. Normally, restarting an affected node helps, but this is something we would like to avoid doing. What makes it worse for us

Re: Extending Cassandra on AWS from single Region to Multi-Region

2018-08-20 Thread Oleksandr Shulgin
On Thu, Aug 9, 2018 at 3:46 AM srinivasarao daruna wrote: > Hi All, > > We have built Cassandra on AWS EC2 instances. Initially when creating > cluster we have not considered multi-region deployment and we have used AWS > EC2Snitch. > > We have used EBS Volumes to save our data and each of those

Re: Configuration parameter to reject incremental repair?

2018-08-20 Thread Oleksandr Shulgin
On Mon, Aug 13, 2018 at 1:31 PM kurt greaves wrote: > No flag currently exists. Probably a good idea considering the serious > issues with incremental repairs since forever, and the change of defaults > since 3.0. > Hi Kurt, Did you mean since 2.2 (when incremental became the default one)? Or

Re: Adding new datacenter to the cluster

2018-08-20 Thread Oleksandr Shulgin
On Mon, Aug 13, 2018 at 3:50 PM Vitali Dyachuk wrote: > Hello, > I'm going to follow this documentation to add a new datacenter to the C* > cluster > > https://docs.datastax.com/en/cassandra/3.0/cassandra/operations/opsAddDCToCluster.html > > The main step is to run nodetool rebuild which will

Re: Repairs are slow after upgrade to 3.11.3

2018-08-29 Thread Oleksandr Shulgin
On Wed, Aug 29, 2018 at 3:06 AM Maxim Parkachov wrote: > couple of days ago I have upgraded Cassandra from 3.11.2 to 3.11.3 and I > see that repair time is practically doubled. Does someone else experience > the same regression ? > We have upgraded from 3.0.16 to 3.0.17 two days ago and we see

Re: [EXTERNAL] Re: adding multiple node to a cluster, cleanup and num_tokens

2018-09-08 Thread Oleksandr Shulgin
On Sat, 8 Sep 2018, 14:47 Jonathan Haddad, wrote: > 256 tokens is a pretty terrible default setting especially post 3.0. I > recommend folks use 4 tokens for new clusters, > I wonder why not setting it to all the way down to 1 then? What's the key difference once you have so few vnodes? with

Re: [EXTERNAL] Re: adding multiple node to a cluster, cleanup and num_tokens

2018-09-08 Thread Oleksandr Shulgin
On Sat, 8 Sep 2018, 19:00 Jeff Jirsa, wrote: > Virtual nodes accomplish two primary goals > > 1) it makes it easier to gradually add/remove capacity to your cluster by > distributing the new host capacity around the ring in smaller increments > > 2) it increases the number of sources for

Re: nodetool cleanup - compaction remaining time

2018-09-06 Thread Oleksandr Shulgin
On Thu, Sep 6, 2018 at 11:50 AM Alain RODRIGUEZ wrote: > > Be aware that this behavior happens when the compaction throughput is set > to *0 *(unthrottled/unlimited). I believe the estimate uses the speed > limit for calculation (which is often very much wrong anyway). > As far as I can

Drop TTLd rows: upgradesstables -a or scrub?

2018-09-10 Thread Oleksandr Shulgin
Hello, We have some tables with significant amount of TTLd rows that have expired by now (and more gc_grace_seconds have passed since the TTL). We have stopped writing more data to these tables quite a while ago, so background compaction isn't running. The compaction strategy is the default

Re: Drop TTLd rows: upgradesstables -a or scrub?

2018-09-10 Thread Oleksandr Shulgin
an 'nodetool garbagecollect' - that command is not available in the version we are using. It only became available in 3.10. -- Alex > > *From: *Oleksandr Shulgin > *Reply-To: *"user@cassandra.apache.org" > *Date: *Monday, September 10, 2018 at 6:53 AM > *To: *"user@cassandr

Re: Drop TTLd rows: upgradesstables -a or scrub?

2018-09-10 Thread Oleksandr Shulgin
harulata Sharma (charshar) > wrote: > >> Scrub takes a very long time and does not remove the tombstones. You >> should do garbage cleaning. It immediately removes the tombstones. >> >> >> >> Thaks, >> >> Charu >> >> >> >

Re: Recommended num_tokens setting for small cluster

2018-08-30 Thread Oleksandr Shulgin
On Thu, Aug 30, 2018 at 12:05 AM kurt greaves wrote: > For 10 nodes you probably want to use between 32 and 64. Make sure you use > the token allocation algorithm by specifying allocate_tokens_for_keyspace > We are using 16 tokens with 30 nodes on Cassandra 3.0. And yes, we have used

Re: Drop TTLd rows: upgradesstables -a or scrub?

2018-09-11 Thread Oleksandr Shulgin
On Tue, Sep 11, 2018 at 9:31 AM Steinmaurer, Thomas < thomas.steinmau...@dynatrace.com> wrote: > As far as I remember, in newer Cassandra versions, with STCS, nodetool > compact offers a ‘-s’ command-line option to split the output into files > with 50%, 25% … in size, thus in this case, not a

Re: Drop TTLd rows: upgradesstables -a or scrub?

2018-09-11 Thread Oleksandr Shulgin
On Tue, Sep 11, 2018 at 9:47 AM Oleksandr Shulgin < oleksandr.shul...@zalando.de> wrote: > On Tue, Sep 11, 2018 at 9:31 AM Steinmaurer, Thomas < > thomas.steinmau...@dynatrace.com> wrote: > >> As far as I remember, in newer Cassandra versions, with STCS, nodetool >&

Re: Drop TTLd rows: upgradesstables -a or scrub?

2018-09-11 Thread Oleksandr Shulgin
On Tue, Sep 11, 2018 at 10:04 AM Oleksandr Shulgin < oleksandr.shul...@zalando.de> wrote: > > Yet another surprising aspect of using `nodetool compact` is that it > triggers major compaction on *all* nodes in the cluster at the same time. > I don't see where

Fresh SSTable files (due to repair?) in a static table (was Re: Drop TTLd rows: upgradesstables -a or scrub?)

2018-09-11 Thread Oleksandr Shulgin
On Tue, Sep 11, 2018 at 9:47 AM Oleksandr Shulgin < oleksandr.shul...@zalando.de> wrote: > On Tue, Sep 11, 2018 at 9:31 AM Steinmaurer, Thomas < > thomas.steinmau...@dynatrace.com> wrote: > >> As far as I remember, in newer Cassandra versions, with STCS, nodetool >&

Re: Drop TTLd rows: upgradesstables -a or scrub?

2018-09-11 Thread Oleksandr Shulgin
On Mon, Sep 10, 2018 at 10:03 PM Jeff Jirsa wrote: > How much free space do you have, and how big is the table? > So there are 2 tables, one is around 120GB and the other is around 250GB on every node. On the node with the most free disk space we still have around 500GB available and on the

Re: Drop TTLd rows: upgradesstables -a or scrub?

2018-09-11 Thread Oleksandr Shulgin
On Tue, Sep 11, 2018 at 11:07 AM Steinmaurer, Thomas < thomas.steinmau...@dynatrace.com> wrote: > > a single (largish) SSTable or any other SSTable for a table, which does > not get any writes (with e.g. deletes) anymore, will most likely not be > part of an automatic minor compaction anymore,

Upgrading sstables not using all available compaction slots on version 2.2

2018-01-24 Thread Oleksandr Shulgin
Hello, In the process of upgrading our cluster from 2.1 to 2.2 we have triggered the SSTable rewriting process like this: $ nodetool upgradesstables -j 4 # concurrent_compactors=5 Then if we would immediately check the compactionstats, we see that 4 compactions of type 'Upgrade sstables' are

Re: Seed nodes of DC2 creating own versions of system keyspaces

2018-03-06 Thread Oleksandr Shulgin
On 6 Mar 2018 16:55, "Jeff Jirsa" <jji...@gmail.com> wrote: On Mar 6, 2018, at 12:32 AM, Oleksandr Shulgin <oleksandr.shul...@zalando.de> wrote: On 5 Mar 2018 16:13, "Jeff Jirsa" <jji...@gmail.com> wrote: On Mar 5, 2018, at 6:40 AM, Oleksandr Shulgin <

Re: Joining a cluster of nodes having multi valued initial_token parameters.

2018-03-08 Thread Oleksandr Shulgin
On Thu, Mar 8, 2018 at 1:41 PM, Mikhail Tsaplin wrote: > Thank you for the answer, are you sure that it at least safe? > I would test in a lab first of course, but I don't see why it should be a problem. I wonder more why did you have tokens listed explicitly on the

Re: Joining a cluster of nodes having multi valued initial_token parameters.

2018-03-08 Thread Oleksandr Shulgin
On Thu, Mar 8, 2018 at 12:09 PM, Mikhail Tsaplin wrote: > Hi, > > I have a three node Cassandra cluster. Every node has initial_token > configuration parameter holding 256 tokens (looks like randomly > distributed). Now I have to add a fourth node. How could this be done? >

Re: Adding new DC?

2018-03-12 Thread Oleksandr Shulgin
On Sun, Mar 11, 2018 at 10:31 PM, Kunal Gangakhedkar < kgangakhed...@gmail.com> wrote: > Hi all, > > We currently have a cluster in GCE for one of the customers. > They want it to be migrated to AWS. > > I have setup one node in AWS to join into the cluster by following: >

Re: Data Deleted After a few days of being off

2018-02-27 Thread Oleksandr Shulgin
On 27 Feb 2018 20:58, "A" wrote: No. I inserted using USING TTL 86400; Well, that answers your question I believe? -- Alex

Re: command to view yaml file setting in use on console

2018-03-13 Thread Oleksandr Shulgin
On Tue, Mar 13, 2018 at 2:43 AM, Jeff Jirsa wrote: > Cassandra-7622 went patch available today > Jeff, Are you sure you didn't mistype the issue number? I see: https://issues.apache.org/jira/browse/CASSANDRA-7622 Summary: Implement virtual tables Status: Open -- Alex

Re: Cassandra storage: Some thoughts

2018-03-09 Thread Oleksandr Shulgin
On 9 Mar 2018 16:56, "Vangelis Koukis" wrote: Hello all, My name is Vangelis Koukis and I am a Founder and the CTO of Arrikto. I'm writing to share our thoughts on how people run distributed, stateful applications such as Cassandra on modern infrastructure, and would love

Re: uneven data movement in one of the disk in Cassandra

2018-03-09 Thread Oleksandr Shulgin
On Fri, Mar 9, 2018 at 11:40 AM, Yasir Saleem wrote: > Thanks, Nicolas Guyomar > > I am new to cassandra, here is the properties which I can see in yaml file: > > # of compaction, including validation compaction. > compaction_throughput_mb_per_sec: 16 >

Re: Seed nodes of DC2 creating own versions of system keyspaces

2018-03-07 Thread Oleksandr Shulgin
On Tue, Mar 6, 2018 at 8:28 PM, Jeff Jirsa wrote: > > Sorry, I wasnt as precise as I should have been: > > In 3.0 and newer, a bootstrapping node will wait until it has schema > before it bootstraps. HOWEVER, we make the ssystem_auth/system_distributed, > etc keyspaces as a

  1   2   3   >