RE: Best compaction strategy for rarely used data

2023-01-06 Thread onmstester onmstester via user
o:l...@lapo.it> Sent: Friday, December 30, 2022 4:12 AM To: mailto:user@cassandra.apache.org Subject: [EXTERNAL] Re: Best compaction strategy for rarely used data   On 2022-12-29 21: 54, Durity, Sean R via user wrote: > At some point you will end up with large sstables (like 1 TB) t

RE: Best compaction strategy for rarely used data

2022-12-30 Thread Durity, Sean R via user
: [EXTERNAL] Re: Best compaction strategy for rarely used data On 2022-12-29 21: 54, Durity, Sean R via user wrote: > At some point you will end up with large sstables (like 1 TB) that won’t > compact because there are not 4 similar-sized ones able to be compacted Yes, that's exa

Re: Best compaction strategy for rarely used data

2022-12-30 Thread Lapo Luchini
On 2022-12-29 21:54, Durity, Sean R via user wrote: At some point you will end up with large sstables (like 1 TB) that won’t compact because there are not 4 similar-sized ones able to be compacted Yes, that's exactly what's happening. I'll see maybe just one more compaction, since the biggest

RE: Best compaction strategy for rarely used data

2022-12-29 Thread Durity, Sean R via user
, December 29, 2022 4:51 AM To: user@cassandra.apache.org Subject: [EXTERNAL] Re: Best compaction strategy for rarely used data Hi Lapo Take a look at TWCS, I think that could help your use case: https:  //thelastpickle. com/blog/2016/12/08/TWCS-part1. html [thelastpickle. com] Regards Paul Chandler

Re: Best compaction strategy for rarely used data

2022-12-29 Thread Paul Chandler
Hi Lapo Take a look at TWCS, I think that could help your use case: https://thelastpickle.com/blog/2016/12/08/TWCS-part1.html Regards Paul Chandler Sent from my iPhone > On 29 Dec 2022, at 08:55, Lapo Luchini wrote: > > Hi, I have a table which gets (a lot of) data that is written once an

Best compaction strategy for rarely used data

2022-12-29 Thread Lapo Luchini
Hi, I have a table which gets (a lot of) data that is written once and very rarely read (it is used for data that is mandatory for regulatory reasons), and almost never deleted. I'm using the default SCTS as at the time I didn't know any better, but SSTables size are getting huge, which is a p

Re: Deleting Compaction Strategy for Cassandra 3.0?

2020-03-02 Thread Jeff Jirsa
et it right that > running `nodetool upgradesstables -a` is going to rewrite all the SSTable > files subject to the defined compaction strategy? You don’t need to do user defined compaction here As soon as the data files are on the server, the next time TWCS looks for compaction candidates

Re: Deleting Compaction Strategy for Cassandra 3.0?

2020-03-02 Thread Oleksandr Shulgin
. We are OK with rewriting all files once, though. Assuming, we get it running on our server version: do I get it right that running `nodetool upgradesstables -a` is going to rewrite all the SSTable files subject to the defined compaction strategy? -- Alex

Re: Deleting Compaction Strategy for Cassandra 3.0?

2020-02-28 Thread Jeff Jirsa
If you’re really really advanced you MIGHT be able to use spark + cqlsstablewriter to create a ton of sstables with just tombstones one them representing deletes, then either nodetool refresh or sstableloader them into the cluster If you create sstables on the right timestamp boundaries to ma

Re: Deleting Compaction Strategy for Cassandra 3.0?

2020-02-28 Thread Oleksandr Shulgin
On Fri, 28 Feb 2020, 23:02 Erick Ramirez, wrote: > I'm not personally aware of anyone who is using it successfully other > than ProtectWise where it was a good fit for their narrow use case. My > limited knowledge of it is that it has some sharp edges which is the reason > they haven't pushed for

Re: Deleting Compaction Strategy for Cassandra 3.0?

2020-02-28 Thread Erick Ramirez
I'm not personally aware of anyone who is using it successfully other than ProtectWise where it was a good fit for their narrow use case. My limited knowledge of it is that it has some sharp edges which is the reason they haven't pushed for it to be added to Cassandra (that's second hand info so pl

Deleting Compaction Strategy for Cassandra 3.0?

2020-02-28 Thread Oleksandr Shulgin
component can be anything and we don't have a way to enumerate all of them. Even if we could, there are millions of records to remove and we would like to avoid creating any tombstones. Our total storage for this cluster is close to 100 TiB. We found that DCS (Deleting Compaction Strateg

Re: Best compaction strategy

2018-10-26 Thread Jeff Jirsa
> within the TTL it's the wrong strategy. > This diagram is a good rule of the thumb > > > >> On Thu, Oct 25, 2018 at 6:28 AM Alexander Dejanovski >> wrote: >> Hi Raman, >> >> TWCS is the best compaction strategy for TTL data, even if you

Re: Best compaction strategy

2018-10-25 Thread Jonathan Haddad
To add to what Alex suggested, if you know what keys use what TTL you could store them in different tables, with different window settings. Jon On Fri, Oct 26, 2018 at 1:28 AM Alexander Dejanovski wrote: > Hi Raman, > > TWCS is the best compaction strategy for TTL data, even if

Re: Best compaction strategy

2018-10-25 Thread Alexander Dejanovski
Hi Raman, TWCS is the best compaction strategy for TTL data, even if you have different TTLs (set the time window based on your largest TTL, so it would be 1 day in your case). Enable unchecked tombstone compaction to clear the data with 2 days TTL along the way. This is done by setting : ALTER

Best compaction strategy

2018-10-25 Thread raman gugnani
Hi All, I have one table in which i have some data which has TTL of 2days and some data which has TTL of 60 days. What compaction strategy will suits the most. 1. LeveledCompactionStrategy (LCS) 2. SizeTieredCompactionStrategy (STCS) 3. TimeWindowCompactionStrategy (TWCS) -- Raman

Re: Newsletter / Marketing: Re: Compaction Strategy

2018-09-21 Thread Ali Hubail
I suspect that you are CPU bound rather than IO bound. There are a lot of areas to look into, but I would start with a few. I could not tell much from the results you shared since at the time, there were no writes happening. Switching to a different compaction strategy will most likely make it

Re: Compaction Strategy

2018-09-20 Thread rajasekhar kommineni
y scattered across different sstables. size compaction strategy (STCS) is much less expensive than level compaction strategy (LCS). Stopping the background compaction should be approached with caution, I think your problem is more to do with why STCS compaction is taking more resources than you e

Re: Compaction Strategy

2018-09-20 Thread Ali Hubail
size compaction strategy (STCS) is much less expensive than level compaction strategy (LCS). Stopping the background compaction should be approached with caution, I think your problem is more to do with why STCS compaction is taking more resources than you expect. Regards, Ali Hubail Petr

Re: Compaction Strategy

2018-09-19 Thread Nitan Kainth
ction > and it is causing timeouts. > > Also will reducing or increasing compaction_throughput_mb_per_sec eliminate > timeouts ? > > Thanks, > > >> On Sep 17, 2018, at 9:38 PM, rajasekhar kommineni >> wrote: >> >> Hello Folks, >> >

Re: Compaction Strategy

2018-09-19 Thread rajasekhar kommineni
17, 2018, at 9:38 PM, rajasekhar kommineni wrote: > > Hello Folks, > > I need advice in deciding the compaction strategy for my C cluster. There are > multiple jobs that will load the data with less inserts and more updates but > no deletes. Currently I am using Size Tired c

Compaction Strategy

2018-09-17 Thread rajasekhar kommineni
Hello Folks, I need advice in deciding the compaction strategy for my C cluster. There are multiple jobs that will load the data with less inserts and more updates but no deletes. Currently I am using Size Tired compaction, but seeing auto compactions after the data load kicks, and also read

Re: Compaction strategy for update heavy workload

2018-06-13 Thread kurt greaves
ll also solve the problem w.r.t old buckets. > > > > In regards to LCS, the only way to really know if it'll be too much > compaction overhead is to test it, but for the most part you should > consider your read/write ratio, rather than the total number of > reads/writes (unle

Re: Compaction strategy for update heavy workload

2018-06-13 Thread Jonathan Haddad
;s so small that it's irrelevant, which it may well be). > > On 13 June 2018 at 19:25, manuj singh wrote: >> >> Hi all, >> I am trying to determine compaction strategy for our use case. >> In our use case we will have updates on a row a few times. And we have a ttl

Re: Compaction strategy for update heavy workload

2018-06-13 Thread kurt greaves
part you should consider your read/write ratio, rather than the total number of reads/writes (unless it's so small that it's irrelevant, which it may well be). On 13 June 2018 at 19:25, manuj singh wrote: > Hi all, > I am trying to determine compaction strategy for our use case. >

Compaction strategy for update heavy workload

2018-06-13 Thread manuj singh
Hi all, I am trying to determine compaction strategy for our use case. In our use case we will have updates on a row a few times. And we have a ttl also defined on the table level. Our typical workload is less then 1000 writes + reads per second. At the max it could go up to 2500 per second. We

Re: Best compaction strategy for counters tables

2018-01-18 Thread Octavian Rinciog
the past I did not care much about the compaction strategy for counters > as I considered it to be negligible in my case (counters were MB big tables > out of a few TB for the entire dataset). > > You can always pick a strategy you think would work better, and test the > change on a canary n

Re: Best compaction strategy for counters tables

2018-01-18 Thread Alain RODRIGUEZ
SSTables. In the past I did not care much about the compaction strategy for counters as I considered it to be negligible in my case (counters were MB big tables out of a few TB for the entire dataset). You can always pick a strategy you think would work better, and test the change on a canary node

Best compaction strategy for counters tables

2018-01-17 Thread Octavian Rinciog
Hello! I am using Cassandra 3.10. I have a counter table, with the following schema and RF=1 CREATE TABLE edges ( src_id text, src_type text, source text weight counter, PRIMARY KEY ((src_id, src_type), source) ); SELECT vs UPDATE requests ratio for this table is 0.1 READ vs W

Re: changing compaction strategy

2017-03-15 Thread kurt greaves
The rogue pending task is likely a non-issue. If your jmx command went through without errors and you got the log message you can assume it worked. It won't show in the schema unless you run the ALTER statement which affects the whole cluster. If you were switching from STCS then you wouldn't expe

Re: changing compaction strategy

2017-03-14 Thread Mike Torra
a way to tell when/if the local node has successfully updated the compaction strategy? Looking at the sstable files, it seems like they are still based on STCS but I don't know how to be sure. Appreciate any tips or suggestions! On Mon, Mar 13, 2017 at 5:30 PM, Mike Torra wrote: >

changing compaction strategy

2017-03-13 Thread Mike Torra
I'm trying to change compaction strategy one node at a time. I'm using jmxterm like this: `echo 'set -b org.apache.cassandra.db:type=ColumnFamilies,keyspace=my_ks,columnfamily=my_cf CompactionParametersJson \{"class":"TimeWindowCompactionStrategy&qu

Re: Which compaction strategy when modeling a dumb set

2017-02-27 Thread Benjamin Roth
queuing table is anti pattern. > > From: Vincent Rischmann > Sent: Friday, February 24, 2017 06:24 > To: user@cassandra.apache.org > Subject: Which compaction strategy when modeling a dumb set > > Hello, > > I'm using a table

Re: Which compaction strategy when modeling a dumb set

2017-02-27 Thread Rakesh Kumar
cent Rischmann Sent: Friday, February 24, 2017 06:24 To: user@cassandra.apache.org Subject: Which compaction strategy when modeling a dumb set Hello, I'm using a table like this: CREATE TABLE myset (id uuid PRIMARY KEY) which is basically a set I use for deduplication, id is a unique id for

Re: Which compaction strategy when modeling a dumb set

2017-02-27 Thread Rakesh Kumar
: user@cassandra.apache.org Subject: Which compaction strategy when modeling a dumb set Hello, I'm using a table like this: CREATE TABLE myset (id uuid PRIMARY KEY) which is basically a set I use for deduplication, id is a unique id for an event, when I process the event I insert the id

Re: Which compaction strategy when modeling a dumb set

2017-02-27 Thread Vincent Rischmann
lication, id is a unique id >> for an event, when I process the event I insert the id, and before >> processing I check if it has already been processed for >> deduplication. >> >> It works well enough, but I'm wondering which compaction strategy I >> should u

Re: Which compaction strategy when modeling a dumb set

2017-02-27 Thread Vladimir Yudovin
f it has already been processed for deduplication. It works well enough, but I'm wondering which compaction strategy I should use. I expect maybe 1% or less of events will end up duplicated (thus not generating an insert), so the workload will probably be 50% writes 50% read. Is LCS a good

Re: Which compaction strategy when modeling a dumb set

2017-02-24 Thread kurt greaves
Probably LCS although what you're implying (read before write) is an anti-pattern in Cassandra. Something like this is a good indicator that you should review your model. ​

Which compaction strategy when modeling a dumb set

2017-02-24 Thread Vincent Rischmann
ation. It works well enough, but I'm wondering which compaction strategy I should use. I expect maybe 1% or less of events will end up duplicated (thus not generating an insert), so the workload will probably be 50% writes 50% read. Is LCS a good strategy here or should I stick with STCS ?

Re: Choosing a compaction strategy (TWCS)

2016-12-21 Thread Voytek Jarnot
it’ll compact it in a >>> single sstable compaction). >>> >>> >>> >>> - Jeff >>> >>> >>> >>> *From: *Voytek Jarnot >>> *Reply-To: *"user@cassandra.apache.org" >>> *Date: *Friday, Decem

Re: Choosing a compaction strategy (TWCS)

2016-12-17 Thread Voytek Jarnot
able >> over 24 hours old that’s at least 80% tombstones, it’ll compact it in a >> single sstable compaction). >> >> >> >> - Jeff >> >> >> >> *From: *Voytek Jarnot >> *Reply-To: *"user@cassandra.apache.org" >> *Date: *Friday,

Re: Choosing a compaction strategy (TWCS)

2016-12-16 Thread Jeff Jirsa
that’s at least 80% tombstones, it’ll compact it in a >> single sstable compaction). >> >> >> >> - Jeff >> >> >> >> From: Voytek Jarnot >> Reply-To: "user@cassandra.apache.org" >> Date: Friday, December 16, 2016 at 7:3

Re: Choosing a compaction strategy (TWCS)

2016-12-16 Thread Voytek Jarnot
ber 16, 2016 at 7:34 PM > > *To: *"user@cassandra.apache.org" > *Subject: *Re: Choosing a compaction strategy (TWCS) > > > > Thanks again, Jeff. > > > > Thinking about this some more, I'm wondering if I'm overthinking or if > there's a p

Re: Choosing a compaction strategy (TWCS)

2016-12-16 Thread Jeff Jirsa
Friday, December 16, 2016 at 11:12 AM To: "user@cassandra.apache.org" Subject: Re: Choosing a compaction strategy (TWCS) Thank you Jeff - always nice to hear straight from the source. Any issues you can see with 3 (my calendar-week bucket not aligning with the arbitrary 7-day window

Re: Choosing a compaction strategy (TWCS)

2016-12-16 Thread Voytek Jarnot
uot; > *Date: *Friday, December 16, 2016 at 11:12 AM > > *To: *"user@cassandra.apache.org" > *Subject: *Re: Choosing a compaction strategy (TWCS) > > > > Thank you Jeff - always nice to hear straight from the source. > > > > Any issues you can see with 3 (my calendar

Re: Choosing a compaction strategy (TWCS)

2016-12-16 Thread Jeff Jirsa
. - Jeff From: Voytek Jarnot Reply-To: "user@cassandra.apache.org" Date: Friday, December 16, 2016 at 11:12 AM To: "user@cassandra.apache.org" Subject: Re: Choosing a compaction strategy (TWCS) Thank you Jeff - always nice to hear straight from the source.

Re: Choosing a compaction strategy (TWCS)

2016-12-16 Thread Voytek Jarnot
table – it > won’t kill you, but if you can avoid it, avoid it. > > > > - Jeff > > > > > > *From: *Jeff Jirsa > *Date: *Friday, December 16, 2016 at 10:47 AM > *To: *"user@cassandra.apache.org" > *Subject: *Re: Choosing a

Re: Choosing a compaction strategy (TWCS)

2016-12-16 Thread Jeff Jirsa
. - Jeff From: Jeff Jirsa Date: Friday, December 16, 2016 at 10:47 AM To: "user@cassandra.apache.org" Subject: Re: Choosing a compaction strategy (TWCS) With a 10 year retention, just ignore the target sstable count (I should remove that guidance, to be honest), and

Re: Choosing a compaction strategy (TWCS)

2016-12-16 Thread Jeff Jirsa
bloom filter is going to make things nice and easy for you. - Jeff From: Voytek Jarnot Reply-To: "user@cassandra.apache.org" Date: Friday, December 16, 2016 at 10:37 AM To: "user@cassandra.apache.org" Subject: Choosing a compaction strategy (TWCS) Scena

Choosing a compaction strategy (TWCS)

2016-12-16 Thread Voytek Jarnot
Scenario: Converting an Oracle table to Cassandra, one Oracle table to 4 Cassandra tables, basically time-series - think log or auditing. Retention is 10 years, but greater than 95% of reads will occur on data written within the last year. 7 day TTL used on a small percentage of the records, major

Re: Question about compaction strategy changes

2016-10-24 Thread kurt Greaves
On 24 October 2016 at 18:11, Seth Edwards wrote: > The other thought is that we currently have data mixed in that does not > have a TTL and we are strongly considering putting this data in it's own > table. You should definitely do that. Having non-TTL'd data mixed in will result in SSTables th

Re: Question about compaction strategy changes

2016-10-24 Thread Seth Edwards
> *Reply-To: *"user@cassandra.apache.org" > *Date: *Sunday, October 23, 2016 at 7:03 PM > *To: *user > *Subject: *Re: Question about compaction strategy changes > > > > More compactions meaning "rows to be compacted" or actual number of > pending comp

Re: Question about compaction strategy changes

2016-10-24 Thread Jeff Jirsa
consumed by compaction in a TWCS setting. From: Seth Edwards Reply-To: "user@cassandra.apache.org" Date: Sunday, October 23, 2016 at 7:03 PM To: user Subject: Re: Question about compaction strategy changes More compactions meaning "rows to be compacted" or actu

Re: Question about compaction strategy changes

2016-10-23 Thread kurt Greaves
​More compactions meaning "actual number of compaction tasks". A compaction task generally operates on many SSTables (how many depends on the chosen compaction strategy). The number of pending tasks does not line up with the number of SSTables that will be compacted. 1 task may co

Re: Question about compaction strategy changes

2016-10-23 Thread Seth Edwards
More compactions meaning "rows to be compacted" or actual number of pending compactions? I assumed when I run nodetool compactionstats the number of pending tasks would line up with number of sstables that will be compacted. Most of the time this is idle, then we hit spots when it could jump into t

Re: Question about compaction strategy changes

2016-10-23 Thread kurt Greaves
On 22 October 2016 at 03:37, Seth Edwards wrote: > We're using TWCS and we notice that if we make changes to the options to > the window unit or size, it seems to implicitly start recompacting all > sstables. If you increase the window unit or size you potentially increase the number of SSTable

Re:Question about compaction strategy changes

2016-10-21 Thread Zhao Yang
hi Edwards, when changibg gc_grace_second, no compaction willbbe triggered. regards, jasonstack Sent from my Mi phoneOn Seth Edwards , Oct 22, 2016 11:37 AM wrote:Hello! We're using TWCS and we notice that if we make changes to the options to the window unit or size, it seems to implicitly s

Question about compaction strategy changes

2016-10-21 Thread Seth Edwards
Hello! We're using TWCS and we notice that if we make changes to the options to the window unit or size, it seems to implicitly start recompacting all sstables. Is this indeed the case and more importantly, does the same happen if we were to adjust the gr_grace_seconds for this table? Thanks!

Re: upgradesstables/cleanup/compaction strategy change

2016-05-23 Thread Vitalii Skakun
e: > >> Hi! >> >> I have a 2.0.13 cluster which I have just extended, and I'm now looking >> into upgrading it to 2.1. >> >> * The cleanup after the extension is partially done. >> * I'm also looking into changing a few tables into Leveled Compaction >

Re: upgradesstables/cleanup/compaction strategy change

2016-05-23 Thread Vitalii Skakun
xtended, and I'm now looking > into upgrading it to 2.1. > > * The cleanup after the extension is partially done. > * I'm also looking into changing a few tables into Leveled Compaction > Strategy. > > In the interest of speeding up things by avoiding unnecessary rewrit

upgradesstables/cleanup/compaction strategy change

2016-05-23 Thread Erik Forsberg
Hi! I have a 2.0.13 cluster which I have just extended, and I'm now looking into upgrading it to 2.1. * The cleanup after the extension is partially done. * I'm also looking into changing a few tables into Leveled Compaction Strategy. In the interest of speeding up things b

Re: Impact of Changing Compaction Strategy

2016-01-15 Thread Jeff Jirsa
When you change compaction strategy, nothing happens until the next flush. On the next flush, the new compaction strategy will decide what to do – if you change from STCS to DTCS, it will look at various timestamps of files, and attempt to group them by time windows based on the sstable’s

Impact of Changing Compaction Strategy

2016-01-15 Thread Anuj Wadehra
Hi, I need to understand whether all existing sstables are recreated/updated when we change compaction strategy from STCS to DTCS? Sstables are immutable by design but do we take an exception for such cases and update same files when an Alter statement is fired to change the compaction

Re: Leveled Compaction Strategy with a really intensive delete workload

2015-05-26 Thread Stefano Ortolani
>>> to my question above would help second guess my a decision a bit less :) >>> >>> Cheers, >>> Stefano >>> >>> On Mon, May 25, 2015 at 9:52 AM, Jason Wee wrote: >>> >>>> , due to a really intensive delete workloads, th

Re: Leveled Compaction Strategy with a really intensive delete workload

2015-05-25 Thread Jason Wee
>>> promoted to t.. >>> >>> Is cassandra design for *delete* workloads? doubt so. Perhaps looking at >>> some other alternative like ttl? >>> >>> jason >>> >>> On Mon, May 25, 2015 at 10:12 AM, Manoj Khangaonkar < >>>

Re: Leveled Compaction Strategy with a really intensive delete workload

2015-05-25 Thread Stefano Ortolani
mmendation seems to be >>> that leveled compaction is suited for read intensive workloads. >>> >>> Depending on your use case, you might better of with data tiered or size >>> tiered strategy. >>> >>> regards >>> >>> regards >&g

Re: Leveled Compaction Strategy with a really intensive delete workload

2015-05-25 Thread Stefano Ortolani
tion seems to be that >> leveled compaction is suited for read intensive workloads. >> >> Depending on your use case, you might better of with data tiered or size >> tiered strategy. >> >> regards >> >> regards >> >>> On Sun, May 24, 2015

Re: Leveled Compaction Strategy with a really intensive delete workload

2015-05-25 Thread Jason Wee
ed strategy. > > regards > > regards > > On Sun, May 24, 2015 at 10:50 AM, Stefano Ortolani > wrote: > >> Hi all, >> >> I have a question re leveled compaction strategy that has been bugging me >> quite a lot lately. Based on what I understood, a compacti

Re: Leveled Compaction Strategy with a really intensive delete workload

2015-05-24 Thread Manoj Khangaonkar
strategy. regards regards On Sun, May 24, 2015 at 10:50 AM, Stefano Ortolani wrote: > Hi all, > > I have a question re leveled compaction strategy that has been bugging me > quite a lot lately. Based on what I understood, a compaction takes place > when the SSTable gets to a specific

Leveled Compaction Strategy with a really intensive delete workload

2015-05-24 Thread Stefano Ortolani
Hi all, I have a question re leveled compaction strategy that has been bugging me quite a lot lately. Based on what I understood, a compaction takes place when the SSTable gets to a specific size (10 times the size of its previous generation). My question is about an edge case where, due to a

Re: Date Tiered Compaction Strategy and collections

2014-11-28 Thread Eric Stevens
The underlying write time is still tracked for each value in the collection - it's part of how conflict resolution is managed - but it's not exposed through CQL. On Fri Nov 28 2014 at 4:18:47 AM Batranut Bogdan wrote: > Hello all, > > If one has a table like this: > id text, > ts timestamp > val

Date Tiered Compaction Strategy and collections

2014-11-28 Thread Batranut Bogdan
Hello all, If one has a table like this:id text,ts timestampvalues list  PK (id,ts)  How will the DTCS work? I am asking this because the writeTime() function does not work on collections.

Re: Compaction Strategy guidance

2014-11-25 Thread Andrei Ivanov
>> >> Hi Andrei, Hi Nicolai, >> >> >> >> Which version of C* are you using ? >> >> >> >> There are some recommendations about the max storage per node : >> >> >> >> http://www.datastax.com/dev/blog/performance-improvements-in-ca

Re: Compaction Strategy guidance

2014-11-25 Thread Nikolai Grigoriev
t the max storage per node : > >> > http://www.datastax.com/dev/blog/performance-improvements-in-cassandra-1-2 > >> > >> "For 1.0 we recommend 300-500GB. For 1.2 we are looking to be able to > >> handle 10x > >> (3-5TB)". > >> > &

Re: Compaction Strategy guidance

2014-11-25 Thread Andrei Ivanov
r 1.0 we recommend 300-500GB. For 1.2 we are looking to be able to >> handle 10x >> (3-5TB)". >> >> I have the feeling that those recommendations are sensitive according many >> criteria such as : >> - your hardware >> - the compaction strategy >> - ...

Re: Compaction Strategy guidance

2014-11-25 Thread Nikolai Grigoriev
uot;For 1.0 we recommend 300-500GB. For 1.2 we are looking to be able to > handle 10x > (3-5TB)". > > I have the feeling that those recommendations are sensitive according many > criteria such as : > - your hardware > - the compaction strategy > - ... > > It looks

Re: Compaction Strategy guidance

2014-11-25 Thread Andrei Ivanov
> There are some recommendations about the max storage per node : >> > >> > http://www.datastax.com/dev/blog/performance-improvements-in-cassandra-1-2 >> > >> > "For 1.0 we recommend 300-500GB. For 1.2 we are looking to be able to >> > handle >> > 10x

Re: Compaction Strategy guidance

2014-11-25 Thread Marcus Eriksson
-1-2 > > > > "For 1.0 we recommend 300-500GB. For 1.2 we are looking to be able to > handle > > 10x > > (3-5TB)". > > > > I have the feeling that those recommendations are sensitive according > many > > criteria such as : > > - your ha

Re: Compaction Strategy guidance

2014-11-25 Thread Andrei Ivanov
d 300-500GB. For 1.2 we are looking to be able to handle > 10x > (3-5TB)". > > I have the feeling that those recommendations are sensitive according many > criteria such as : > - your hardware > - the compaction strategy > - ... > > It looks that LCS lower those limita

Re: Compaction Strategy guidance

2014-11-25 Thread Jean-Armel Luce
". I have the feeling that those recommendations are sensitive according many criteria such as : - your hardware - the compaction strategy - ... It looks that LCS lower those limitations. Increasing the size of sstables might help if you have enough CPU and you can put more load on your

Re: Compaction Strategy guidance

2014-11-24 Thread Robert Coli
On Mon, Nov 24, 2014 at 6:48 AM, Nikolai Grigoriev wrote: > One of the obvious recommendations I have received was to run more than > one instance of C* per host. Makes sense - it will reduce the amount of > data per node and will make better use of the resources. > This is usually a Bad Idea to

Re: Compaction Strategy guidance

2014-11-24 Thread Andrei Ivanov
sure if this node will ever get back to normal life. And >> >> >>>> believe me >> >> >>>> - this is not because of I/O, I have SSDs everywhere and 16 >> >> >>>> physical >> >> >>>> cores. >> >> >>&

Re: Compaction Strategy guidance

2014-11-24 Thread Nikolai Grigoriev
lose to 2x disk > space > >> >>>> on > >> >>>> EVERY disk in my JBOD configuration...this will kill the node > sooner > >> >>>> or > >> >>>> later. This is all because all sstables after bootstrap end at L0 > and > &g

Re: Compaction Strategy guidance

2014-11-24 Thread Andrei Ivanov
otstrap end at L0 and >> >>>> then >> >>>> the process slowly slowly moves them to other levels. If you have >> >>>> write >> >>>> traffic to that CF then the number of sstables and L0 will grow >> >>

Re: Compaction Strategy guidance

2014-11-24 Thread Nikolai Grigoriev
e > https://issues.apache.org/jira/browse/CASSANDRA-8301 > >>>> is implemented it may be better. > >>>> > >>>> > >>>> On Sun, Nov 23, 2014 at 4:53 AM, Andrei Ivanov > >>>> wrote: > >>>>> &g

Re: Compaction Strategy guidance

2014-11-24 Thread Andrei Ivanov
ewhat similar C* load profile. Hence some comments >>>>> in addition Nikolai's answer. >>>>> 1. Fallback to STCS - you can disable it actually >>>>> 2. Based on our experience, if you have a lot of data per node, LCS >>>>> may work j

Re: Compaction Strategy guidance

2014-11-24 Thread Nikolai Grigoriev
ill not be able >>>> to compact what it gets from old nodes. In your case, if you switch >>>> strategy the same thing may happen. This is all due to limitations >>>> mentioned by Nikolai. >>>> >>>> Andrei, >>>> >>>>

Re: Compaction Strategy guidance

2014-11-23 Thread Andrei Ivanov
may work just fine. That is, till the moment you decide to join >>>> another node - chances are that the newly added node will not be able >>>> to compact what it gets from old nodes. In your case, if you switch >>>> strategy the same thing may happen. This is all due to li

Re: Compaction Strategy guidance

2014-11-23 Thread Jean-Armel Luce
ai. >>> >>> Andrei, >>> >>> >>> On Sun, Nov 23, 2014 at 8:51 AM, Servando Muñoz G. >>> wrote: >>> > ABUSE >>> > >>> > >>> > >>> > YA NO QUIERO MAS MAILS SOY DE MEXICO >>> >

Re: Compaction Strategy guidance

2014-11-23 Thread Jean-Armel Luce
t; >> > >> > YA NO QUIERO MAS MAILS SOY DE MEXICO >> > >> > >> > >> > De: Nikolai Grigoriev [mailto:ngrigor...@gmail.com] >> > Enviado el: sábado, 22 de noviembre de 2014 07:13 p. m. >> > Para: user@cassandra.apache.org >>

Re: Compaction Strategy guidance

2014-11-23 Thread Nikolai Grigoriev
gt; > > > > > > De: Nikolai Grigoriev [mailto:ngrigor...@gmail.com] > > Enviado el: sábado, 22 de noviembre de 2014 07:13 p. m. > > Para: user@cassandra.apache.org > > Asunto: Re: Compaction Strategy guidance > > Importancia: Alta > > > > > >

Re: Compaction Strategy guidance

2014-11-23 Thread Andrei Ivanov
t; ABUSE > > > > YA NO QUIERO MAS MAILS SOY DE MEXICO > > > > De: Nikolai Grigoriev [mailto:ngrigor...@gmail.com] > Enviado el: sábado, 22 de noviembre de 2014 07:13 p. m. > Para: user@cassandra.apache.org > Asunto: Re: Compaction Strategy guidance > Importancia: A

RE: Compaction Strategy guidance

2014-11-22 Thread Servando Muñoz G .
ABUSE YA NO QUIERO MAS MAILS SOY DE MEXICO De: Nikolai Grigoriev [mailto:ngrigor...@gmail.com] Enviado el: sábado, 22 de noviembre de 2014 07:13 p. m. Para: user@cassandra.apache.org Asunto: Re: Compaction Strategy guidance Importancia: Alta Stephane, As everything good, LCS comes at

Re: Compaction Strategy guidance

2014-11-22 Thread Nikolai Grigoriev
Stephane, As everything good, LCS comes at certain price. LCS will put most load on you I/O system (if you use spindles - you may need to be careful about that) and on CPU. Also LCS (by default) may fall back to STCS if it is falling behind (which is very possible with heavy writing activity) and

Compaction Strategy guidance

2014-11-22 Thread Stephane Legay
Hi there, use case: - Heavy write app, few reads. - Lots of updates of rows / columns. - Current performance is fine, for both writes and reads.. - Currently using SizedCompactionStrategy We're trying to limit the amount of storage used during compaction. Should we switch to LeveledCompactionStr

Re: Could table partitioning be implemented using a customer compaction strategy?

2014-08-15 Thread DuyHai Doan
Check that: https://issues.apache.org/jira/browse/CASSANDRA-6602 There is a patch for a compaction strategy dedicated to time series data. The discussion is also interesting in the comments. On Fri, Aug 15, 2014 at 6:28 AM, Kevin Burton wrote: > We use log structured tables to hold logs

Could table partitioning be implemented using a customer compaction strategy?

2014-08-14 Thread Kevin Burton
space, you can just drop the oldest day's worth of data without having to use tombstones. MySQL has a somewhat decent partition engine: http://dev.mysql.com/doc/refman/5.1/en/partitioning.html It seems like this come be easily implemented using a custom compaction strategy. Essentially, you

Re: Customized Compaction Strategy: Dev Questions

2014-06-04 Thread Russell Bradberry
able as well? Are you talking about making a new compaction strategy that creates SSTables by day? On June 4, 2014 at 1:36:10 PM, Redmumba (redmu...@gmail.com) wrote: Let's say I run a major compaction every day, so that the "oldest" sstable contains only the data for January 1st. 

Re: Customized Compaction Strategy: Dev Questions

2014-06-04 Thread Redmumba
>> >> >> On Wed, Jun 4, 2014 at 10:40 AM, Russell Bradberry >> wrote: >> >>> Maybe I’m misunderstanding something, but what makes you think that >>> running a major compaction every day will cause they data from January 1st >>> to exist in only on

Re: Customized Compaction Strategy: Dev Questions

2014-06-04 Thread Jonathan Haddad
r days in the >> SSTable as well? Are you talking about making a new compaction strategy >> that creates SSTables by day? >> >> >> >> On June 4, 2014 at 1:36:10 PM, Redmumba (redmu...@gmail.com) wrote: >> >> Let's say I run a major compaction

  1   2   >