Re: Cassandra Tombstone

2019-06-20 Thread Alain RODRIGUEZ
Hello Aneesh, Reading your message and answers given, I really think this post I wrote about 3 years ago now (how quickly time goes through...) about tombstone might be of interest to you: https://thelastpickle.com/blog/2016/07/27/about-deletes-and-tombstones.html. Your problem is not related

Re: Cassandra Tombstone

2019-06-18 Thread Oleksandr Shulgin
On Tue, Jun 18, 2019 at 8:06 AM ANEESH KUMAR K.M wrote: > > I am using Cassandra cluster with 3 nodes which is hosted on AWS. Also we > have NodeJS web Application which is on AWS ELB. Now the issue is that, > when I add 2 or more servers (nodeJS) in AWS ELB then the delete queries > are not

Re: Cassandra Tombstone

2019-06-18 Thread Jeff Jirsa
This is nearly impossible to answer without much more info, but suspect you’re either: Using very weak consistency levels or some weirdness with data centers / availability zones (like simplestrategy and local_*), or Have bad clocks / no ntp / wrong time zones, > On Jun 17, 2019, at 11:05

Cassandra Tombstone

2019-06-18 Thread ANEESH KUMAR K.M
Hi, I am using Cassandra cluster with 3 nodes which is hosted on AWS. Also we have NodeJS web Application which is on AWS ELB. Now the issue is that, when I add 2 or more servers (nodeJS) in AWS ELB then the delete queries are not working on Cassandra. Its working when there is only one server

Re: Questions about C* performance related to tombstone

2019-04-10 Thread Li, George
s are fine. > > Sadly there's a lot of hand wringing about tombstones in the generic > sense which leads people to try to work around *every* case where > they're used. This is unnecessary. A tombstone over a single row > isn't a problem, especially if you're only fetching that one

Re: Questions about C* performance related to tombstone

2019-04-10 Thread Alok Dwivedi
reappearing. Regards Alok > On 9 Apr 2019, at 15:56, Jon Haddad wrote: > > Normal deletes are fine. > > Sadly there's a lot of hand wringing about tombstones in the generic > sense which leads people to try to work around *every* case where > they're used. This is unneces

Re: Questions about C* performance related to tombstone

2019-04-09 Thread Jon Haddad
Normal deletes are fine. Sadly there's a lot of hand wringing about tombstones in the generic sense which leads people to try to work around *every* case where they're used. This is unnecessary. A tombstone over a single row isn't a problem, especially if you're only fetching that one row back

Questions about C* performance related to tombstone

2019-04-09 Thread Li, George
Would query "SELECT * FROM myTable WHERE course_id = 'C' AND assignment_id = 'A2';" be affected too? For query "SELECT * FROM myTable WHERE course_id = 'C';", to workaround the tombstone problem, we are thinking about not doing hard deletes, instead doing soft deletes. So ins

Re: TWCS and tombstone purging

2019-03-18 Thread Alexander Dejanovski
n our blog <http://thelastpickle.com/blog/> that cover the tombstones and compaction strategies topic (search for "tombstone" on that page), notably this one: http://thelastpickle.com/blog/2016/07/27/about-deletes-and-tombstones.html Cheers, On Sat, Mar 16, 2019 at 1:04 AM Nick Hatfield wrote:

TWCS and tombstone purging

2019-03-15 Thread Nick Hatfield
Hey guys, Can someone give me some idea or link some good material for determining a good / aggressive tombstone strategy? I want to make sure my tombstones are getting purged as soon as possible to reclaim disk. Thanks

Re: Tombstone removal optimization and question

2018-11-06 Thread DuyHai Doan
>> Hello all >> >> I have tried to sum up all rules related to tombstone removal: >> >> ---- >> -- >> >> Given a tombstone written at timestamp (t) for a partition key

Re: Tombstone removal optimization and question

2018-11-06 Thread kurt greaves
Yes it does. Consider if it didn't and you kept writing to the same partition, you'd never be able to remove any tombstones for that partition. On Tue., 6 Nov. 2018, 19:40 DuyHai Doan Hello all > > I have tried to sum up all rules related to tombstone r

Tombstone removal optimization and question

2018-11-06 Thread DuyHai Doan
Hello all I have tried to sum up all rules related to tombstone removal: -- Given a tombstone written at timestamp (t) for a partition key (P) in SSTable (S1). This tombstone will be removed: 1) after

Re: Tombstone experience

2018-08-24 Thread Rahul Singh
Executive Officer m 202.905.2818 Anant Corporation 1010 Wisconsin Ave NW, Suite 250 Washington, D.C. 20007 We build and manage digital business technology platforms. On Aug 24, 2018, 1:46 AM -0400, Charulata Sharma (charshar) , wrote: > Hi All, > >    I have shared my experience of

Tombstone experience

2018-08-23 Thread Charulata Sharma (charshar)
Hi All, I have shared my experience of tombstone clearing in this blog post. Sharing it in this forum for wider distribution. https://medium.com/cassandra-tombstones-clearing-use-case/the-curios-case-of-tombstones-d897f681a378 Thanks, Charu

RE: [EXTERNAL] Re: Tombstone

2018-06-21 Thread Rahul Singh
hing else lined up properly to solve a queue problem. > > > Sean Durity > > From: Abhishek Singh > Sent: Tuesday, June 19, 2018 10:41 AM > To: user@cassandra.apache.org > Subject: [EXTERNAL] Re: Tombstone > >                        The Partition key is made of datet

RE: [EXTERNAL] Re: Tombstone

2018-06-19 Thread Durity, Sean R
:41 AM To: user@cassandra.apache.org Subject: [EXTERNAL] Re: Tombstone The Partition key is made of datetime(basically date truncated to hour) and bucket.I think your RCA may be correct since we are deleting the partition rows one by one not in a batch files maybe overlapping

Re: Tombstone

2018-06-19 Thread Abhishek Singh
> > > > Hi all, > >We using Cassandra for storing events which are time series > based for batch processing once a particular batch based on hour is > processed we delete the entries but we were left with almost 18% deletes > marked as Tombstones. > >

Re: Tombstone

2018-06-19 Thread Jeff Jirsa
d as > Tombstones. > I ran compaction on the particular CF tombstone didn't come > down. > Can anyone suggest what is the optimal tunning/recommended > practice used for compaction strategy and GC_grace period with 100k entries > and deletes every hour

Re: Tombstone

2018-06-19 Thread Evelyn Smith
18% deletes marked as > Tombstones. > I ran compaction on the particular CF tombstone didn't come > down. > Can anyone suggest what is the optimal tunning/recommended > practice used for compaction strategy and GC_grace period with 100k entries > and d

Tombstone

2018-06-19 Thread Abhishek Singh
tombstone didn't come down. Can anyone suggest what is the optimal tunning/recommended practice used for compaction strategy and GC_grace period with 100k entries and deletes every hour. Warm Regards Abhishek Singh

Re: about the tombstone and hinted handoff

2018-04-16 Thread kurt greaves
> handoff to replay the database mutationsthe node missed while it was > down. Cassandra does not replay a mutation for a tombstoned record > during its grace period.". > > The tombstone here is on the recovered node or coordinator? > The tombstone is a special write record, so i

about the tombstone and hinted handoff

2018-04-12 Thread Jinhua Luo
toned record during its grace period.". The tombstone here is on the recovered node or coordinator? The tombstone is a special write record, so it must have writetime. We could compare the writetime between the version in the hint and the version of the tombstone, which is enough to make choice,

Re: TWCS enabling tombstone compaction

2018-03-12 Thread Lerh Chuan Low
;lu...@maurobenevides.com.br> wrote: > Dear community, > > I have been using TWCS in my lab, with TTL'd data. > In the debug log there is always the sentence: > "TimeWindowCompactionStrategy.java:65 Disabling tombstone compactions for > TWCS". Indeed, the line is always r

TWCS enabling tombstone compaction

2018-03-09 Thread Lucas Benevides
Dear community, I have been using TWCS in my lab, with TTL'd data. In the debug log there is always the sentence: "TimeWindowCompactionStrategy.java:65 Disabling tombstone compactions for TWCS". Indeed, the line is always repeated. What does it actually mean? If my data gets expired

Re: Tombstone warnings in log file

2017-12-07 Thread Alain RODRIGUEZ
Hello Simon. Tombstone is a tricky topic in Cassandra that brought a lot of questions over time. I exposed my understanding in a blog post last year and thought it might be of interest for you, even though things probably evolved a bit, principles and tuning did not change that much I guess

RE: Tombstone warnings in log file

2017-12-04 Thread wxn...@zjqunshuo.com
Got it. Thank you. From: Meg Mara Date: 2017-12-05 01:54 To: user@cassandra.apache.org Subject: RE: Tombstone warnings in log file Simon, It means that in processing your queries, Cassandra is going through that many tombstone cells in order to return your results. It is because some

RE: Tombstone warnings in log file

2017-12-04 Thread Meg Mara
Simon, It means that in processing your queries, Cassandra is going through that many tombstone cells in order to return your results. It is because some of the partitions that you are querying for have already expired. The warning is just cassandra's way of letting you know that your reads

Tombstone warnings in log file

2017-12-03 Thread wxn...@zjqunshuo.com
Hi, My cluster is running 2.2.8, no update and deletion, only insertion with TTL. I saw below warnings reacently. What's the meaning of them and what's the impact? WARN [SharedPool-Worker-2] 2017-12-04 09:32:48,833 SliceQueryFilter.java:308 - Read 2461 live and 1978 tombstone cells

Re: old big tombstone data file occupy much disk space

2017-09-04 Thread Shalom Sagges
sonInc> We >>>> Create Meaningful Connections >>>> >>>> >>>> >>>> On Sat, Sep 2, 2017 at 8:34 PM, Jeff Jirsa <jji...@gmail.com> wrote: >>>> >>>>> If you're on 3.0 (3.0.6 or 3.0.8 or newer I don't remem

Re: old big tombstone data file occupy much disk space

2017-09-04 Thread Nicolas Guyomar
ebook.com/LivePersonInc> We Create Meaningful Connections >>> >>> >>> >>> On Sat, Sep 2, 2017 at 8:34 PM, Jeff Jirsa <jji...@gmail.com> wrote: >>> >>>> If you're on 3.0 (3.0.6 or 3.0.8 or newer I don't remember which), TWCS >>&g

Re: old big tombstone data file occupy much disk space

2017-09-04 Thread Nicolas Guyomar
:34 PM, Jeff Jirsa <jji...@gmail.com> wrote: >> >>> If you're on 3.0 (3.0.6 or 3.0.8 or newer I don't remember which), TWCS >>> was designed for ttl-only time series use cases >>> >>> Alternatively, if you have IO to spare, you may find LCS works a

Re: old big tombstone data file occupy much disk space

2017-09-04 Thread Shalom Sagges
as well >> (it'll cause quite a bit more compaction, but a much higher chance to >> compact away tombstones) >> >> There are also tombstone focused sub properties to more aggressively >> compact sstables that have a lot of tombstones - check the docs for >> "unchecked tombst

Re: old big tombstone data file occupy much disk space

2017-09-03 Thread Shalom Sagges
it'll cause quite a bit more compaction, but a much higher chance to > compact away tombstones) > > There are also tombstone focused sub properties to more aggressively > compact sstables that have a lot of tombstones - check the docs for > "unchecked tombstone compaction" and &quo

Re: old big tombstone data file occupy much disk space

2017-09-02 Thread Jeff Jirsa
) There are also tombstone focused sub properties to more aggressively compact sstables that have a lot of tombstones - check the docs for "unchecked tombstone compaction" and "tombstone threshold" - enabling those will enable more aggressive automatic single-sstable compacti

Re: old big tombstone data file occupy much disk space

2017-09-02 Thread qf zhou
Yes, your are right. I am using STCS compaction strategy with some kind of timeseries model. Too much disk space has been occupied. What should I do to stop the disk full ? I only want to keep 100 days data most recently, so I set default_time_to_live = 864(100 days ). I know I

Re: old big tombstone data file occupy much disk space

2017-09-02 Thread Nicolas Guyomar
-uber.jar -l >>>> localhost:7199 >>>> >>>> In the above, I am using a jmx method. But it seems that the file size >>>> doesn’t change. My command is wrong ? >>>> >>>> > 在 2017年9月1日,下午2:17,Jeff Jirsa <jji...@gmail.com> 写道:

Re: old big tombstone data file occupy much disk space

2017-09-01 Thread qf zhou
>> In the above, I am using a jmx method. But it seems that the file size >>>> doesn’t change. My command is wrong ? >>>> >>>> > 在 2017年9月1日,下午2:17,Jeff Jirsa <jji...@gmail.com >>>> > <mailto:jji...@gmail.com>> 写道: >>>&g

Re: old big tombstone data file occupy much disk space

2017-09-01 Thread Nicolas Guyomar
er defined compaction to do a single sstable compaction on just that >>> sstable >>> > >>> > It's a nodetool command in very recent versions, or a jmx method in >>> older versions >>> > >>> > >>> > -- >>> > Jeff Jirs

Re: old big tombstone data file occupy much disk space

2017-09-01 Thread qf zhou
as about 1.5T data in the disk. >> I found some sstables file are over 300G. Using the  sstablemetadata command,  I found it:  Estimated droppable tombstones: 0.9622972799707109. >> It is obvious that too much tombstone data exists. >> The default_time_to_live = 864(100 days)

Re: old big tombstone data file occupy much disk space

2017-09-01 Thread Nicolas Guyomar
In the above, I am using a jmx method. But it seems that the file size >>>> doesn’t change. My command is wrong ? >>>> >>>> > 在 2017年9月1日,下午2:17,Jeff Jirsa <jji...@gmail.com> 写道: >>>> > >>>> > User defined compaction to do a si

Re: old big tombstone data file occupy much disk space

2017-09-01 Thread Nicolas Guyomar
;>> sstable >>> > >>> > It's a nodetool command in very recent versions, or a jmx method in >>> older versions >>> > >>> > >>> > -- >>> > Jeff Jirsa >>> > >>> > >>> >> O

Re: old big tombstone data file occupy much disk space

2017-09-01 Thread qf zhou
>> > It's a nodetool command in very recent versions, or a jmx method in older >> > versions >> > >> > >> > -- >> > Jeff Jirsa >> > >> > >> >> On Aug 31, 2017, at 11:04 PM, qf zhou <zhouqf2...@gmail.co

Re: old big tombstone data file occupy much disk space

2017-09-01 Thread Nicolas Guyomar
-- >> > Jeff Jirsa >> > >> > >> >> On Aug 31, 2017, at 11:04 PM, qf zhou <zhouqf2...@gmail.com> wrote: >> >> >> >> I am using a cluster with 3 nodes and the cassandra version is >> 3.0.9. I have used it about 6 months. Now e

Re: old big tombstone data file occupy much disk space

2017-09-01 Thread qf zhou
sion is 3.0.9. I > >> have used it about 6 months. Now each node has about 1.5T data in the disk. > >> I found some sstables file are over 300G. Using the sstablemetadata > >> command, I found it: Estimated droppable tombstones: 0.9622972799707109. > >> It i

Re: old big tombstone data file occupy much disk space

2017-09-01 Thread Nicolas Guyomar
.0.9. I have used it about 6 months. Now each node has about 1.5T data in > the disk. > >> I found some sstables file are over 300G. Using the sstablemetadata > command, I found it: Estimated droppable tombstones: 0.9622972799707109. > >> It is obvious that too much tombston

Re: old big tombstone data file occupy much disk space

2017-09-01 Thread qf zhou
gt;> have used it about 6 months. Now each node has about 1.5T data in the disk. >> I found some sstables file are over 300G. Using the sstablemetadata >> command, I found it: Estimated droppable tombstones: 0.9622972799707109. >> It is obvious that too much tombstone dat

Re: old big tombstone data file occupy much disk space

2017-09-01 Thread Jeff Jirsa
It is obvious that too much tombstone data exists. > The default_time_to_live = 864(100 days) and gc_grace_seconds = > 432000(5 days). Using nodetool compactionstats, I found the some compaction > processes exists. > So I really want to know how to clear tombstone data ? otherwise th

old big tombstone data file occupy much disk space

2017-09-01 Thread qf zhou
. It is obvious that too much tombstone data exists. The default_time_to_live = 864(100 days) and gc_grace_seconds = 432000(5 days). Using nodetool compactionstats, I found the some compaction processes exists. So I really want to know how to clear tombstone data ? otherwise the disk

tombstone limit reaches 100K cells

2017-06-21 Thread web master
Accoding to http://docs.datastax.com/en/cql/3.1/cql/ddl/ddl_when_use_ index_c.html#concept_ds_sgh_yzz_zj__upDatIndx > Cassandra stores tombstones in the index until the tombstone limit reaches 100K cells. After exceeding the tombstone limit, the query that uses the indexed value will fail.

RE: Some questions to updating and tombstone

2016-11-17 Thread Lu, Boying
Very appreciate to all of you, I’ll study the blog. From: Alain RODRIGUEZ [mailto:arodr...@gmail.com] Sent: 2016年11月16日 23:26 To: user@cassandra.apache.org Cc: Fabrice Facorat Subject: Re: Some questions to updating and tombstone Hi Boying, Old value is not tombstone, but remains until

Re: Some questions to updating and tombstone

2016-11-16 Thread Alain RODRIGUEZ
Hi Boying, Old value is not tombstone, but remains until compaction Be careful, the above is generally true but not necessary. Tombstones can actually be generated while using update in some corner cases. Using collections or prepared statements. I wrote a detailed blog post about deletes

Re: Some questions to updating and tombstone

2016-11-16 Thread Shalom Sagges
t tombstones, don't generate them ;) > > More seriously, tombstones are generated when: > - doing a DELETE > - TTL expiration > - set a column to NULL > > However tombstones are an issue only if for the same value, you have many > tombstones (i.e you keep overwriting the same values wit

Re: Some questions to updating and tombstone

2016-11-15 Thread Fabrice Facorat
and tombstones). Having 1 tombstone for 1 value is not an issue, having 1000 tombstone for 1 value is a problem. Do really your use case overwrite data with DELETE or NULL ? So that's why what you may want to know is how many tombstones you have on average when reading a value. This is available

RE: Some questions to updating and tombstone

2016-11-15 Thread Lu, Boying
to updating and tombstone Hi Boying, I agree with Vladimir.If compaction is not compacting the two sstables with updates soon, disk space issues will be wasted. For example, if the updates are not closer in time, first update might be in a big table by the time second update is being written

Re: Some questions to updating and tombstone

2016-11-14 Thread Anuj Wadehra
write new value with new time stamp. Old value is not tombstone, but remains until compaction. gc_grace_period is not related to this. Best regards, Vladimir Yudovin, Winguzone - Hosted Cloud Cassandra Launch your cluster in minutes. On Mon, 14 Nov 2016 03:02:21 -0500Lu, Boying &

Re: Some questions to updating and tombstone

2016-11-14 Thread Vladimir Yudovin
Hi Boying, UPDATE write new value with new time stamp. Old value is not tombstone, but remains until compaction. gc_grace_period is not related to this. Best regards, Vladimir Yudovin, Winguzone - Hosted Cloud Cassandra Launch your cluster in minutes. On Mon, 14 Nov 2016 03:02

Some questions to updating and tombstone

2016-11-14 Thread Lu, Boying
Hi, All, Will the Cassandra generates a new tombstone when updating a column by using CQL update statement? And is there any way to get the number of tombstones of a column family since we want to void generating too many tombstones within gc_grace_period? Thanks Boying

Re: Secondary index tombstone limit

2016-11-07 Thread anil ahlawat
unsubsribe Sent from Yahoo Mail on Android On Tue, 8 Nov, 2016 at 2:11 pm, Oleg Krayushkin<allight...@gmail.com> wrote: Hi, could you please clarify: 100k tombstone limit for SE is per CF, cf-node, original sstable or (very unlikely) partition? Thanks!-- Oleg Krayushkin

Secondary index tombstone limit

2016-11-07 Thread Oleg Krayushkin
Hi, could you please clarify: 100k tombstone limit for SE is per CF, cf-node, original sstable or (very unlikely) partition? Thanks! -- Oleg Krayushkin

Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Alexander Dejanovski
be more specific on how you are running repair ? What's the >>>>> precise command line for that, does it run on several nodes at the same >>>>> time, etc... >>>>> What is your gc_grace_seconds ? >>>>> Do you see errors in your logs that would be

Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Atul Saroha
>>>> What is your gc_grace_seconds ? >>>> Do you see errors in your logs that would be linked to repairs >>>> (Validation failure or failure to create a merkle tree)? >>>> >>>> You seem to mention a single node that went down but say the

Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Atul Saroha
ween the node that went down and the >>> fact that deleted data comes back to life ? >>> What is your strategy for cyclic maintenance repair (schedule, command >>> line or tool, etc...) ? >>> >>> Thanks, >>> >>> On Thu, Sep 29, 2016 at 1

Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Alexander Dejanovski
anks, >> >> On Thu, Sep 29, 2016 at 10:40 AM Atul Saroha <atul.sar...@snapdeal.com> >> wrote: >> >>> Hi, >>> >>> We have seen a weird behaviour in cassandra 3.6. >>> Once our node was went down more than 10 hrs. After that, w

Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Atul Saroha
, Sep 29, 2016 at 10:40 AM Atul Saroha <atul.sar...@snapdeal.com> > wrote: > >> Hi, >> >> We have seen a weird behaviour in cassandra 3.6. >> Once our node was went down more than 10 hrs. After that, we had ran >> Nodetool repair multiple times. But tombstone

Re: [cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Alexander Dejanovski
, command line or tool, etc...) ? Thanks, On Thu, Sep 29, 2016 at 10:40 AM Atul Saroha <atul.sar...@snapdeal.com> wrote: > Hi, > > We have seen a weird behaviour in cassandra 3.6. > Once our node was went down more than 10 hrs. After that, we had ran > Nodetool repair multiple

[cassandra 3.6.] Nodetool Repair + tombstone behaviour

2016-09-29 Thread Atul Saroha
Hi, We have seen a weird behaviour in cassandra 3.6. Once our node was went down more than 10 hrs. After that, we had ran Nodetool repair multiple times. But tombstone are not getting sync properly over the cluster. On day- today basis, on expiry of every grace period, deleted records start

lots of tombstone after compaction

2015-12-07 Thread Kai Wang
I bulkloaded a few tables using CQLSStableWrite/sstableloader. The data are large amount of wide rows with lots of null's. It takes one day or two for the compaction to complete. sstable count is at single digit. Maximum partition size is ~50M and mean size is ~5M. However I am seeing frequent

Re: lots of tombstone after compaction

2015-12-07 Thread Kai Wang
10505 > > There are buggy versions of cassandra that will multiple tombstones during > compaction. 2.1.12 SHOULD correct that, if you’re on 2.1. > > > > From: Kai Wang > Reply-To: "user@cassandra.apache.org" > Date: Monday, December 7, 2015 at 3:46 PM > To: "us

Re: lots of tombstone after compaction

2015-12-07 Thread Robert Wille
The nulls in the original data created the tombstones. They won’t go away until gc_grace_seconds have passed (default is 10 days). On Dec 7, 2015, at 4:46 PM, Kai Wang wrote: > I bulkloaded a few tables using CQLSStableWrite/sstableloader. The data are > large amount of wide

Re: lots of tombstone after compaction

2015-12-07 Thread Jeff Jirsa
;user@cassandra.apache.org" Date: Monday, December 7, 2015 at 3:46 PM To: "user@cassandra.apache.org" Subject: lots of tombstone after compaction I bulkloaded a few tables using CQLSStableWrite/sstableloader. The data are large amount of wide rows with lots of null's. It takes one day or two

Manual Tombstone Removal for Leveled Compaction

2015-08-01 Thread Mir Tanvir Hossain
Hello, I have a question about the tombstone removal process for leveled compaction strategy. I am migrating a lot of text data from a cassandra column family to elastic search. The column family uses leveled compaction strategy. As part of the migration, I am deleting the migrated rows from

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-23 Thread Andrei Ivanov
in understanding how tombstone threshold is implemented. And ticket also says that running major compaction weekly is an alternative. I actually want to understand if I run major compaction on a cf with 500gb of data and a single giant file is created. Do you see any problems with Cassandra processing

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-23 Thread Anuj Wadehra
Great !!! Thanks Andrei !!! Thats the answer I was looking for :) Thanks Anuj Wadehra Sent from Yahoo Mail on Android From:Andrei Ivanov aiva...@iponweb.net Date:Thu, 23 Apr, 2015 at 11:57 pm Subject:Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists Just

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-22 Thread Anuj Wadehra
Thanks Robert!! The JIRA was very helpful in understanding how tombstone threshold is implemented. And ticket also says that running major compaction weekly is an alternative. I actually want to understand if I run major compaction on a cf with 500gb of data and a single giant file is created

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-17 Thread Robert Coli
On Tue, Apr 14, 2015 at 8:29 PM, Anuj Wadehra anujw_2...@yahoo.co.in wrote: By automatic tombstone compaction, I am referring to tombstone_threshold sub property under compaction strategy in CQL. It is 0.2 by default. So what I understand from the Datastax documentation is that even

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-16 Thread Anuj Wadehra
Hi Robert, Any comments or suggestions ? Thanks Anuj Wadehra Sent from Yahoo Mail on Android From:Anuj Wadehra anujw_2...@yahoo.co.in Date:Wed, 15 Apr, 2015 at 8:59 am Subject:Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists Hi Robert, By automatic

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-14 Thread Anuj Wadehra
Hi Robert, By automatic tombstone compaction, I am referring to tombstone_threshold sub property under compaction strategy in CQL. It is 0.2 by default. So what I understand from the Datastax documentation is that even if a sstable does not find sstables of similar size (STCS) , an automatic

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-13 Thread Robert Coli
On Mon, Apr 13, 2015 at 10:52 AM, Anuj Wadehra anujw_2...@yahoo.co.in wrote: Any comments on side effects of Major compaction especially when sstable generated is 100+ GB? I have no idea how this interacts with the automatic compaction stuff; if you find out, let us know? But if you want to

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-13 Thread Anuj Wadehra
Any comments on side effects of Major compaction especially when sstable generated is 100+ GB?  After Cassandra 1.2 , automated tombstone compaction occurs even on a single sstable if tombstone percentage increases the tombstone_threshold sub property specified in compaction strategy. So

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-13 Thread Rahul Neelakantan
Rob, Does that mean once you split it back into small ones, automatic compaction a will continue to happen on a more frequent basis now that it's no longer a single large monolith? Rahul On Apr 13, 2015, at 3:23 PM, Robert Coli rc...@eventbrite.com wrote: On Mon, Apr 13, 2015 at 10:52 AM,

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-13 Thread Robert Coli
On Mon, Apr 13, 2015 at 12:26 PM, Rahul Neelakantan ra...@rahul.be wrote: Does that mean once you split it back into small ones, automatic compaction a will continue to happen on a more frequent basis now that it's no longer a single large monolith? That's what the word size tiered means in

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-12 Thread Sebastian Estevez
are not impacted, we were left with no option but to run major compaction to ensure that thousands of tiny sstables are compacted. Queries: Does major compaction has any drawback after automatic tombstone compaction got implemented in 1.2 via tombstone_threshold sub-property(CASSANDRA-3442

Re: Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-12 Thread Anuj Wadehra
drawback after automatic tombstone compaction got implemented in 1.2 via tombstone_threshold sub-property(CASSANDRA-3442)? I understand that the huge SSTable created after major compaction wont be compacted with new data any time soon but is that a problem if purged data is removed via automatic

Drawbacks of Major Compaction now that Automatic Tombstone Compaction Exists

2015-04-12 Thread Anuj Wadehra
are compacted. Queries: Does major compaction has any drawback after automatic tombstone compaction got implemented in 1.2 via tombstone_threshold sub-property(CASSANDRA-3442)? I understand that the huge SSTable created after major compaction wont be compacted with new data any time soon

Re: Tombstone gc after gc grace seconds

2015-02-03 Thread Alain RODRIGUEZ
Feel of the data that we have is - 8000 rowkeys per day and columns are added throughout the day. 300K columns on an average per rowKey. *From:* Alain RODRIGUEZ [mailto:arodr...@gmail.com] *Sent:* Friday, January 30, 2015 4:26 AM *To:* user@cassandra.apache.org *Subject:* Re: Tombstone

RE: Tombstone gc after gc grace seconds

2015-01-30 Thread Ravi Agrawal
and columns are added throughout the day. 300K columns on an average per rowKey. From: Alain RODRIGUEZ [mailto:arodr...@gmail.com] Sent: Friday, January 30, 2015 4:26 AM To: user@cassandra.apache.org Subject: Re: Tombstone gc after gc grace seconds The point is that all the parts or fragments

Re: Tombstone gc after gc grace seconds

2015-01-30 Thread Alain RODRIGUEZ
rows / time series) your only way to get rid of tombstone is a major compaction. That's how I understand this. Hope this help, C*heers, Alain 2015-01-30 1:29 GMT+01:00 Mohammed Guller moham...@glassbeam.com: Ravi - It may help. What version are you running? Do you know if minor

RE: Tombstone gc after gc grace seconds

2015-01-29 Thread Ravi Agrawal
. unchecked_tombstone_compaction - True enables more aggressive than normal tombstone compactions. A single SSTable tombstone compaction runs without checking the likelihood of success. Cassandra 2.0.9 and later. Could I use these to get what I want? Problem I am encountering is even long

RE: Tombstone gc after gc grace seconds

2015-01-29 Thread Mohammed Guller
To: user@cassandra.apache.org Subject: RE: Tombstone gc after gc grace seconds Hi, I saw there are 2 more interesting parameters – a. tombstone_threshold - A ratio of garbage-collectable tombstones to all contained columns, which if exceeded by the SSTable triggers compaction (with no other

Re: Tombstone gc after gc grace seconds

2015-01-26 Thread Eric Stevens
My understanding is consistent with Alain's, there's no way to force a tombstone-only compaction, your only option is major compaction. If you're using size tiered, that comes with its own drawbacks. I wonder if there's a technical limitation that prevents introducing a shadowed data cleanup

Re: Tombstone gc after gc grace seconds

2015-01-26 Thread Alain RODRIGUEZ
...@clearpoolgroup.com: Hi, I want to trigger just tombstone compaction after gc grace seconds is completed not nodetool compact keyspace column family. Anyway I can do that? Thanks

Re: get partition key from tombstone warnings?

2015-01-22 Thread Paulo Ricardo Motta Gomes
Yep, you may register and log into the Apache JIRA and click Vote for this issue, in the upper right-side of the ticket. On Wed, Jan 21, 2015 at 11:30 PM, Ian Rose ianr...@fullstory.com wrote: Ah, thanks for the pointer Philip. Is there any kind of formal way to vote up issues? I'm assuming

Re: get partition key from tombstone warnings?

2015-01-22 Thread Philip Thompson
Ian, Leaving a comment explaining your situation and how, as an operator of a Cassandra Cluster, this would be valuable, would probably help most. On Thu, Jan 22, 2015 at 6:06 AM, Paulo Ricardo Motta Gomes paulo.mo...@chaordicsystems.com wrote: Yep, you may register and log into the Apache

Tombstone gc after gc grace seconds

2015-01-22 Thread Ravi Agrawal
Hi, I want to trigger just tombstone compaction after gc grace seconds is completed not nodetool compact keyspace column family. Anyway I can do that? Thanks

Re: get partition key from tombstone warnings?

2015-01-21 Thread Philip Thompson
There is an open ticket for this improvement at https://issues.apache.org/jira/browse/CASSANDRA-8561 On Wed, Jan 21, 2015 at 4:55 PM, Ian Rose ianr...@fullstory.com wrote: When I see a warning like Read 9 live and 5769 tombstoned cells in ... etc is there a way for me to see the partition key

get partition key from tombstone warnings?

2015-01-21 Thread Ian Rose
When I see a warning like Read 9 live and 5769 tombstoned cells in ... etc is there a way for me to see the partition key that this query was operating on? The description in the original JIRA ticket ( https://issues.apache.org/jira/browse/CASSANDRA-6042) reads as though exposing this information

Re: get partition key from tombstone warnings?

2015-01-21 Thread Ian Rose
Ah, thanks for the pointer Philip. Is there any kind of formal way to vote up issues? I'm assuming that adding a comment of +1 or the like is more likely to be *counter*productive. - Ian On Wed, Jan 21, 2015 at 5:02 PM, Philip Thompson philip.thomp...@datastax.com wrote: There is an open

Re: nodetool compact cannot remove tombstone in system keyspace

2015-01-13 Thread Xu Zhongxing
columns was requested, slices=[-], delInfo={deletedAt=-9223372036854775808, localDeletion=2147483647} I run the command: nodetool compact system But the tombstone number does not decrease. I still see the warnings with the exact number of tombstones. Why is this happening? What should I do

Re: nodetool compact cannot remove tombstone in system keyspace

2015-01-13 Thread Rahul Neelakantan
tombstone_warn_threshold). 2147449199 columns was requested, slices=[-], delInfo={deletedAt=-9223372036854775808, localDeletion=2147483647} I run the command: nodetool compact system But the tombstone number does not decrease. I still see the warnings with the exact number of tombstones. Why

  1   2   >