[
https://issues.apache.org/jira/browse/CASSANDRA-8140?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Davide updated CASSANDRA-8140:
------------------------------
Description:
Hi there,
I'm on cassandra 2.1 since then I figured out that in some circumstances (I
can't find a way to reproduce them constantly) minor compactions and full
compactions had no effects.
We are on a cluster composed of 5 nodes with around 500gb of data, no deletions
around 1.5k updates/s and same on reads.
After a repair I saw that a couple of nodes were `slow`, I investigate further
and I found that on these two nodes the number of sstable were around 20.000+ !
We use STC.
So with node tool I triggered a full compaction, It took less than I minute
(with nothing in the logs) and of course the number of sstable didn't go down.
Then I drained the node, and I ran again with `nodetool compact`, at that point
the number of sstables went down to less than 10.
I tough was a strange spot problem. However after a week I noticed that one
node had ~100 sstabels where others just 8-10.
I ran again the compaction (It last less than a minute with nothing in logs)
and didn't change anything. I drained it and restarted then compacted and took
several hours to get it back close to 2/3 sstables.
What could be? We never incurred this behavior before.
Here informations about the table:
{code}
CREATE TABLE xyz (
ppk text PRIMARY KEY,
.. ten more columns...
) WITH bloom_filter_fp_chance = 0.01
AND caching = '{"keys":"ALL", "rows_per_partition":"NONE"}'
AND comment = ''
AND compaction = {'min_threshold': '4', 'cold_reads_to_omit': '0.0',
'class': 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy',
'max_threshold': '32'}
AND compression = {'sstable_compression':
'org.apache.cassandra.io.compress.SnappyCompressor'}
AND dclocal_read_repair_chance = 0.0
AND default_time_to_live = 0
AND gc_grace_seconds = 864000
AND max_index_interval = 2048
AND memtable_flush_period_in_ms = 0
AND min_index_interval = 128
AND read_repair_chance = 0.0
AND speculative_retry = '99.0PERCENTILE';
{code}
Here the current cf stats:
{code}
SSTable count: 11
Space used (live), bytes: 118007220865
Space used (total), bytes: 118007220865
Space used by snapshots (total), bytes: 170591332257
SSTable Compression Ratio: 0.3643916626015517
Memtable cell count: 920306
Memtable data size, bytes: 70034097
Memtable switch count: 25
Local read count: 5358772
Local read latency: 54.621 ms
Local write count: 4715106
Local write latency: 0.069 ms
Pending flushes: 0
Bloom filter false positives: 53757
Bloom filter false ratio: 0.04103
Bloom filter space used, bytes: 220634056
Compacted partition minimum bytes: 18
Compacted partition maximum bytes: 61214
Compacted partition mean bytes: 1935
Average live cells per slice (last five minutes):
0.8139232271871242
Average tombstones per slice (last five minutes):
0.5493417148555677
{code}
Is there anything else that I can provide?
Thanks!
DD
was:
Hi there,
I'm on cassandra 2.1 since then I figure out that in some circumstances (I
can't find a way to reproduce them constantly) minor compactions and full
compactions takes no effects.
So we are on a cluster composed of 5 nodes with around 500gb of data, no
deletions around 1.5k updates/s and same on reads.
After a repair I saw that a couple of nodes were `slow`, I investigate further
and I found that on these two nodes the number of sstable were around 20.000+ !
We use STC.
So with node tool I triggered a full compaction, It took less than I minute
(with noting in the logs) and of course the number of sstable didn't go down.
Then I drained the node, and I ran again the `nodetool compact`, at that point
the number of sstables went down to less than 10.
I tough was a strange spot problem. However after a week I noticed that one
node had ~100 sstabels where others just 8-10.
I ran again the compaction (It last less than a minute with nothing in logs)
and didn't change anything. I drained it and restarted then compacted and took
several hours to get it back close to 2/3 sstables.
What could be? We never incurred this behavior before.
Here informations about the table:
{code}
CREATE TABLE xyz (
ppk text PRIMARY KEY,
.. ten more columns...
) WITH bloom_filter_fp_chance = 0.01
AND caching = '{"keys":"ALL", "rows_per_partition":"NONE"}'
AND comment = ''
AND compaction = {'min_threshold': '4', 'cold_reads_to_omit': '0.0',
'class': 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy',
'max_threshold': '32'}
AND compression = {'sstable_compression':
'org.apache.cassandra.io.compress.SnappyCompressor'}
AND dclocal_read_repair_chance = 0.0
AND default_time_to_live = 0
AND gc_grace_seconds = 864000
AND max_index_interval = 2048
AND memtable_flush_period_in_ms = 0
AND min_index_interval = 128
AND read_repair_chance = 0.0
AND speculative_retry = '99.0PERCENTILE';
{code}
Here the current cf stats:
{code}
SSTable count: 11
Space used (live), bytes: 118007220865
Space used (total), bytes: 118007220865
Space used by snapshots (total), bytes: 170591332257
SSTable Compression Ratio: 0.3643916626015517
Memtable cell count: 920306
Memtable data size, bytes: 70034097
Memtable switch count: 25
Local read count: 5358772
Local read latency: 54.621 ms
Local write count: 4715106
Local write latency: 0.069 ms
Pending flushes: 0
Bloom filter false positives: 53757
Bloom filter false ratio: 0.04103
Bloom filter space used, bytes: 220634056
Compacted partition minimum bytes: 18
Compacted partition maximum bytes: 61214
Compacted partition mean bytes: 1935
Average live cells per slice (last five minutes):
0.8139232271871242
Average tombstones per slice (last five minutes):
0.5493417148555677
{code}
Is there anything else that I can provide?
Thanks!
DD
> Compaction has no effects
> -------------------------
>
> Key: CASSANDRA-8140
> URL: https://issues.apache.org/jira/browse/CASSANDRA-8140
> Project: Cassandra
> Issue Type: Bug
> Components: Core
> Reporter: Davide
>
> Hi there,
> I'm on cassandra 2.1 since then I figured out that in some circumstances (I
> can't find a way to reproduce them constantly) minor compactions and full
> compactions had no effects.
> We are on a cluster composed of 5 nodes with around 500gb of data, no
> deletions around 1.5k updates/s and same on reads.
> After a repair I saw that a couple of nodes were `slow`, I investigate
> further and I found that on these two nodes the number of sstable were around
> 20.000+ ! We use STC.
> So with node tool I triggered a full compaction, It took less than I minute
> (with nothing in the logs) and of course the number of sstable didn't go down.
> Then I drained the node, and I ran again with `nodetool compact`, at that
> point the number of sstables went down to less than 10.
> I tough was a strange spot problem. However after a week I noticed that one
> node had ~100 sstabels where others just 8-10.
> I ran again the compaction (It last less than a minute with nothing in logs)
> and didn't change anything. I drained it and restarted then compacted and
> took several hours to get it back close to 2/3 sstables.
> What could be? We never incurred this behavior before.
> Here informations about the table:
> {code}
> CREATE TABLE xyz (
> ppk text PRIMARY KEY,
> .. ten more columns...
> ) WITH bloom_filter_fp_chance = 0.01
> AND caching = '{"keys":"ALL", "rows_per_partition":"NONE"}'
> AND comment = ''
> AND compaction = {'min_threshold': '4', 'cold_reads_to_omit': '0.0',
> 'class': 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy',
> 'max_threshold': '32'}
> AND compression = {'sstable_compression':
> 'org.apache.cassandra.io.compress.SnappyCompressor'}
> AND dclocal_read_repair_chance = 0.0
> AND default_time_to_live = 0
> AND gc_grace_seconds = 864000
> AND max_index_interval = 2048
> AND memtable_flush_period_in_ms = 0
> AND min_index_interval = 128
> AND read_repair_chance = 0.0
> AND speculative_retry = '99.0PERCENTILE';
> {code}
> Here the current cf stats:
> {code}
> SSTable count: 11
> Space used (live), bytes: 118007220865
> Space used (total), bytes: 118007220865
> Space used by snapshots (total), bytes: 170591332257
> SSTable Compression Ratio: 0.3643916626015517
> Memtable cell count: 920306
> Memtable data size, bytes: 70034097
> Memtable switch count: 25
> Local read count: 5358772
> Local read latency: 54.621 ms
> Local write count: 4715106
> Local write latency: 0.069 ms
> Pending flushes: 0
> Bloom filter false positives: 53757
> Bloom filter false ratio: 0.04103
> Bloom filter space used, bytes: 220634056
> Compacted partition minimum bytes: 18
> Compacted partition maximum bytes: 61214
> Compacted partition mean bytes: 1935
> Average live cells per slice (last five minutes):
> 0.8139232271871242
> Average tombstones per slice (last five minutes):
> 0.5493417148555677
> {code}
> Is there anything else that I can provide?
> Thanks!
> DD
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)