If there were no other messages about anti-compaction similar to:
>
> SSTable YYY (ranges) will be anticompacted on range [range]


Then no anti-compaction needed to occur and yes, it was not the cause.

On 5 April 2018 at 13:52, Dmitry Simonov <dimmobor...@gmail.com> wrote:

> Hi, Evelyn!
>
> I've found the following messages:
>
> INFO RepairRunnable.java Starting repair command #41, repairing keyspace
> XXX with repair options (parallelism: parallel, primary range: false,
> incremental: false, job threads: 1, ColumnFamilies: [YYY], dataCenters: [],
> hosts: [], # of ranges: 768)
> INFO CompactionExecutor:6 CompactionManager.java Starting anticompaction
> for XXX.YYY on 5132/5846 sstables
>
> After that many similar messages go:
> SSTable BigTableReader(path='/mnt/cassandra/data/XXX/YYY-
> 4c12fd9029e611e8810ac73ddacb37d1/lb-12688-big-Data.db') fully contained
> in range (-9223372036854775808,-9223372036854775808], mutating repairedAt
> instead of anticompacting
>
> Does it means that anti-compaction is not the cause?
>
> 2018-04-05 18:01 GMT+05:00 Evelyn Smith <u5015...@gmail.com>:
>
>> It might not be what cause it here. But check your logs for
>> anti-compactions.
>>
>>
>> On 5 Apr 2018, at 8:35 pm, Dmitry Simonov <dimmobor...@gmail.com> wrote:
>>
>> Thank you!
>> I'll check this out.
>>
>> 2018-04-05 15:00 GMT+05:00 Alexander Dejanovski <a...@thelastpickle.com>:
>>
>>> 40 pending compactions is pretty high and you should have way less than
>>> that most of the time, otherwise it means that compaction is not keeping up
>>> with your write rate.
>>>
>>> If you indeed have SSDs for data storage, increase your compaction
>>> throughput to 100 or 200 (depending on how the CPUs handle the load). You
>>> can experiment with compaction throughput using : nodetool
>>> setcompactionthroughput 100
>>>
>>> You can raise the number of concurrent compactors as well and set it to
>>> a value between 4 and 6 if you have at least 8 cores and CPUs aren't
>>> overwhelmed.
>>>
>>> I'm not sure why you ended up with only one node having 6k SSTables and
>>> not the others, but you should apply the above changes so that you can
>>> lower the number of pending compactions and see if it prevents the issue
>>> from happening again.
>>>
>>> Cheers,
>>>
>>>
>>> On Thu, Apr 5, 2018 at 11:33 AM Dmitry Simonov <dimmobor...@gmail.com>
>>> wrote:
>>>
>>>> Hi, Alexander!
>>>>
>>>> SizeTieredCompactionStrategy is used for all CFs in problematic
>>>> keyspace.
>>>> Current compaction throughput is 16 MB/s (default value).
>>>>
>>>> We always have about 40 pending and 2 active "CompactionExecutor" tasks
>>>> in "tpstats".
>>>> Mostly because of another (bigger) keyspace in this cluster.
>>>> But the situation is the same on each node.
>>>>
>>>> According to "nodetool compactionhistory", compactions on this CF run
>>>> (sometimes several times per day, sometimes one time per day, the last run
>>>> was yesterday).
>>>> We run "repair -full" regulary for this keyspace (every 24 hours on
>>>> each node), because gc_grace_seconds is set to 24 hours.
>>>>
>>>> Should we consider increasing compaction throughput and
>>>> "concurrent_compactors" (as recommended for SSDs) to keep
>>>> "CompactionExecutor" pending tasks low?
>>>>
>>>> 2018-04-05 14:09 GMT+05:00 Alexander Dejanovski <a...@thelastpickle.com
>>>> >:
>>>>
>>>>> Hi Dmitry,
>>>>>
>>>>> could you tell us which compaction strategy that table is currently
>>>>> using ?
>>>>> Also, what is the compaction max throughput and is auto-compaction
>>>>> correctly enabled on that node ?
>>>>>
>>>>> Did you recently run repair ?
>>>>>
>>>>> Thanks,
>>>>>
>>>>> On Thu, Apr 5, 2018 at 10:53 AM Dmitry Simonov <dimmobor...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hello!
>>>>>>
>>>>>> Could you please give some ideas on the following problem?
>>>>>>
>>>>>> We have a cluster with 3 nodes, running Cassandra 2.2.11.
>>>>>>
>>>>>> We've recently discovered high CPU usage on one cluster node, after
>>>>>> some investigation we found that number of sstables for one CF on it is
>>>>>> very big: 5800 sstables, on other nodes: 3 sstable.
>>>>>>
>>>>>> Data size in this keyspace was not very big ~100-200Mb per node.
>>>>>>
>>>>>> There is no such problem with other CFs of that keyspace.
>>>>>>
>>>>>> nodetool compact solved the issue as a quick-fix.
>>>>>>
>>>>>> But I'm wondering, what was the cause? How prevent it from repeating?
>>>>>>
>>>>>> --
>>>>>> Best Regards,
>>>>>> Dmitry Simonov
>>>>>>
>>>>> --
>>>>> -----------------
>>>>> Alexander Dejanovski
>>>>> France
>>>>> @alexanderdeja
>>>>>
>>>>> Consultant
>>>>> Apache Cassandra Consulting
>>>>> http://www.thelastpickle.com
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Best Regards,
>>>> Dmitry Simonov
>>>>
>>> --
>>> -----------------
>>> Alexander Dejanovski
>>> France
>>> @alexanderdeja
>>>
>>> Consultant
>>> Apache Cassandra Consulting
>>> http://www.thelastpickle.com
>>>
>>
>>
>>
>> --
>> Best Regards,
>> Dmitry Simonov
>>
>>
>>
>
>
> --
> Best Regards,
> Dmitry Simonov
>

Reply via email to