Thank you! I'll check this out. 2018-04-05 15:00 GMT+05:00 Alexander Dejanovski <a...@thelastpickle.com>:
> 40 pending compactions is pretty high and you should have way less than > that most of the time, otherwise it means that compaction is not keeping up > with your write rate. > > If you indeed have SSDs for data storage, increase your compaction > throughput to 100 or 200 (depending on how the CPUs handle the load). You > can experiment with compaction throughput using : nodetool > setcompactionthroughput 100 > > You can raise the number of concurrent compactors as well and set it to a > value between 4 and 6 if you have at least 8 cores and CPUs aren't > overwhelmed. > > I'm not sure why you ended up with only one node having 6k SSTables and > not the others, but you should apply the above changes so that you can > lower the number of pending compactions and see if it prevents the issue > from happening again. > > Cheers, > > > On Thu, Apr 5, 2018 at 11:33 AM Dmitry Simonov <dimmobor...@gmail.com> > wrote: > >> Hi, Alexander! >> >> SizeTieredCompactionStrategy is used for all CFs in problematic keyspace. >> Current compaction throughput is 16 MB/s (default value). >> >> We always have about 40 pending and 2 active "CompactionExecutor" tasks >> in "tpstats". >> Mostly because of another (bigger) keyspace in this cluster. >> But the situation is the same on each node. >> >> According to "nodetool compactionhistory", compactions on this CF run >> (sometimes several times per day, sometimes one time per day, the last run >> was yesterday). >> We run "repair -full" regulary for this keyspace (every 24 hours on each >> node), because gc_grace_seconds is set to 24 hours. >> >> Should we consider increasing compaction throughput and >> "concurrent_compactors" (as recommended for SSDs) to keep >> "CompactionExecutor" pending tasks low? >> >> 2018-04-05 14:09 GMT+05:00 Alexander Dejanovski <a...@thelastpickle.com>: >> >>> Hi Dmitry, >>> >>> could you tell us which compaction strategy that table is currently >>> using ? >>> Also, what is the compaction max throughput and is auto-compaction >>> correctly enabled on that node ? >>> >>> Did you recently run repair ? >>> >>> Thanks, >>> >>> On Thu, Apr 5, 2018 at 10:53 AM Dmitry Simonov <dimmobor...@gmail.com> >>> wrote: >>> >>>> Hello! >>>> >>>> Could you please give some ideas on the following problem? >>>> >>>> We have a cluster with 3 nodes, running Cassandra 2.2.11. >>>> >>>> We've recently discovered high CPU usage on one cluster node, after >>>> some investigation we found that number of sstables for one CF on it is >>>> very big: 5800 sstables, on other nodes: 3 sstable. >>>> >>>> Data size in this keyspace was not very big ~100-200Mb per node. >>>> >>>> There is no such problem with other CFs of that keyspace. >>>> >>>> nodetool compact solved the issue as a quick-fix. >>>> >>>> But I'm wondering, what was the cause? How prevent it from repeating? >>>> >>>> -- >>>> Best Regards, >>>> Dmitry Simonov >>>> >>> -- >>> ----------------- >>> Alexander Dejanovski >>> France >>> @alexanderdeja >>> >>> Consultant >>> Apache Cassandra Consulting >>> http://www.thelastpickle.com >>> >> >> >> >> -- >> Best Regards, >> Dmitry Simonov >> > -- > ----------------- > Alexander Dejanovski > France > @alexanderdeja > > Consultant > Apache Cassandra Consulting > http://www.thelastpickle.com > -- Best Regards, Dmitry Simonov