I have disabled autocompaction and stop it on highload node. Freezes all nodes sequentially, 2-6 simultaneously.
Heap is 8Gb. gc_grace is 86400 All sstables is about 200-300 Mb. $ nodetool compactionstats pending tasks: 14 $ dstat -lvnr 10 ---load-avg--- ---procs--- ------memory-usage----- ---paging-- -dsk/total- ---system-- ----total-cpu-usage---- -net/total- --io/total- 1m 5m 15m |run blk new| used buff cach free| in out | read writ| int csw |usr sys idl wai hiq siq| recv send| read writ 29.4 28.6 23.5|0.0 0 1.2|11.3G 190M 17.6G 407M| 0 0 |7507k 7330k| 13k 40k| 11 1 88 0 0 0| 0 0 |96.5 64.6 29.3 28.6 23.5| 29 0 0.9|11.3G 190M 17.6G 408M| 0 0 | 0 189k|9822 2319 | 99 0 0 0 0 0| 138k 120k| 0 4.30 29.4 28.6 23.6| 30 0 2.0|11.3G 190M 17.6G 408M| 0 0 | 0 26k|8689 2189 |100 0 0 0 0 0| 139k 120k| 0 2.70 29.4 28.7 23.6| 29 0 3.0|11.3G 190M 17.6G 408M| 0 0 | 0 20k|8722 1846 | 99 0 0 0 0 0| 136k 120k| 0 1.50 ^C JvmTop 0.8.0 alpha - 15:20:37, amd64, 16 cpus, Linux 3.14.44-3, load avg 28.09 http://code.google.com/p/jvmtop PID 32505: org.apache.cassandra.service.CassandraDaemon ARGS: VMARGS: -ea -javaagent:/usr/share/cassandra/lib/jamm-0.3.0.jar -XX:+CMSCl[...] VM: Oracle Corporation Java HotSpot(TM) 64-Bit Server VM 1.8.0_65 UP: 8:31m #THR: 334 #THRPEAK: 437 #THRCREATED: 4694 USER: cassandra GC-Time: 0: 8m #GC-Runs: 6378 #TotalLoadedClasses: 5926 CPU: 97.96% GC: 0.00% HEAP:6049m /7540m NONHEAP: 82m / n/a TID NAME STATE CPU TOTALCPU BLOCKEDBY 447 SharedPool-Worker-45 RUNNABLE 60.47% 1.03% 343 SharedPool-Worker-2 RUNNABLE 56.46% 3.07% 349 SharedPool-Worker-8 RUNNABLE 56.43% 1.61% 456 SharedPool-Worker-25 RUNNABLE 55.25% 1.06% 483 SharedPool-Worker-40 RUNNABLE 53.06% 1.04% 475 SharedPool-Worker-53 RUNNABLE 52.31% 1.03% 464 SharedPool-Worker-20 RUNNABLE 52.00% 1.11% 577 SharedPool-Worker-71 RUNNABLE 51.73% 1.02% 404 SharedPool-Worker-10 RUNNABLE 51.10% 1.29% 486 SharedPool-Worker-34 RUNNABLE 51.06% 1.03% Note: Only top 10 threads (according cpu load) are shown! > On 12 Feb 2016, at 18:14, Julien Anguenot <jul...@anguenot.org> wrote: > > At the time when the load is high and you have to restart, do you see any > pending compactions when using `nodetool compactionstats`? > > Possible to see a `nodetool compactionstats` taken *when* the load is too > high? Have you checked the size of your SSTables for that big table? Any > large ones in there? What about the Java HEAP configuration on these nodes? > > If you have too many tombstones I would try to decrease gc_grace_seconds so > they get cleared out earlier during compactions. > > J. > >> On Feb 12, 2016, at 8:45 AM, Skvazh Roman <r...@skvazh.com> wrote: >> >> There is 1-4 compactions at that moment. >> We have many tombstones, which does not removed. >> DroppableTombstoneRatio is 5-6 (greater than 1) >> >>> On 12 Feb 2016, at 15:53, Julien Anguenot <jul...@anguenot.org> wrote: >>> >>> Hey, >>> >>> What about compactions count when that is happening? >>> >>> J. >>> >>> >>>> On Feb 12, 2016, at 3:06 AM, Skvazh Roman <r...@skvazh.com> wrote: >>>> >>>> Hello! >>>> We have a cluster of 25 c3.4xlarge nodes (16 cores, 32 GiB) with attached >>>> 1.5 TB 4000 PIOPS EBS drive. >>>> Sometimes one or two nodes user cpu spikes to 100%, load average to 20-30 >>>> - read requests drops of. >>>> Only restart of this cassandra services helps. >>>> Please advice. >>>> >>>> One big table with wide rows. 600 Gb per node. >>>> LZ4Compressor >>>> LeveledCompaction >>>> >>>> concurrent compactors: 4 >>>> compactor throughput: tried from 16 to 128 >>>> Concurrent_readers: from 16 to 32 >>>> Concurrent_writers: 128 >>>> >>>> >>>> https://gist.github.com/rskvazh/de916327779b98a437a6 >>>> >>>> >>>> JvmTop 0.8.0 alpha - 06:51:10, amd64, 16 cpus, Linux 3.14.44-3, load avg >>>> 19.35 >>>> http://code.google.com/p/jvmtop >>>> >>>> Profiling PID 9256: org.apache.cassandra.service.CassandraDa >>>> >>>> 95.73% ( 4.31s) >>>> ....google.common.collect.AbstractIterator.tryToComputeN() >>>> 1.39% ( 0.06s) com.google.common.base.Objects.hashCode() >>>> 1.26% ( 0.06s) io.netty.channel.epoll.Native.epollWait() >>>> 0.85% ( 0.04s) net.jpountz.lz4.LZ4JNI.LZ4_compress_limitedOutput() >>>> 0.46% ( 0.02s) net.jpountz.lz4.LZ4JNI.LZ4_decompress_fast() >>>> 0.26% ( 0.01s) com.google.common.collect.Iterators$7.computeNext() >>>> 0.06% ( 0.00s) io.netty.channel.epoll.Native.eventFdWrite() >>>> >>>> >>>> ttop: >>>> >>>> 2016-02-12T08:20:25.605+0000 Process summary >>>> process cpu=1565.15% >>>> application cpu=1314.48% (user=1354.48% sys=-40.00%) >>>> other: cpu=250.67% >>>> heap allocation rate 146mb/s >>>> [000405] user=76.25% sys=-0.54% alloc= 0b/s - SharedPool-Worker-9 >>>> [000457] user=75.54% sys=-1.26% alloc= 0b/s - SharedPool-Worker-14 >>>> [000451] user=73.52% sys= 0.29% alloc= 0b/s - SharedPool-Worker-16 >>>> [000311] user=76.45% sys=-2.99% alloc= 0b/s - SharedPool-Worker-4 >>>> [000389] user=70.69% sys= 2.62% alloc= 0b/s - SharedPool-Worker-6 >>>> [000388] user=86.95% sys=-14.28% alloc= 0b/s - SharedPool-Worker-5 >>>> [000404] user=70.69% sys= 0.10% alloc= 0b/s - SharedPool-Worker-8 >>>> [000390] user=72.61% sys=-1.82% alloc= 0b/s - SharedPool-Worker-7 >>>> [000255] user=87.86% sys=-17.87% alloc= 0b/s - SharedPool-Worker-1 >>>> [000444] user=72.21% sys=-2.30% alloc= 0b/s - SharedPool-Worker-12 >>>> [000310] user=71.50% sys=-2.31% alloc= 0b/s - SharedPool-Worker-3 >>>> [000445] user=69.68% sys=-0.83% alloc= 0b/s - SharedPool-Worker-13 >>>> [000406] user=72.61% sys=-4.40% alloc= 0b/s - SharedPool-Worker-10 >>>> [000446] user=69.78% sys=-1.65% alloc= 0b/s - SharedPool-Worker-11 >>>> [000452] user=66.86% sys= 0.22% alloc= 0b/s - SharedPool-Worker-15 >>>> [000256] user=69.08% sys=-2.42% alloc= 0b/s - SharedPool-Worker-2 >>>> [004496] user=29.99% sys= 0.59% alloc= 30mb/s - CompactionExecutor:15 >>>> [004906] user=29.49% sys= 0.74% alloc= 39mb/s - CompactionExecutor:16 >>>> [010143] user=28.58% sys= 0.25% alloc= 26mb/s - CompactionExecutor:17 >>>> [000785] user=27.87% sys= 0.70% alloc= 38mb/s - CompactionExecutor:12 >>>> [012723] user= 9.09% sys= 2.46% alloc= 2977kb/s - RMI TCP >>>> Connection(2673)-127.0.0.1 >>>> [000555] user= 5.35% sys=-0.08% alloc= 474kb/s - SharedPool-Worker-24 >>>> [000560] user= 3.94% sys= 0.07% alloc= 434kb/s - SharedPool-Worker-22 >>>> [000557] user= 3.94% sys=-0.17% alloc= 339kb/s - SharedPool-Worker-25 >>>> [000447] user= 2.73% sys= 0.60% alloc= 436kb/s - SharedPool-Worker-19 >>>> [000563] user= 3.33% sys=-0.04% alloc= 460kb/s - SharedPool-Worker-20 >>>> [000448] user= 2.73% sys= 0.27% alloc= 414kb/s - SharedPool-Worker-21 >>>> [000554] user= 1.72% sys= 0.70% alloc= 232kb/s - SharedPool-Worker-26 >>>> [000558] user= 1.41% sys= 0.39% alloc= 213kb/s - SharedPool-Worker-23 >>>> [000450] user= 1.41% sys=-0.03% alloc= 158kb/s - SharedPool-Worker-17 >>> > > >