Also, that particular logger is for the internal chunk / page cache. If it can’t allocate from within that pool, it’ll just use a normal bytebuffer.
It’s not really a problem, but if you see performance suffer, upgrade to latest 3.11.4, there was a bit of a perf improvement in the case where that cache fills up. -- Jeff Jirsa > On Mar 6, 2019, at 11:40 AM, Jonathan Haddad <j...@jonhaddad.com> wrote: > > That’s not an error. To the left of the log message is the severity, level > INFO. > > Generally, I don’t recommend running Cassandra on only 2GB ram or for small > datasets that can easily fit in memory. Is there a reason why you’re picking > Cassandra for this dataset? > >> On Thu, Mar 7, 2019 at 8:04 AM Kyrylo Lebediev <klebed...@conductor.com> >> wrote: >> Hi All, >> >> >> >> We have a tiny 3-node cluster >> >> C* version 3.9 (I know 3.11 is better/stable, but can’t upgrade immediately) >> >> HEAP_SIZE is 2G >> >> JVM options are default >> >> All setting in cassandra.yaml are default (file_cache_size_in_mb not set) >> >> >> >> Data per node – just ~ 1Gbyte >> >> >> >> We’re getting following errors messages: >> >> >> >> DEBUG [CompactionExecutor:87412] 2019-03-06 11:00:13,545 >> CompactionTask.java:150 - Compacting (ed4a4d90-4028-11e9-adc0-230e0d6622df) >> [/cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/mc-23248-big-Data.db:level=0, >> >> /cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/mc-23247-big-Data.db:level=0, >> >> /cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/mc-23246-big-Data.db:level=0, >> >> /cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/mc-23245-big-Data.db:level=0, >> ] >> >> DEBUG [CompactionExecutor:87412] 2019-03-06 11:00:13,582 >> CompactionTask.java:230 - Compacted (ed4a4d90-4028-11e9-adc0-230e0d6622df) 4 >> sstables to >> [/cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/mc-23249-big,] >> to level=0. 6.264KiB to 1.485KiB (~23% of original) in 36ms. Read >> Throughput = 170.754KiB/s, Write Throughput = 40.492KiB/s, Row Throughput = >> ~106/s. 194 total partitions merged to 44. Partition merge counts were >> {1:18, 4:44, } >> >> INFO [IndexSummaryManager:1] 2019-03-06 11:00:22,007 >> IndexSummaryRedistribution.java:75 - Redistributing index summaries >> >> INFO [pool-1-thread-1] 2019-03-06 11:11:24,903 NoSpamLogger.java:91 - >> Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB >> >> INFO [pool-1-thread-1] 2019-03-06 11:26:24,926 NoSpamLogger.java:91 - >> Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB >> >> INFO [pool-1-thread-1] 2019-03-06 11:41:25,010 NoSpamLogger.java:91 - >> Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB >> >> INFO [pool-1-thread-1] 2019-03-06 11:56:25,018 NoSpamLogger.java:91 - >> Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB >> >> >> >> What’s interesting that “Maximum memory usage reached” messages appears each >> 15 minutes. >> >> Reboot temporary solve the issue, but it then appears again after some time >> >> >> >> Checked, there are no huge partitions (max partition size is ~2Mbytes ) >> >> >> >> How such small amount of data may cause this issue? >> >> How to debug this issue further? >> >> >> >> >> >> Regards, >> >> Kyrill >> >> >> >> >> > -- > Jon Haddad > http://www.rustyrazorblade.com > twitter: rustyrazorblade