Also, that particular logger is for the internal chunk / page cache. If it 
can’t allocate from within that pool, it’ll just use a normal bytebuffer. 

It’s not really a problem, but if you see performance suffer, upgrade to latest 
3.11.4, there was a bit of a perf improvement in the case where that cache 
fills up.


-- 
Jeff Jirsa


> On Mar 6, 2019, at 11:40 AM, Jonathan Haddad <j...@jonhaddad.com> wrote:
> 
> That’s not an error. To the left of the log message is the severity, level 
> INFO. 
> 
> Generally, I don’t recommend running Cassandra on only 2GB ram or for small 
> datasets that can easily fit in memory. Is there a reason why you’re picking 
> Cassandra for this dataset?
> 
>> On Thu, Mar 7, 2019 at 8:04 AM Kyrylo Lebediev <klebed...@conductor.com> 
>> wrote:
>> Hi All,
>> 
>>  
>> 
>> We have a tiny 3-node cluster
>> 
>> C* version 3.9 (I know 3.11 is better/stable, but can’t upgrade immediately)
>> 
>> HEAP_SIZE is 2G
>> 
>> JVM options are default
>> 
>> All setting in cassandra.yaml are default (file_cache_size_in_mb not set)
>> 
>>  
>> 
>> Data per node – just ~ 1Gbyte
>> 
>>  
>> 
>> We’re getting following errors messages:
>> 
>>  
>> 
>> DEBUG [CompactionExecutor:87412] 2019-03-06 11:00:13,545 
>> CompactionTask.java:150 - Compacting (ed4a4d90-4028-11e9-adc0-230e0d6622df) 
>> [/cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/mc-23248-big-Data.db:level=0,
>>  
>> /cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/mc-23247-big-Data.db:level=0,
>>  
>> /cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/mc-23246-big-Data.db:level=0,
>>  
>> /cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/mc-23245-big-Data.db:level=0,
>>  ]
>> 
>> DEBUG [CompactionExecutor:87412] 2019-03-06 11:00:13,582 
>> CompactionTask.java:230 - Compacted (ed4a4d90-4028-11e9-adc0-230e0d6622df) 4 
>> sstables to 
>> [/cassandra/data/data/system/sstable_activity-5a1ff267ace03f128563cfae6103c65e/mc-23249-big,]
>>  to level=0.  6.264KiB to 1.485KiB (~23% of original) in 36ms.  Read 
>> Throughput = 170.754KiB/s, Write Throughput = 40.492KiB/s, Row Throughput = 
>> ~106/s.  194 total partitions merged to 44.  Partition merge counts were 
>> {1:18, 4:44, }
>> 
>> INFO  [IndexSummaryManager:1] 2019-03-06 11:00:22,007 
>> IndexSummaryRedistribution.java:75 - Redistributing index summaries
>> 
>> INFO  [pool-1-thread-1] 2019-03-06 11:11:24,903 NoSpamLogger.java:91 - 
>> Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB
>> 
>> INFO  [pool-1-thread-1] 2019-03-06 11:26:24,926 NoSpamLogger.java:91 - 
>> Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB
>> 
>> INFO  [pool-1-thread-1] 2019-03-06 11:41:25,010 NoSpamLogger.java:91 - 
>> Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB
>> 
>> INFO  [pool-1-thread-1] 2019-03-06 11:56:25,018 NoSpamLogger.java:91 - 
>> Maximum memory usage reached (512.000MiB), cannot allocate chunk of 1.000MiB
>> 
>>  
>> 
>> What’s interesting that “Maximum memory usage reached” messages appears each 
>> 15 minutes.
>> 
>> Reboot temporary solve the issue, but it then appears again after some time
>> 
>>  
>> 
>> Checked, there are no huge partitions (max partition size is ~2Mbytes )
>> 
>>  
>> 
>> How such small amount of data may cause this issue?
>> 
>> How to debug this issue further?
>> 
>>  
>> 
>>  
>> 
>> Regards,
>> 
>> Kyrill
>> 
>>  
>> 
>>  
>> 
> -- 
> Jon Haddad
> http://www.rustyrazorblade.com
> twitter: rustyrazorblade

Reply via email to