By default Cassandra is set to generate a heap dump on OOM. It can be a bit
tricky to figure out what’s going on exactly but it’s the best evidence you
can work with.

On Tue, Aug 7, 2018 at 6:30 AM Laszlo Szabo <laszlo.viktor.sz...@gmail.com>
wrote:

> Hi,
>
> Thanks for the fast response!
>
> We are not using any materialized views, but there are several indexes.  I
> don't have a recent heap dump, and it will be about 24 before I can
> generate an interesting one, but most of the memory was allocated to byte
> buffers, so not entirely helpful.
>
> nodetool cfstats is also below.
>
> I also see a lot of flushing happening, but it seems like there are too
> many small allocations to be effective.  Here are the messages I see,
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,459
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='gpsmessages') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,459 ColumnFamilyStore.java:915
>> - Enqueuing flush of gpsmessages: 0.000KiB (0%) on-heap, 0.014KiB (0%)
>> off-heap
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,460
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='user_history') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,461 ColumnFamilyStore.java:915
>> - Enqueuing flush of user_history: 0.000KiB (0%) on-heap, 0.011KiB (0%)
>> off-heap
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,465
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='tweets') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,465 ColumnFamilyStore.java:915
>> - Enqueuing flush of tweets: 0.000KiB (0%) on-heap, 0.188KiB (0%) off-heap
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='user_history') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:915
>> - Enqueuing flush of user_history: 0.000KiB (0%) on-heap, 0.024KiB (0%)
>> off-heap
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='tweets') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:915
>> - Enqueuing flush of tweets: 0.000KiB (0%) on-heap, 0.188KiB (0%) off-heap
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,472
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='gpsmessages') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,472 ColumnFamilyStore.java:915
>> - Enqueuing flush of gpsmessages: 0.000KiB (0%) on-heap, 0.013KiB (0%)
>> off-heap
>
>
>>
>
> Stack traces from errors are below.
>
>
>> java.io.IOException: Broken pipe
>
>         at sun.nio.ch.FileDispatcherImpl.write0(Native Method)
>> ~[na:1.8.0_181]
>
>         at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:47)
>> ~[na:1.8.0_181]
>
>         at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93)
>> ~[na:1.8.0_181]
>
>         at sun.nio.ch.IOUtil.write(IOUtil.java:51) ~[na:1.8.0_181]
>
>         at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:471)
>> ~[na:1.8.0_181]
>
>         at
>> org.apache.cassandra.io.util.BufferedDataOutputStreamPlus.doFlush(BufferedDataOutputStreamPlus.java:323)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.io.util.BufferedDataOutputStreamPlus.flush(BufferedDataOutputStreamPlus.java:331)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.streaming.ConnectionHandler$OutgoingMessageHandler.sendMessage(ConnectionHandler.java:409)
>> [apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.streaming.ConnectionHandler$OutgoingMessageHandler.run(ConnectionHandler.java:380)
>> [apache-cassandra-3.11.1.jar:3.11.1]
>
>         at java.lang.Thread.run(Thread.java:748) [na:1.8.0_181]
>
> ERROR [MutationStage-226] 2018-08-06 07:16:08,236
>> JVMStabilityInspector.java:142 - JVM state determined to be unstable.
>> Exiting forcefully due to:
>
> java.lang.OutOfMemoryError: Direct buffer memory
>
>         at java.nio.Bits.reserveMemory(Bits.java:694) ~[na:1.8.0_181]
>
>         at java.nio.DirectByteBuffer.<init>(DirectByteBuffer.java:123)
>> ~[na:1.8.0_181]
>
>         at java.nio.ByteBuffer.allocateDirect(ByteBuffer.java:311)
>> ~[na:1.8.0_181]
>
>         at
>> org.apache.cassandra.utils.memory.SlabAllocator.getRegion(SlabAllocator.java:139)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.utils.memory.SlabAllocator.allocate(SlabAllocator.java:104)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.utils.memory.ContextAllocator.allocate(ContextAllocator.java:57)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.utils.memory.ContextAllocator.clone(ContextAllocator.java:47)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.utils.memory.MemtableBufferAllocator.clone(MemtableBufferAllocator.java:40)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at org.apache.cassandra.db.Memtable.put(Memtable.java:269)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.db.ColumnFamilyStore.apply(ColumnFamilyStore.java:1332)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.db.Keyspace.applyInternal(Keyspace.java:618)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at org.apache.cassandra.db.Keyspace.applyFuture(Keyspace.java:425)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at org.apache.cassandra.db.Mutation.applyFuture(Mutation.java:222)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.db.MutationVerbHandler.doVerb(MutationVerbHandler.java:68)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.net.MessageDeliveryTask.run(MessageDeliveryTask.java:66)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>> ~[na:1.8.0_181]
>
>         at
>> org.apache.cassandra.concurrent.AbstractLocalAwareExecutorService$FutureTask.run(AbstractLocalAwareExecutorService.java:162)
>> ~[apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.concurrent.AbstractLocalAwareExecutorService$LocalSessionFutureTask.run(AbstractLocalAwareExecutorService.java:134)
>> [apache-cassandra-3.11.1.jar:3.11.1]
>
>         at
>> org.apache.cassandra.concurrent.SEPWorker.run(SEPWorker.java:109)
>> [apache-cassandra-3.11.1.jar:3.11.1]
>
>         at java.lang.Thread.run(Thread.java:748) [na:1.8.0_181]
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,459
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='gpsmessages') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,459 ColumnFamilyStore.java:915
>> - Enqueuing flush of gpsmessages: 0.000KiB (0%) on-heap, 0.014KiB (0%)
>> off-heap
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,460
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='user_history') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,461 ColumnFamilyStore.java:915
>> - Enqueuing flush of user_history: 0.000KiB (0%) on-heap, 0.011KiB (0%)
>> off-heap
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,465
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='tweets') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,465 ColumnFamilyStore.java:915
>> - Enqueuing flush of tweets: 0.000KiB (0%) on-heap, 0.188KiB (0%) off-heap
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='user_history') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:915
>> - Enqueuing flush of user_history: 0.000KiB (0%) on-heap, 0.024KiB (0%)
>> off-heap
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='tweets') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,470 ColumnFamilyStore.java:915
>> - Enqueuing flush of tweets: 0.000KiB (0%) on-heap, 0.188KiB (0%) off-heap
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,472
>> ColumnFamilyStore.java:1305 - Flushing largest CFS(Keyspace='userinfo',
>> ColumnFamily='gpsmessages') to free up room. Used total: 0.54/0.05, live:
>> 0.00/0.00, flushing: 0.40/0.04, this: 0.00/0.00
>
> DEBUG [SlabPoolCleaner] 2018-08-06 07:16:08,472 ColumnFamilyStore.java:915
>> - Enqueuing flush of gpsmessages: 0.000KiB (0%) on-heap, 0.013KiB (0%)
>> off-heap
>
>
> Total number of tables: 40
> ----------------
> Keyspace : userinfo
> Read Count: 143301
> Read Latency: 14.945587623254548 ms.
> Write Count: 2754603904
> Write Latency: 0.020883145284324698 ms.
> Pending Flushes: 0
> Table (index):
> gpsmessages.gpsmessages_addresscount_idxgpsmessages.gpsmessages_addresscount_idx
> SSTable count: 9
> Space used (live): 19043463189
> Space used (total): 19043463189
> Space used by snapshots (total): 0
> Off heap memory used (total): 6259448
> SSTable Compression Ratio: 0.3704785164266614
> Number of partitions (estimate): 1025
> Memtable cell count: 309066
> Memtable data size: 13602774
> Memtable off heap memory used: 0
> Memtable switch count: 0
> Local read count: 0
> Local read latency: NaN ms
> Local write count: 46025778
> Local write latency: 0.034 ms
> Pending flushes: 0
> Percent repaired: 0.0
> Bloom filter false positives: 0
> Bloom filter false ratio: 0.00000
> Bloom filter space used: 2504
> Bloom filter off heap memory used: 2432
> Index summary off heap memory used: 320
> Compression metadata off heap memory used: 6256696
> Compacted partition minimum bytes: 43
> Compacted partition maximum bytes: 44285675122
> Compacted partition mean bytes: 30405277
> Average live cells per slice (last five minutes): NaN
> Maximum live cells per slice (last five minutes): 0
> Average tombstones per slice (last five minutes): NaN
> Maximum tombstones per slice (last five minutes): 0
> Dropped Mutations: 0
>
> Table (index):
> gpsmessages.gpsmessages_addresses_idxgpsmessages.gpsmessages_addresses_idx
> SSTable count: 18
> Space used (live): 409514565570
> Space used (total): 409514565570
> Space used by snapshots (total): 0
> Off heap memory used (total): 153405673
> SSTable Compression Ratio: 0.4447731157134059
> Number of partitions (estimate): 6013125
> Memtable cell count: 1110334
> Memtable data size: 67480140
> Memtable off heap memory used: 0
> Memtable switch count: 0
> Local read count: 0
> Local read latency: NaN ms
> Local write count: 147639252
> Local write latency: 0.015 ms
> Pending flushes: 0
> Percent repaired: 0.0
> Bloom filter false positives: 0
> Bloom filter false ratio: 0.00000
> Bloom filter space used: 34175400
> Bloom filter off heap memory used: 34175256
> Index summary off heap memory used: 7432177
> Compression metadata off heap memory used: 111798240
> Compacted partition minimum bytes: 61
> Compacted partition maximum bytes: 322381140
> Compacted partition mean bytes: 36692
> Average live cells per slice (last five minutes): NaN
> Maximum live cells per slice (last five minutes): 0
> Average tombstones per slice (last five minutes): NaN
> Maximum tombstones per slice (last five minutes): 0
> Dropped Mutations: 0
>
> Table (index):
> gpsmessages.addressreceivedtime_idxgpsmessages.addressreceivedtime_idx
> SSTable count: 10
> Space used (live): 52738155477
> Space used (total): 52738155477
> Space used by snapshots (total): 0
> Off heap memory used (total): 1909362628
> SSTable Compression Ratio: 0.4106961621795128
> Number of partitions (estimate): 1338730016
> Memtable cell count: 308990
> Memtable data size: 13410867
> Memtable off heap memory used: 0
> Memtable switch count: 0
> Local read count: 0
> Local read latency: NaN ms
> Local write count: 46012614
> Local write latency: 0.012 ms
> Pending flushes: 0
> Percent repaired: 0.0
> Bloom filter false positives: 0
> Bloom filter false ratio: 0.00000
> Bloom filter space used: 1687550888
> Bloom filter off heap memory used: 1687550808
> Index summary off heap memory used: 213249180
> Compression metadata off heap memory used: 8562640
> Compacted partition minimum bytes: 36
> Compacted partition maximum bytes: 2759
> Compacted partition mean bytes: 54
> Average live cells per slice (last five minutes): NaN
> Maximum live cells per slice (last five minutes): 0
> Average tombstones per slice (last five minutes): NaN
> Maximum tombstones per slice (last five minutes): 0
> Dropped Mutations: 0
>
> Table: gpsmessages
> SSTable count: 13
> Space used (live): 337974446627
> Space used (total): 337974446627
> Space used by snapshots (total): 0
> Off heap memory used (total): 77833540
> SSTable Compression Ratio: 0.5300637241381126
> Number of partitions (estimate): 22034
> Memtable cell count: 308904
> Memtable data size: 72074512
> Memtable off heap memory used: 0
> Memtable switch count: 110
> Local read count: 0
> Local read latency: NaN ms
> Local write count: 45996652
> Local write latency: 0.281 ms
> Pending flushes: 0
> Percent repaired: 0.0
> Bloom filter false positives: 0
> Bloom filter false ratio: 0.00000
> Bloom filter space used: 67904
> Bloom filter off heap memory used: 67800
> Index summary off heap memory used: 11756
> Compression metadata off heap memory used: 77753984
> Compacted partition minimum bytes: 73
> Compacted partition maximum bytes: 1155149911
> Compacted partition mean bytes: 13158224
> Average live cells per slice (last five minutes): NaN
> Maximum live cells per slice (last five minutes): 0
> Average tombstones per slice (last five minutes): NaN
> Maximum tombstones per slice (last five minutes): 0
> Dropped Mutations: 13699
>
> Table: user_history
> SSTable count: 17
> Space used (live): 116361158882
> Space used (total): 116361158882
> Space used by snapshots (total): 0
> Off heap memory used (total): 29562319
> SSTable Compression Ratio: 0.5683114352331539
> Number of partitions (estimate): 1337206
> Memtable cell count: 773277
> Memtable data size: 40623368
> Memtable off heap memory used: 0
> Memtable switch count: 57
> Local read count: 209
> Local read latency: NaN ms
> Local write count: 145853733
> Local write latency: 0.020 ms
> Pending flushes: 0
> Percent repaired: 0.0
> Bloom filter false positives: 0
> Bloom filter false ratio: 0.00000
> Bloom filter space used: 3844416
> Bloom filter off heap memory used: 3844280
> Index summary off heap memory used: 800991
> Compression metadata off heap memory used: 24917048
> Compacted partition minimum bytes: 61
> Compacted partition maximum bytes: 464228842
> Compacted partition mean bytes: 72182
> Average live cells per slice (last five minutes): NaN
> Maximum live cells per slice (last five minutes): 0
> Average tombstones per slice (last five minutes): NaN
> Maximum tombstones per slice (last five minutes): 0
> Dropped Mutations: 66702
>
> Table: users
> SSTable count: 3
> Space used (live): 89945186
> Space used (total): 89945186
> Space used by snapshots (total): 0
> Off heap memory used (total): 2092053
> SSTable Compression Ratio: 0.5712127629253333
> Number of partitions (estimate): 1365645
> Memtable cell count: 3556
> Memtable data size: 150903
> Memtable off heap memory used: 0
> Memtable switch count: 42
> Local read count: 143087
> Local read latency: 6.094 ms
> Local write count: 250971
> Local write latency: 0.024 ms
> Pending flushes: 0
> Percent repaired: 0.0
> Bloom filter false positives: 0
> Bloom filter false ratio: 0.00000
> Bloom filter space used: 1709848
> Bloom filter off heap memory used: 1709824
> Index summary off heap memory used: 372125
> Compression metadata off heap memory used: 10104
> Compacted partition minimum bytes: 36
> Compacted partition maximum bytes: 310
> Compacted partition mean bytes: 66
> Average live cells per slice (last five minutes): 1.0
> Maximum live cells per slice (last five minutes): 1
> Average tombstones per slice (last five minutes): 1.0
> Maximum tombstones per slice (last five minutes): 1
> Dropped Mutations: 114
>
> Table: tweets
> SSTable count: 18
> Space used (live): 1809145656486
> Space used (total): 1809145656486
> Space used by snapshots (total): 0
> Off heap memory used (total): 435915908
> SSTable Compression Ratio: 0.5726200929451171
> Number of partitions (estimate): 26217889
> Memtable cell count: 710146
> Memtable data size: 31793929
> Memtable off heap memory used: 0
> Memtable switch count: 399
> Local read count: 5
> Local read latency: NaN ms
> Local write count: 2322829524
> Local write latency: 0.019 ms
> Pending flushes: 0
> Percent repaired: 0.0
> Bloom filter false positives: 0
> Bloom filter false ratio: 0.00000
> Bloom filter space used: 35019224
> Bloom filter off heap memory used: 35019080
> Index summary off heap memory used: 16454076
> Compression metadata off heap memory used: 384442752
> Compacted partition minimum bytes: 104
> Compacted partition maximum bytes: 3379391
> Compacted partition mean bytes: 124766
> Average live cells per slice (last five minutes): NaN
> Maximum live cells per slice (last five minutes): 0
> Average tombstones per slice (last five minutes): NaN
> Maximum tombstones per slice (last five minutes): 0
> Dropped Mutations: 697696
>
> ----------------
>
>
> On Mon, Aug 6, 2018 at 8:57 PM, Jeff Jirsa <jji...@gmail.com> wrote:
>
>>
>>
>> Upgrading to 3.11.3 May fix it (there were some memory recycling bugs
>> fixed recently), but analyzing the heap will be the best option
>>
>> If you can print out the heap histogram and stack trace or open a heap
>> dump in your kit or visualvm or MAT and show us what’s at the top of the
>> reclaimed objects, we may be able to figure out what’s going on
>>
>> --
>> Jeff Jirsa
>>
>>
>> On Aug 6, 2018, at 5:42 PM, Jeff Jirsa <jji...@gmail.com> wrote:
>>
>> Are you using materialized views or secondary indices?
>>
>> --
>> Jeff Jirsa
>>
>>
>> On Aug 6, 2018, at 3:49 PM, Laszlo Szabo <laszlo.viktor.sz...@gmail.com>
>> wrote:
>>
>> Hello All,
>>
>> I'm having JVM unstable / OOM errors when attempting to auto bootstrap a
>> 9th node to an existing 8 node cluster (256 tokens).  Each machine has 24
>> cores 148GB RAM and 10TB (2TB used).  Under normal operation the 8 nodes
>> have JVM memory configured with Xms35G and Xmx35G, and handle 2-4 billion
>> inserts per day.  There are never updates, deletes, or sparsely populated
>> rows.
>>
>> For the bootstrap node, I've tried memory values from 35GB to 135GB in
>> 10GB increments. I've tried using both memtable_allocation_types
>> (heap_buffers and offheap_buffers).  I've not tried modifying the
>> memtable_cleanup_threshold but instead have tried memtable_flush_writers
>> from 2 to 8.  I've tried memtable_(off)heap_space_in_mb from 20000 to
>> 60000.  I've tried both CMS and G1 garbage collection with various
>> settings.
>>
>> Typically, after streaming about ~2TB of data, CPU load will hit a
>> maximum, and the "nodetool info" heap memory will, over the course of an
>> hour, approach the maximum.  At that point, CPU load will drop to a single
>> thread with minimal activity until the system becomes unstable and
>> eventually the OOM error occurs.
>>
>> Excerpt of the system log is below, and what I consistently see is the
>> MemtableFlushWriter and the MemtableReclaimMemory pending queues grow as
>> the memory becomes depleted, but the number of completed seems to stop
>> changing a few minutes after the CPU load spikes.
>>
>> One other data point is there seems to be a huge number of mutations that
>> occur after most of the stream has occured.  Concurrent_writes is set at
>> 256 with the queue getting as high as 200K before dropping down.
>>
>> Any suggestions for yaml changes or jvm changes?  JVM.options is
>> currently the default with the memory set to the max, the current YAML file
>> is below.
>>
>> Thanks!
>>
>>
>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,329 StatusLogger.java:51 -
>>> MutationStage                     1         2      191498052         0
>>>            0
>>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,331 StatusLogger.java:51 -
>>> ViewMutationStage                 0         0              0         0
>>>            0
>>> INFO  [Service Thread] 2018-08-06 17:49:26,338 StatusLogger.java:51 -
>>> PerDiskMemtableFlushWriter_0         0         0           5865         0
>>>                0
>>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,343 StatusLogger.java:51 -
>>> ReadStage                         0         0              0         0
>>>            0
>>> INFO  [Service Thread] 2018-08-06 17:49:26,347 StatusLogger.java:51 -
>>> ValidationExecutor                0         0              0         0
>>>            0
>>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,360 StatusLogger.java:51 -
>>> RequestResponseStage              0         0              8         0
>>>            0
>>> INFO  [Service Thread] 2018-08-06 17:49:26,380 StatusLogger.java:51 -
>>> Sampler                           0         0              0         0
>>>            0
>>> INFO  [Service Thread] 2018-08-06 17:49:26,382 StatusLogger.java:51 - 
>>> *MemtableFlushWriter
>>>              8     74293           4716         0  *               0
>>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,388 StatusLogger.java:51 -
>>> ReadRepairStage                   0         0              0         0
>>>            0
>>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,389 StatusLogger.java:51 -
>>> CounterMutationStage              0         0              0         0
>>>            0
>>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,404 StatusLogger.java:51 -
>>> MiscStage                         0         0              0         0
>>>            0
>>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,407 StatusLogger.java:51 -
>>> CompactionExecutor                8        13            493         0
>>>            0
>>> INFO  [Service Thread] 2018-08-06 17:49:26,410 StatusLogger.java:51 -
>>> InternalResponseStage             0         0             16         0
>>>            0
>>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,413 StatusLogger.java:51 - 
>>> *MemtableReclaimMemory
>>>            1      6066            356         0     *            0
>>> INFO  [Service Thread] 2018-08-06 17:49:26,421 StatusLogger.java:51 -
>>> AntiEntropyStage                  0         0              0         0
>>>            0
>>> INFO  [Service Thread] 2018-08-06 17:49:26,430 StatusLogger.java:51 -
>>> CacheCleanupExecutor              0         0              0         0
>>>            0
>>> INFO  [ScheduledTasks:1] 2018-08-06 17:49:26,431 StatusLogger.java:51 -
>>> PendingRangeCalculator            0         0              9         0
>>>            0
>>> INFO  [Service Thread] 2018-08-06 17:49:26,436 StatusLogger.java:61 -
>>> CompactionManager                 8        19
>>
>>
>>
>>
>>  Current Yaml
>>
>> num_tokens: 256
>>
>> hinted_handoff_enabled: true
>>
>> hinted_handoff_throttle_in_kb: 10240
>>
>> max_hints_delivery_threads: 8
>>
>> hints_flush_period_in_ms: 10000
>>
>> max_hints_file_size_in_mb: 128
>>
>> batchlog_replay_throttle_in_kb: 10240
>>
>> authenticator: AllowAllAuthenticator
>>
>> authorizer: AllowAllAuthorizer
>>
>> role_manager: CassandraRoleManager
>>
>> roles_validity_in_ms: 2000
>>
>> permissions_validity_in_ms: 2000
>>
>> credentials_validity_in_ms: 2000
>>
>> partitioner: org.apache.cassandra.dht.Murmur3Partitioner
>>
>> data_file_directories:
>>
>>     - /data/cassandra/data
>>
>> commitlog_directory: /data/cassandra/commitlog
>>
>> cdc_enabled: false
>>
>> disk_failure_policy: stop
>>
>> commit_failure_policy: stop
>>
>> prepared_statements_cache_size_mb:
>>
>> thrift_prepared_statements_cache_size_mb:
>>
>> key_cache_size_in_mb:
>>
>> key_cache_save_period: 14400
>>
>> row_cache_size_in_mb: 0
>>
>> row_cache_save_period: 0
>>
>> counter_cache_size_in_mb:
>>
>> counter_cache_save_period: 7200
>>
>> saved_caches_directory: /data/cassandra/saved_caches
>>
>> commitlog_sync: periodic
>>
>> commitlog_sync_period_in_ms: 10000
>>
>> commitlog_segment_size_in_mb: 32
>>
>> seed_provider:
>>
>>     - class_name: org.apache.cassandra.locator.SimpleSeedProvider
>>
>>       parameters:
>>
>>           - seeds: "10.1.1.11,10.1.1.12,10.1.1.13"
>>
>> concurrent_reads: 128
>>
>> concurrent_writes: 256
>>
>> concurrent_counter_writes: 96
>>
>> concurrent_materialized_view_writes: 32
>>
>> disk_optimization_strategy: spinning
>>
>> memtable_heap_space_in_mb: 61440
>>
>> memtable_offheap_space_in_mb: 61440
>>
>> memtable_allocation_type: heap_buffers
>>
>> commitlog_total_space_in_mb: 81920
>>
>> memtable_flush_writers: 8
>>
>>
>>
> --
Jon Haddad
http://www.rustyrazorblade.com
twitter: rustyrazorblade

Reply via email to