[
https://issues.apache.org/jira/browse/CASSANDRA-11447?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15215864#comment-15215864
]
Mark Manley commented on CASSANDRA-11447:
-----------------------------------------
Fair enough. I am unclear as to how all these tasks were cancelled though.
> Flush writer deadlock in Cassandra 2.2.5
> ----------------------------------------
>
> Key: CASSANDRA-11447
> URL: https://issues.apache.org/jira/browse/CASSANDRA-11447
> Project: Cassandra
> Issue Type: Bug
> Reporter: Mark Manley
> Attachments: cassandra.jstack.out
>
>
> When writing heavily to one of my Cassandra tables, I got a deadlock similar
> to CASSANDRA-9882:
> {code}
> "MemtableFlushWriter:4589" #34721 daemon prio=5 os_prio=0
> tid=0x0000000005fc11d0 nid=0x7664 waiting for monitor entry
> [0x00007fb83f0e5000]
> java.lang.Thread.State: BLOCKED (on object monitor)
> at
> org.apache.cassandra.db.compaction.WrappingCompactionStrategy.handleNotification(WrappingCompactionStrategy.java:266)
> - waiting to lock <0x0000000400956258> (a
> org.apache.cassandra.db.compaction.WrappingCompactionStrategy)
> at
> org.apache.cassandra.db.lifecycle.Tracker.notifyAdded(Tracker.java:400)
> at
> org.apache.cassandra.db.lifecycle.Tracker.replaceFlushed(Tracker.java:332)
> at
> org.apache.cassandra.db.compaction.AbstractCompactionStrategy.replaceFlushed(AbstractCompactionStrategy.java:235)
> at
> org.apache.cassandra.db.ColumnFamilyStore.replaceFlushed(ColumnFamilyStore.java:1580)
> at
> org.apache.cassandra.db.Memtable$FlushRunnable.runMayThrow(Memtable.java:362)
> at
> org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28)
> at
> com.google.common.util.concurrent.MoreExecutors$SameThreadExecutorService.execute(MoreExecutors.java:297)
> at
> org.apache.cassandra.db.ColumnFamilyStore$Flush.run(ColumnFamilyStore.java:1139)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> {code}
> The compaction strategies in this keyspace are mixed with one table using LCS
> and the rest using DTCS. None of the tables here save for the LCS one seem
> to have large SSTable counts:
> {code}
> Table: active_counters
> SSTable count: 2
> --
> Table: aggregation_job_entries
> SSTable count: 2
> --
> Table: dsp_metrics_log
> SSTable count: 207
> --
> Table: dsp_metrics_ts_5min
> SSTable count: 3
> --
> Table: dsp_metrics_ts_day
> SSTable count: 2
> --
> Table: dsp_metrics_ts_hour
> SSTable count: 2
> {code}
> Yet the symptoms are similar.
> The "dsp_metrics_ts_5min" table had had a major compaction shortly before all
> this to get rid of the 400+ SStable files before this system went into use,
> but they should have been eliminated.
> Have other people seen this? I am attaching a strack trace.
> Thanks!
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)