Do you have any logs you can share for operations on these SSTables? Thinking of things like flush, compaction, streaming, etc.
> On Oct 15, 2025, at 11:37 AM, Ivan Zalozhnykh <[email protected]> > wrote: > > We use LeveledCompactionStrategy on both affected tables. > > The issue is not deterministic: the same query may fail once and then succeed > on retry. > Each time the error occurs, it's on a different token range, node and file. > > We’ll try to reproduce the issue with sstable_preemptive_open_interval: null > this week and I’ll follow up with the results. > > Best regards > Ivan Zalozhnykh > > > ср, 15 окт. 2025 г. в 16:58, Branimir Lambov <[email protected] > <mailto:[email protected]>>: >> Hello, >> >> To help pinpoint the problem, could you check if it still happens if early >> open is disabled (i.e. `preemptive_open_interval` is set to null in >> `cassandra.yaml`)? >> >> Regards, >> Branimir >> >> On 2025/10/15 12:07:06 Ivan Zalozhnykh wrote: >> > Hello, >> > >> > We're experiencing a recurring issue in Cassandra 5.0.5 when performing >> > token-range read queries on a table using the BTI sstable format. The issue >> > results in a fatal `AssertionError` during read, which causes read failures >> > in the client with `ReadFailureException`. >> > >> > The exception from Cassandra logs looks like this: >> > >> > ``` >> > ERROR [ReadStage-6] 2025-08-27 19:22:49,331 JVMStabilityInspector.java:70 - >> > Exception in thread Thread[ReadStage-6,5,SharedPool] >> > java.lang.RuntimeException: java.lang.AssertionError: Caught an error while >> > trying to process the command: SELECT * FROM keyspace.table WHERE >> > token(profile_id) > -8987898918496238646 AND token(profile_id) <= >> > -8718169305572166374 LIMIT 1000 ALLOW FILTERING >> > at org.apache.cassandra.net >> > <http://org.apache.cassandra.net/>.InboundSink.accept(InboundSink.java:108) >> > at org.apache.cassandra.net >> > <http://org.apache.cassandra.net/>.InboundSink.accept(InboundSink.java:45) >> > at >> > org.apache.cassandra.net >> > <http://org.apache.cassandra.net/>.InboundMessageHandler$ProcessMessage.run(InboundMessageHandler.java:430) >> > at >> > org.apache.cassandra.concurrent.ExecutionFailure$1.run(ExecutionFailure.java:133) >> > at >> > org.apache.cassandra.concurrent.SEPWorker.run(SEPWorker.java:143) >> > at >> > io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) >> > at java.base/java.lang.Thread.run(Thread.java:829) >> > Caused by: java.lang.AssertionError: Caught an error while trying to >> > process the command: SELECT * FROM keyspace.table WHERE token(profile_id) > >> > -8987898918496238646 AND token(profile_id) <= -8718169305572166374 LIMIT >> > 1000 ALLOW FILTERING >> > at >> > org.apache.cassandra.db.ReadCommandVerbHandler.doVerb(ReadCommandVerbHandler.java:112) >> > at >> > org.apache.cassandra.net >> > <http://org.apache.cassandra.net/>.InboundSink.lambda$new$0(InboundSink.java:78) >> > at org.apache.cassandra.net >> > <http://org.apache.cassandra.net/>.InboundSink.accept(InboundSink.java:97) >> > ... 6 common frames omitted >> > Caused by: java.lang.AssertionError: 406005 > 393216 >> > at >> > org.apache.cassandra.io.util.MmappedRegions$State.floor(MmappedRegions.java:362) >> > at >> > org.apache.cassandra.io.util.MmappedRegions.floor(MmappedRegions.java:241) >> > at >> > org.apache.cassandra.io.util.MmapRebufferer.rebuffer(MmapRebufferer.java:40) >> > at org.apache.cassandra.io.tries.Walker.<init>(Walker.java:75) >> > at >> > org.apache.cassandra.io.tries.ValueIterator.<init>(ValueIterator.java:96) >> > at >> > org.apache.cassandra.io.tries.ValueIterator.<init>(ValueIterator.java:80) >> > at >> > org.apache.cassandra.io >> > <http://org.apache.cassandra.io/>.sstable.format.bti.PartitionIndex$IndexPosIterator.<init>(PartitionIndex.java:407) >> > at >> > org.apache.cassandra.io >> > <http://org.apache.cassandra.io/>.sstable.format.bti.PartitionIterator.<init>(PartitionIterator.java:113) >> > at >> > org.apache.cassandra.io >> > <http://org.apache.cassandra.io/>.sstable.format.bti.PartitionIterator.create(PartitionIterator.java:75) >> > at >> > org.apache.cassandra.io >> > <http://org.apache.cassandra.io/>.sstable.format.bti.BtiTableReader.coveredKeysIterator(BtiTableReader.java:293) >> > at >> > org.apache.cassandra.io >> > <http://org.apache.cassandra.io/>.sstable.format.bti.BtiTableScanner$BtiScanningIterator.prepareToIterateRow(BtiTableScanner.java:93) >> > at >> > org.apache.cassandra.io >> > <http://org.apache.cassandra.io/>.sstable.format.SSTableScanner$BaseKeyScanningIterator.computeNext(SSTableScanner.java:248) >> > at >> > org.apache.cassandra.io >> > <http://org.apache.cassandra.io/>.sstable.format.SSTableScanner$BaseKeyScanningIterator.computeNext(SSTableScanner.java:228) >> > at >> > org.apache.cassandra.utils.AbstractIterator.hasNext(AbstractIterator.java:47) >> > at >> > org.apache.cassandra.io >> > <http://org.apache.cassandra.io/>.sstable.format.SSTableScanner.hasNext(SSTableScanner.java:190) >> > at >> > org.apache.cassandra.db.transform.BasePartitions.hasNext(BasePartitions.java:90) >> > at >> > org.apache.cassandra.utils.MergeIterator$Candidate.advance(MergeIterator.java:375) >> > at >> > org.apache.cassandra.utils.MergeIterator$ManyToOne.advance(MergeIterator.java:187) >> > at >> > org.apache.cassandra.utils.MergeIterator$ManyToOne.computeNext(MergeIterator.java:156) >> > at >> > org.apache.cassandra.utils.AbstractIterator.hasNext(AbstractIterator.java:47) >> > at >> > org.apache.cassandra.db.partitions.UnfilteredPartitionIterators$4.hasNext(UnfilteredPartitionIterators.java:264) >> > at >> > org.apache.cassandra.db.transform.BasePartitions.hasNext(BasePartitions.java:90) >> > at >> > org.apache.cassandra.db.partitions.UnfilteredPartitionIterators$Serializer.serialize(UnfilteredPartitionIterators.java:334) >> > at >> > org.apache.cassandra.db.ReadResponse$LocalDataResponse.build(ReadResponse.java:201) >> > at >> > org.apache.cassandra.db.ReadResponse$LocalDataResponse.<init>(ReadResponse.java:186) >> > at >> > org.apache.cassandra.db.ReadResponse.createDataResponse(ReadResponse.java:48) >> > at >> > org.apache.cassandra.db.ReadCommand.createResponse(ReadCommand.java:374) >> > at >> > org.apache.cassandra.db.ReadCommandVerbHandler.doVerb(ReadCommandVerbHandler.java:93) >> > ... 8 common frames omitted >> > ``` >> > >> > The query itself is a standard paginated scan over token ranges with a >> > LIMIT. Here's a simplified version of the code performing it: >> > >> > ```kotlin >> > @Query("select profile_id, token(profile_id) as page_token from >> > keyspace.table where token(profile_id) > :pageToken limit :limitCount") >> > fun selectWithToken( >> > pageToken: Long, >> > limitCount: Int, >> > ): MappedReactiveResultSet<MigratorEntityToken> >> > ``` >> > >> > This results in read failures like this on the client side: >> > >> > ``` >> > Caused by: >> > com.datastax.oss.driver.api.core.servererrors.ReadFailureException: >> > Cassandra failure during read query at consistency LOCAL_QUORUM (2 >> > responses were required but only 0 replica responded, 1 failed) >> > ``` >> > >> > The issue occurs with multiple SSTables, across multiple token ranges, on >> > different nodes. >> > All the affected Trie-indexed SSTables, and the error disappears when >> > switching the sstable format back to Big. >> > >> > Has anyone else seen similar behavior with BTI sstables in Cassandra 5.0.x? >> > Is this a known issue or something we should file a Jira ticket for? >> > >> > Thanks in advance, >> > Ivan Zalozhnykh >> >
