Michael Fong created CASSANDRA-11624: ----------------------------------------
Summary: Scrub does not seem to work on previously marked corrupted SSTables Key: CASSANDRA-11624 URL: https://issues.apache.org/jira/browse/CASSANDRA-11624 Project: Cassandra Issue Type: Bug Reporter: Michael Fong We ran into a scenario that scrub does not seem to work on a previously marked-as-corrupted SSTable. Here is the log snippet related to the corrupted SSTable and scrub-attempt : ERROR [ReadStage:174] 2016-03-17 04:14:39,658 CassandraDaemon.java (line 258) Exception in thread Thread[ReadStage:174,5,main] java.lang.RuntimeException: org.apache.cassandra.io.sstable.CorruptSSTableException: java.io.IOException: mmap segment underflow; remaining is 10197 but 30062 requested for /data/ng/db/data/wsg/dpStatusRealTime/wsg-dpStatusRealTime-ic-2-Data.db at org.apache.cassandra.service.StorageProxy$DroppableRunnable.run(StorageProxy.java:2022) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.cassandra.io.sstable.CorruptSSTableException: java.io.IOException: mmap segment underflow; remaining is 10197 but 30062 requested for /data/ng/db/data/wsg/dpStatusRealTime/wsg-dpStatusRealTime-ic-2-Data.db at org.apache.cassandra.db.columniterator.IndexedSliceReader.<init>(IndexedSliceReader.java:97) at org.apache.cassandra.db.columniterator.SSTableSliceIterator.createReader(SSTableSliceIterator.java:65) at org.apache.cassandra.db.columniterator.SSTableSliceIterator.<init>(SSTableSliceIterator.java:42) at org.apache.cassandra.db.filter.SliceQueryFilter.getSSTableColumnIterator(SliceQueryFilter.java:238) at org.apache.cassandra.db.filter.QueryFilter.getSSTableColumnIterator(QueryFilter.java:62) at org.apache.cassandra.db.CollationController.collectAllData(CollationController.java:250) at org.apache.cassandra.db.CollationController.getTopLevelColumns(CollationController.java:53) at org.apache.cassandra.db.ColumnFamilyStore.getTopLevelColumns(ColumnFamilyStore.java:1642) at org.apache.cassandra.db.ColumnFamilyStore.getColumnFamily(ColumnFamilyStore.java:1461) at org.apache.cassandra.db.Keyspace.getRow(Keyspace.java:340) at org.apache.cassandra.db.SliceFromReadCommand.getRow(SliceFromReadCommand.java:89) at org.apache.cassandra.service.StorageProxy$LocalReadRunnable.runMayThrow(StorageProxy.java:1445) at org.apache.cassandra.service.StorageProxy$DroppableRunnable.run(StorageProxy.java:2010) ... 3 more INFO [CompactionExecutor:98] 2016-03-17 04:14:39,693 OutputHandler.java (line 42) Scrubbing SSTableReader(path='/data/ng/db/data/wsg/dpStatusRealTime/wsg-dpStatusRealTime-jb-11-Data.db') (2230223 bytes) INFO [CompactionExecutor:98] 2016-03-17 04:14:39,751 OutputHandler.java (line 42) Scrub of SSTableReader(path='/data/ng/db/data/wsg/dpStatusRealTime/wsg-dpStatusRealTime-jb-11-Data.db') complete: 2 rows in new sstable and 0 empty (tombstoned) rows dropped ------------------ Below is the file information around that time ------------------ -bash-4.1$ ls -alF /data/ng/db/data/wsg/dpStatusRealTime/ total 2328 drwxr-xr-x 2 root root 4096 Mar 17 04:14 ./ drwxr-xr-x 264 root root 12288 Mar 16 06:48 ../ -rw-r--r-- 1 root root 72995 Mar 16 07:08 wsg-dpStatusRealTime-ic-2-Data.db -rw-r--r-- 1 root root 75 Mar 16 07:08 wsg-dpStatusRealTime-ic-2-Digest.sha1 -rw-r--r-- 1 root root 16 Mar 16 07:08 wsg-dpStatusRealTime-ic-2-Filter.db -rw-r--r-- 1 root root 132 Mar 16 07:08 wsg-dpStatusRealTime-ic-2-Index.db -rw-r--r-- 1 root root 5956 Mar 16 07:08 wsg-dpStatusRealTime-ic-2-Statistics.db -rw-r--r-- 1 root root 244 Mar 16 07:20 wsg-dpStatusRealTime-ic-2-Summary.db -rw-r--r-- 1 root root 72 Mar 16 07:08 wsg-dpStatusRealTime-ic-2-TOC.txt -rw-r--r-- 1 root root 144 Mar 17 04:14 wsg-dpStatusRealTime-jb-12-CRC.db -rw-r--r-- 1 root root 2230223 Mar 17 04:14 wsg-dpStatusRealTime-jb-12-Data.db -rw-r--r-- 1 root root 76 Mar 17 04:14 wsg-dpStatusRealTime-jb-12-Digest.sha1 -rw-r--r-- 1 root root 336 Mar 17 04:14 wsg-dpStatusRealTime-jb-12-Filter.db -rw-r--r-- 1 root root 1424 Mar 17 04:14 wsg-dpStatusRealTime-jb-12-Index.db -rw-r--r-- 1 root root 6004 Mar 17 04:14 wsg-dpStatusRealTime-jb-12-Statistics.db -rw-r--r-- 1 root root 244 Mar 17 04:14 wsg-dpStatusRealTime-jb-12-Summary.db -rw-r--r-- 1 root root 79 Mar 17 04:14 wsg-dpStatusRealTime-jb-12-TOC.txt -------------------------- 1. Please note that the corrupted file is in (ic) version, which is 1.2.19. This test bed was upgraded and attempted to upgradesstable a day ago. There has been some read/write to that column family before upgradesstable was performed - but I do not have log for that :( 2. I manually performed the scrub/compact operation with nodetool and the result is the same: INFO [CompactionExecutor:103] 2016-03-17 07:09:53,444 CompactionTask.java (line 123) Compacting [SSTableReader(path='/data/ng/db/data/wsg/dpStatusRealTime/wsg-dpStatusRealTime-jb-14-Data.db'), SSTableReader(path='/data/ng/db/data/wsg/dpStatusRealTime/wsg-dpStatusRealTime-jb-15-Data.db')] Our test environment was Cassandra 2.0.17 with max heap set with 8 gigabytes. -- This message was sent by Atlassian JIRA (v6.3.4#6332)