[jira] [Commented] (CASSANDRA-8267) Only stream from unrepaired sstables during incremental repair

2015-01-26 Thread Flavien Charlon (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-8267?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14291694#comment-14291694
 ] 

Flavien Charlon commented on CASSANDRA-8267:


Won't this be part of 2.1.3? I hope it will.

 Only stream from unrepaired sstables during incremental repair
 --

 Key: CASSANDRA-8267
 URL: https://issues.apache.org/jira/browse/CASSANDRA-8267
 Project: Cassandra
  Issue Type: Bug
Reporter: Marcus Eriksson
Assignee: Marcus Eriksson
 Fix For: 3.0

 Attachments: 
 0001-Only-stream-from-unrepaired-sstables-during-incremen.patch, 
 8267-trunk.patch


 Seems we stream from all sstables even if we do incremental repair, we should 
 limit this to only stream from the unrepaired sstables if we do incremental 
 repair



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (CASSANDRA-8641) Repair causes a large number of tiny SSTables

2015-01-20 Thread Flavien Charlon (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-8641?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14284160#comment-14284160
 ] 

Flavien Charlon commented on CASSANDRA-8641:


Memory pressure with 32 GB of RAM? If that's the case, then you should probably 
look into making some optimizations. That's a bug either way.

In the meantime we have disabled vnodes as they don't seem to be production 
ready.

 Repair causes a large number of tiny SSTables
 -

 Key: CASSANDRA-8641
 URL: https://issues.apache.org/jira/browse/CASSANDRA-8641
 Project: Cassandra
  Issue Type: Bug
  Components: Core
 Environment: Ubuntu 14.04
Reporter: Flavien Charlon
 Fix For: 2.1.3


 I have a 3 nodes cluster with RF = 3, quad core and 32 GB or RAM. I am 
 running 2.1.2 with all the default settings. I'm seeing some strange 
 behaviors during incremental repair (under write load).
 Taking the example of one particular column family, before running an 
 incremental repair, I have about 13 SSTables. After finishing the incremental 
 repair, I have over 114000 SSTables.
 {noformat}
 Table: customers
 SSTable count: 114688
 Space used (live): 97203707290
 Space used (total): 99175455072
 Space used by snapshots (total): 0
 SSTable Compression Ratio: 0.28281112416526505
 Memtable cell count: 0
 Memtable data size: 0
 Memtable switch count: 1069
 Local read count: 0
 Local read latency: NaN ms
 Local write count: 11548705
 Local write latency: 0.030 ms
 Pending flushes: 0
 Bloom filter false positives: 0
 Bloom filter false ratio: 0.0
 Bloom filter space used: 144145152
 Compacted partition minimum bytes: 311
 Compacted partition maximum bytes: 1996099046
 Compacted partition mean bytes: 3419
 Average live cells per slice (last five minutes): 0.0
 Maximum live cells per slice (last five minutes): 0.0
 Average tombstones per slice (last five minutes): 0.0
 Maximum tombstones per slice (last five minutes): 0.0
 {noformat}
 Looking at the logs during the repair, it seems Cassandra is struggling to 
 compact minuscule memtables (often just a few kilobytes):
 {noformat}
 INFO  [CompactionExecutor:337] 2015-01-17 01:44:27,011 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-228341,].
   8,332 bytes to 6,547 (~78% of original) in 80,476ms = 0.78MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 INFO  [CompactionExecutor:337] 2015-01-17 01:45:35,519 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229348,].
   8,384 bytes to 6,563 (~78% of original) in 6,880ms = 0.000910MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 INFO  [CompactionExecutor:339] 2015-01-17 01:47:46,475 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229351,].
   8,423 bytes to 6,401 (~75% of original) in 10,416ms = 0.000586MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 {noformat}
  
 Here is an excerpt of the system logs showing the abnormal flushing:
 {noformat}
 INFO  [AntiEntropyStage:1] 2015-01-17 15:28:43,807 ColumnFamilyStore.java:840 
 - Enqueuing flush of customers: 634484 (0%) on-heap, 2599489 (0%) off-heap
 INFO  [AntiEntropyStage:1] 2015-01-17 15:29:06,823 ColumnFamilyStore.java:840 
 - Enqueuing flush of levels: 129504 (0%) on-heap, 222168 (0%) off-heap
 INFO  [AntiEntropyStage:1] 2015-01-17 15:29:07,940 ColumnFamilyStore.java:840 
 - Enqueuing flush of chain: 4508 (0%) on-heap, 6880 (0%) off-heap
 INFO  [AntiEntropyStage:1] 2015-01-17 15:29:08,124 ColumnFamilyStore.java:840 
 - Enqueuing flush of invoices: 1469772 (0%) on-heap, 2542675 (0%) off-heap
 INFO  [AntiEntropyStage:1] 2015-01-17 15:29:09,471 ColumnFamilyStore.java:840 
 - Enqueuing flush of customers: 809844 (0%) on-heap, 3364728 (0%) off-heap
 INFO  [AntiEntropyStage:1] 2015-01-17 15:29:24,368 ColumnFamilyStore.java:840 
 - Enqueuing flush of levels: 28212 (0%) on-heap, 44220 (0%) off-heap
 INFO  [AntiEntropyStage:1] 2015-01-17 15:29:24,822 ColumnFamilyStore.java:840 
 - Enqueuing flush of chain: 860 (0%) on-heap, 1130 (0%) off-heap
 INFO  [AntiEntropyStage:1] 2015-01-17 15:29:24,985 ColumnFamilyStore.java:840 
 - Enqueuing flush of invoices: 334480 (0%) on-heap, 568959 (0%) off-heap
 INFO  [AntiEntropyStage:1] 2015-01-17 15:29:27,375 ColumnFamilyStore.java:840 
 - Enqueuing flush of customers: 221568 (0%) on-heap, 929962 (0%) off-heap
 INFO  [AntiEntropyStage:1] 2015-01-17 15:29:35,755 ColumnFamilyStore.java:840 
 - Enqueuing flush of invoices: 7916 (0%) 

[jira] [Updated] (CASSANDRA-8641) Repair causes a large number of tiny SSTables

2015-01-18 Thread Flavien Charlon (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-8641?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flavien Charlon updated CASSANDRA-8641:
---
Description: 
I have a 3 nodes cluster with RF = 3, quad core and 32 GB or RAM. I am running 
2.1.2 with all the default settings. I'm seeing some strange behaviors during 
incremental repair (under write load).

Taking the example of one particular column family, before running an 
incremental repair, I have about 13 SSTables. After finishing the incremental 
repair, I have over 114000 SSTables.

{noformat}
Table: customers
SSTable count: 114688
Space used (live): 97203707290
Space used (total): 99175455072
Space used by snapshots (total): 0
SSTable Compression Ratio: 0.28281112416526505
Memtable cell count: 0
Memtable data size: 0
Memtable switch count: 1069
Local read count: 0
Local read latency: NaN ms
Local write count: 11548705
Local write latency: 0.030 ms
Pending flushes: 0
Bloom filter false positives: 0
Bloom filter false ratio: 0.0
Bloom filter space used: 144145152
Compacted partition minimum bytes: 311
Compacted partition maximum bytes: 1996099046
Compacted partition mean bytes: 3419
Average live cells per slice (last five minutes): 0.0
Maximum live cells per slice (last five minutes): 0.0
Average tombstones per slice (last five minutes): 0.0
Maximum tombstones per slice (last five minutes): 0.0
{noformat}

Looking at the logs during the repair, it seems Cassandra is struggling to 
compact minuscule memtables (often just a few kilobytes):

{noformat}
INFO  [CompactionExecutor:337] 2015-01-17 01:44:27,011 CompactionTask.java:251 
- Compacted 32 sstables to 
[/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-228341,].
  8,332 bytes to 6,547 (~78% of original) in 80,476ms = 0.78MB/s.  32 total 
partitions merged to 32.  Partition merge counts were {1:32, }
INFO  [CompactionExecutor:337] 2015-01-17 01:45:35,519 CompactionTask.java:251 
- Compacted 32 sstables to 
[/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229348,].
  8,384 bytes to 6,563 (~78% of original) in 6,880ms = 0.000910MB/s.  32 total 
partitions merged to 32.  Partition merge counts were {1:32, }
INFO  [CompactionExecutor:339] 2015-01-17 01:47:46,475 CompactionTask.java:251 
- Compacted 32 sstables to 
[/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229351,].
  8,423 bytes to 6,401 (~75% of original) in 10,416ms = 0.000586MB/s.  32 total 
partitions merged to 32.  Partition merge counts were {1:32, }
{noformat}
 
Here is an excerpt of the system logs showing the abnormal flushing:

{noformat}
INFO  [AntiEntropyStage:1] 2015-01-17 15:28:43,807 ColumnFamilyStore.java:840 - 
Enqueuing flush of customers: 634484 (0%) on-heap, 2599489 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:06,823 ColumnFamilyStore.java:840 - 
Enqueuing flush of levels: 129504 (0%) on-heap, 222168 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:07,940 ColumnFamilyStore.java:840 - 
Enqueuing flush of chain: 4508 (0%) on-heap, 6880 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:08,124 ColumnFamilyStore.java:840 - 
Enqueuing flush of invoices: 1469772 (0%) on-heap, 2542675 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:09,471 ColumnFamilyStore.java:840 - 
Enqueuing flush of customers: 809844 (0%) on-heap, 3364728 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:24,368 ColumnFamilyStore.java:840 - 
Enqueuing flush of levels: 28212 (0%) on-heap, 44220 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:24,822 ColumnFamilyStore.java:840 - 
Enqueuing flush of chain: 860 (0%) on-heap, 1130 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:24,985 ColumnFamilyStore.java:840 - 
Enqueuing flush of invoices: 334480 (0%) on-heap, 568959 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:27,375 ColumnFamilyStore.java:840 - 
Enqueuing flush of customers: 221568 (0%) on-heap, 929962 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:35,755 ColumnFamilyStore.java:840 - 
Enqueuing flush of invoices: 7916 (0%) on-heap, 11080 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:36,239 ColumnFamilyStore.java:840 - 
Enqueuing flush of customers: 9968 (0%) on-heap, 33041 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:37,935 ColumnFamilyStore.java:840 - 
Enqueuing flush of invoices: 42108 (0%) on-heap, 69494 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:41,182 ColumnFamilyStore.java:840 - 
Enqueuing flush of customers: 40936 (0%) on-heap, 159099 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:49,573 ColumnFamilyStore.java:840 - 
Enqueuing flush of levels: 17236 (0%) on-heap, 27048 (0%) off-heap
INFO  [AntiEntropyStage:1] 2015-01-17 15:29:50,440 ColumnFamilyStore.java:840 - 
Enqueuing flush of chain: 548 

[jira] [Commented] (CASSANDRA-8641) Repair causes a large number of tiny SSTables

2015-01-17 Thread Flavien Charlon (JIRA)

[ 
https://issues.apache.org/jira/browse/CASSANDRA-8641?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14281407#comment-14281407
 ] 

Flavien Charlon commented on CASSANDRA-8641:


I have rebuilt the cluster, and repeated the same setup (constant insert load), 
while doing a non-incremental, non-parallel repair, and the same thing happens: 
tons of flushes, creating hundreds of SSTables on disk.

Then I did it again, with the insert load only (no repair), and everything is 
normal: rare flushes, and the number of SSTables remains stable.

So this is definitely triggered by repair in some way.

 Repair causes a large number of tiny SSTables
 -

 Key: CASSANDRA-8641
 URL: https://issues.apache.org/jira/browse/CASSANDRA-8641
 Project: Cassandra
  Issue Type: Bug
  Components: Core
 Environment: Ubuntu 14.04
Reporter: Flavien Charlon
 Fix For: 2.1.3


 I have a 3 nodes cluster with RF = 3, quad core and 32 GB or RAM. I am 
 running 2.1.2 with all the default settings. I'm seeing some strange 
 behaviors during incremental repair (under write load).
 Taking the example of one particular column family, before running an 
 incremental repair, I have about 13 SSTables. After finishing the incremental 
 repair, I have over 114000 SSTables.
 {noformat}
 Table: customers
 SSTable count: 114688
 Space used (live): 97203707290
 Space used (total): 99175455072
 Space used by snapshots (total): 0
 SSTable Compression Ratio: 0.28281112416526505
 Memtable cell count: 0
 Memtable data size: 0
 Memtable switch count: 1069
 Local read count: 0
 Local read latency: NaN ms
 Local write count: 11548705
 Local write latency: 0.030 ms
 Pending flushes: 0
 Bloom filter false positives: 0
 Bloom filter false ratio: 0.0
 Bloom filter space used: 144145152
 Compacted partition minimum bytes: 311
 Compacted partition maximum bytes: 1996099046
 Compacted partition mean bytes: 3419
 Average live cells per slice (last five minutes): 0.0
 Maximum live cells per slice (last five minutes): 0.0
 Average tombstones per slice (last five minutes): 0.0
 Maximum tombstones per slice (last five minutes): 0.0
 {noformat}
 Looking at the logs during the repair, it seems Cassandra is struggling to 
 compact minuscule memtables (often just a few kilobytes):
 {noformat}
 INFO  [CompactionExecutor:337] 2015-01-17 01:44:27,011 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-228341,].
   8,332 bytes to 6,547 (~78% of original) in 80,476ms = 0.78MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 INFO  [CompactionExecutor:337] 2015-01-17 01:45:35,519 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229348,].
   8,384 bytes to 6,563 (~78% of original) in 6,880ms = 0.000910MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 INFO  [CompactionExecutor:339] 2015-01-17 01:47:46,475 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229351,].
   8,423 bytes to 6,401 (~75% of original) in 10,416ms = 0.000586MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 {noformat}
  
 At the end of the repair, the cluster has become unusable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-8641) Repair causes a large number of tiny SSTables

2015-01-16 Thread Flavien Charlon (JIRA)
Flavien Charlon created CASSANDRA-8641:
--

 Summary: Repair causes a large number of tiny SSTables
 Key: CASSANDRA-8641
 URL: https://issues.apache.org/jira/browse/CASSANDRA-8641
 Project: Cassandra
  Issue Type: Bug
  Components: Core
 Environment: Ubuntu 14.04
Reporter: Flavien Charlon


I have a 3 nodes cluster with RF = 3, quad core and 32 GB or RAM. I am running 
2.1.2 with all the default settings. I'm seeing some strange behaviors during 
incremental repair (under write load).

Taking the example of one particular column family, before running an 
incremental repair, I have about 13 SSTables. After finishing the incremental 
repair, I have over 114000 SSTables.

{noformat}
Table: customers
SSTable count: 114688
Space used (live): 97203707290
Space used (total): 99175455072
Space used by snapshots (total): 0
SSTable Compression Ratio: 0.28281112416526505
Memtable cell count: 0
Memtable data size: 0
Memtable switch count: 1069
Local read count: 0
Local read latency: NaN ms
Local write count: 11548705
Local write latency: 0.030 ms
Pending flushes: 0
Bloom filter false positives: 0
Bloom filter false ratio: 0.0
Bloom filter space used: 144145152
Compacted partition minimum bytes: 311
Compacted partition maximum bytes: 1996099046
Compacted partition mean bytes: 3419
Average live cells per slice (last five minutes): 0.0
Maximum live cells per slice (last five minutes): 0.0
Average tombstones per slice (last five minutes): 0.0
Maximum tombstones per slice (last five minutes): 0.0
{noformat}

Looking at the logs during the repair, it seems Cassandra is struggling to 
compact minuscule memtables (often just a few kilobytes):

{noformat}
INFO  [CompactionExecutor:337] 2015-01-17 01:44:27,011 CompactionTask.java:251 
- Compacted 32 sstables to 
[/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-228341,].
  8,332 bytes to 6,547 (~78% of original) in 80,476ms = 0.78MB/s.  32 total 
partitions merged to 32.  Partition merge counts were {1:32, }
INFO  [CompactionExecutor:337] 2015-01-17 01:45:35,519 CompactionTask.java:251 
- Compacted 32 sstables to 
[/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229348,].
  8,384 bytes to 6,563 (~78% of original) in 6,880ms = 0.000910MB/s.  32 total 
partitions merged to 32.  Partition merge counts were {1:32, }
INFO  [CompactionExecutor:339] 2015-01-17 01:47:46,475 CompactionTask.java:251 
- Compacted 32 sstables to 
[/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229351,].
  8,423 bytes to 6,401 (~75% of original) in 10,416ms = 0.000586MB/s.  32 total 
partitions merged to 32.  Partition merge counts were {1:32, }
{noformat}
 
At the end of the repair, the cluster has become unusable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-8641) Repair causes a large number of tiny SSTables

2015-01-16 Thread Flavien Charlon (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-8641?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flavien Charlon updated CASSANDRA-8641:
---
Since Version:   (was: 2.1.2)

 Repair causes a large number of tiny SSTables
 -

 Key: CASSANDRA-8641
 URL: https://issues.apache.org/jira/browse/CASSANDRA-8641
 Project: Cassandra
  Issue Type: Bug
  Components: Core
 Environment: Ubuntu 14.04
Reporter: Flavien Charlon

 I have a 3 nodes cluster with RF = 3, quad core and 32 GB or RAM. I am 
 running 2.1.2 with all the default settings. I'm seeing some strange 
 behaviors during incremental repair (under write load).
 Taking the example of one particular column family, before running an 
 incremental repair, I have about 13 SSTables. After finishing the incremental 
 repair, I have over 114000 SSTables.
 {noformat}
 Table: customers
 SSTable count: 114688
 Space used (live): 97203707290
 Space used (total): 99175455072
 Space used by snapshots (total): 0
 SSTable Compression Ratio: 0.28281112416526505
 Memtable cell count: 0
 Memtable data size: 0
 Memtable switch count: 1069
 Local read count: 0
 Local read latency: NaN ms
 Local write count: 11548705
 Local write latency: 0.030 ms
 Pending flushes: 0
 Bloom filter false positives: 0
 Bloom filter false ratio: 0.0
 Bloom filter space used: 144145152
 Compacted partition minimum bytes: 311
 Compacted partition maximum bytes: 1996099046
 Compacted partition mean bytes: 3419
 Average live cells per slice (last five minutes): 0.0
 Maximum live cells per slice (last five minutes): 0.0
 Average tombstones per slice (last five minutes): 0.0
 Maximum tombstones per slice (last five minutes): 0.0
 {noformat}
 Looking at the logs during the repair, it seems Cassandra is struggling to 
 compact minuscule memtables (often just a few kilobytes):
 {noformat}
 INFO  [CompactionExecutor:337] 2015-01-17 01:44:27,011 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-228341,].
   8,332 bytes to 6,547 (~78% of original) in 80,476ms = 0.78MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 INFO  [CompactionExecutor:337] 2015-01-17 01:45:35,519 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229348,].
   8,384 bytes to 6,563 (~78% of original) in 6,880ms = 0.000910MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 INFO  [CompactionExecutor:339] 2015-01-17 01:47:46,475 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229351,].
   8,423 bytes to 6,401 (~75% of original) in 10,416ms = 0.000586MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 {noformat}
  
 At the end of the repair, the cluster has become unusable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-8641) Repair causes a large number of tiny SSTables

2015-01-16 Thread Flavien Charlon (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-8641?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flavien Charlon updated CASSANDRA-8641:
---
Fix Version/s: 2.1.3

 Repair causes a large number of tiny SSTables
 -

 Key: CASSANDRA-8641
 URL: https://issues.apache.org/jira/browse/CASSANDRA-8641
 Project: Cassandra
  Issue Type: Bug
  Components: Core
 Environment: Ubuntu 14.04
Reporter: Flavien Charlon
 Fix For: 2.1.3


 I have a 3 nodes cluster with RF = 3, quad core and 32 GB or RAM. I am 
 running 2.1.2 with all the default settings. I'm seeing some strange 
 behaviors during incremental repair (under write load).
 Taking the example of one particular column family, before running an 
 incremental repair, I have about 13 SSTables. After finishing the incremental 
 repair, I have over 114000 SSTables.
 {noformat}
 Table: customers
 SSTable count: 114688
 Space used (live): 97203707290
 Space used (total): 99175455072
 Space used by snapshots (total): 0
 SSTable Compression Ratio: 0.28281112416526505
 Memtable cell count: 0
 Memtable data size: 0
 Memtable switch count: 1069
 Local read count: 0
 Local read latency: NaN ms
 Local write count: 11548705
 Local write latency: 0.030 ms
 Pending flushes: 0
 Bloom filter false positives: 0
 Bloom filter false ratio: 0.0
 Bloom filter space used: 144145152
 Compacted partition minimum bytes: 311
 Compacted partition maximum bytes: 1996099046
 Compacted partition mean bytes: 3419
 Average live cells per slice (last five minutes): 0.0
 Maximum live cells per slice (last five minutes): 0.0
 Average tombstones per slice (last five minutes): 0.0
 Maximum tombstones per slice (last five minutes): 0.0
 {noformat}
 Looking at the logs during the repair, it seems Cassandra is struggling to 
 compact minuscule memtables (often just a few kilobytes):
 {noformat}
 INFO  [CompactionExecutor:337] 2015-01-17 01:44:27,011 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-228341,].
   8,332 bytes to 6,547 (~78% of original) in 80,476ms = 0.78MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 INFO  [CompactionExecutor:337] 2015-01-17 01:45:35,519 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229348,].
   8,384 bytes to 6,563 (~78% of original) in 6,880ms = 0.000910MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 INFO  [CompactionExecutor:339] 2015-01-17 01:47:46,475 
 CompactionTask.java:251 - Compacted 32 sstables to 
 [/mnt/data/cassandra/data/business/customers-d9d42d209ccc11e48ca54553c90a9d45/business-customers-ka-229351,].
   8,423 bytes to 6,401 (~75% of original) in 10,416ms = 0.000586MB/s.  32 
 total partitions merged to 32.  Partition merge counts were {1:32, }
 {noformat}
  
 At the end of the repair, the cluster has become unusable.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-8529) Cassandra suddenly stops responding to clients though process is still running

2014-12-20 Thread Flavien Charlon (JIRA)
Flavien Charlon created CASSANDRA-8529:
--

 Summary: Cassandra suddenly stops responding to clients though 
process is still running
 Key: CASSANDRA-8529
 URL: https://issues.apache.org/jira/browse/CASSANDRA-8529
 Project: Cassandra
  Issue Type: Bug
  Components: Core
Reporter: Flavien Charlon


I am running a moderate write-only load onto a 3 nodes cluster.

After some time, nodes become completely unresponsive to clients, even though 
the process is still running.

tpstats on affected nodes indicate pending compaction, which never gets 
executued.
This is tpstats on the affected node hours after the load has stopped:

Pool NameActive   Pending  Completed   Blocked  All 
time blocked
CounterMutationStage  0 0  0 0  
   0
ReadStage 0 0 243384 0  
   0
RequestResponseStage  0 03336833 0  
   0
MutationStage32  19024775909 0  
   0
ReadRepairStage   0 0  14445 0  
   0
GossipStage   0 0 128499 0  
   0
CacheCleanupExecutor  0 0  0 0  
   0
AntiEntropyStage  0 0  0 0  
   0
MigrationStage0 0 36 0  
   0
ValidationExecutor0 0  0 0  
   0
CommitLogArchiver 0 0  0 0  
   0
MiscStage 0 0  0 0  
   0
MemtableFlushWriter   2 7947 0  
   0
MemtableReclaimMemory 0 0947 0  
   0
PendingRangeCalculator0 0  5 0  
   0
MemtablePostFlush 1 8   1241 0  
   0
CompactionExecutor2 8   1035 0  
   0
InternalResponseStage 0 0  9 0  
   0
HintedHandoff 0 0  6 0  
   0

Message type   Dropped
RANGE_SLICE  0
READ_REPAIR  0
PAGED_RANGE  0
BINARY   0
READ 0
MUTATION 0
_TRACE   0
REQUEST_RESPONSE 0
COUNTER_MUTATION 0

Also, compactionstats shows that compaction is stalled:

   compaction type   keyspace  table   completed   totalunit   
progress
Compactiontestnet   transactions   117833347   117834891   bytes
100.00%
Compactiontestnetscripts   206418064   206419414   bytes
100.00%
Active compaction remaining time :   0h00m00s

And again, this has been like this for hours.

I have reproduced this on several clusters with various memory configurations.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (CASSANDRA-8502) Static columns returning null when paging is used

2014-12-17 Thread Flavien Charlon (JIRA)
Flavien Charlon created CASSANDRA-8502:
--

 Summary: Static columns returning null when paging is used
 Key: CASSANDRA-8502
 URL: https://issues.apache.org/jira/browse/CASSANDRA-8502
 Project: Cassandra
  Issue Type: Bug
  Components: Core
Reporter: Flavien Charlon
 Attachments: null-static-column.txt

When paging is used for a query containing a static column, the first page 
contains the right value for the static column, but subsequent pages have null 
null for the static column instead of the expected value.

Repro steps:

- Create a table with a static column
- Create a partition with 500 cells
- Using cqlsh, query that partition

Actual result:
- You will see that first, the static column appears as expected, but if you 
press a key after ---MORE---, the static columns will appear as null.

See the attached file for a repro of the output.

I am using a single node cluster.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (CASSANDRA-8502) Static columns returning null for pages after first

2014-12-17 Thread Flavien Charlon (JIRA)

 [ 
https://issues.apache.org/jira/browse/CASSANDRA-8502?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flavien Charlon updated CASSANDRA-8502:
---
Summary: Static columns returning null for pages after first  (was: Static 
columns returning null when paging is used)

 Static columns returning null for pages after first
 ---

 Key: CASSANDRA-8502
 URL: https://issues.apache.org/jira/browse/CASSANDRA-8502
 Project: Cassandra
  Issue Type: Bug
  Components: Core
Reporter: Flavien Charlon
 Attachments: null-static-column.txt


 When paging is used for a query containing a static column, the first page 
 contains the right value for the static column, but subsequent pages have 
 null null for the static column instead of the expected value.
 Repro steps:
 - Create a table with a static column
 - Create a partition with 500 cells
 - Using cqlsh, query that partition
 Actual result:
 - You will see that first, the static column appears as expected, but if you 
 press a key after ---MORE---, the static columns will appear as null.
 See the attached file for a repro of the output.
 I am using a single node cluster.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)