Re: [Marketing Mail] Re: Memory leak and lockup on our 2.2.7 Cassandra cluster.

2016-08-03 Thread Henrik Schröder
Have you tried using the G1 garbage collector instead of CMS? We had the same issues that things were normally fine, but as soon as something extraordinary happened, a node could go into GC hell and never recover, and that could then spread to other nodes as they took up the slack, trapping them

Re: Testing compaction strategies on a single production server?

2013-02-21 Thread Henrik Schröder
. And not the fun Disney kind. While it may be more work I personally would use one node in write survey to test LCS Cheers - Aaron Morton Freelance Cassandra Developer New Zealand @aaronmorton http://www.thelastpickle.com On 20/02/2013, at 6:28 AM, Henrik Schröder skro...@gmail.com

Testing compaction strategies on a single production server?

2013-02-19 Thread Henrik Schröder
Hey, Version 1.1 of Cassandra introduced live traffic sampling, which allows you to measure the performance of a node without it really joining the cluster: http://www.datastax.com/dev/blog/whats-new-in-cassandra-1-1-live-traffic-sampling That page mentions that you can change the compaction

Re: Testing compaction strategies on a single production server?

2013-02-19 Thread Henrik Schröder
copies. *From:* Henrik Schröder [mailto:skro...@gmail.com] *Sent:* Tuesday, February 19, 2013 15:57 *To:* user *Subject:* Testing compaction strategies on a single production server?*** * ** ** Hey, Version 1.1 of Cassandra introduced live traffic sampling, which allows you

Re: Replication Factor and Consistency Level Confusion

2012-12-20 Thread Henrik Schröder
Don't run with a replication factor of 2, use 3 instead, and do all reads and writes using quorum consistency. That way, if a single node is down, all your operations will complete. In fact, if every third node is down, you'll still be fine and able to handle all requests. However, if two

Re: Compact and Repair

2012-11-16 Thread Henrik Schröder
caused by major compaction. For maintenance it is better to have a set of small sstables then one big. Andrey On Thu, Nov 8, 2012 at 2:55 AM, Henrik Schröder skro...@gmail.com wrote: Hi, We recently ran a major compaction across our cluster, which reduced the storage used by about 50

Compact and Repair

2012-11-08 Thread Henrik Schröder
Hi, We recently ran a major compaction across our cluster, which reduced the storage used by about 50%. This is fine, since we do a lot of updates to existing data, so that's the expected result. The day after, we ran a full repair -pr across the cluster, and when that finished, each storage

Re: Compact and Repair

2012-11-08 Thread Henrik Schröder
are describing sounds like a lot to me. Does this increase happen due to repair entirely? Or was the load maybe increasing gradually over the week and you just checked for the first time? cheers, Christian On Thu, Nov 8, 2012 at 11:55 AM, Henrik Schröder skro...@gmail.comwrote: Hi, We

Re: Compact and Repair

2012-11-08 Thread Henrik Schröder
repaired last time ? 2012/11/8 Henrik Schröder skro...@gmail.com No, we're not using columns with TTL, and I performed a major compaction before the repair, so there shouldn't be vast amounts of tombstones moving around. And the increase happened during the repair, the nodes gained ~20

Re: Suggestions for tuning and monitoring.

2012-09-15 Thread Henrik Schröder
, at 2:05 AM, Henrik Schröder skro...@gmail.com wrote: Hi all, We're running a small Cassandra cluster (v1.0.10) serving data to our web application, and as our traffic grows, we're starting to see some weird issues. The biggest of these is that sometimes, a single node becomes unresponsive

Suggestions for tuning and monitoring.

2012-09-10 Thread Henrik Schröder
Hi all, We're running a small Cassandra cluster (v1.0.10) serving data to our web application, and as our traffic grows, we're starting to see some weird issues. The biggest of these is that sometimes, a single node becomes unresponsive. It's impossible to start new connections, or impossible to

Re: High commit size

2012-09-10 Thread Henrik Schröder
When we ran Cassandra on windows, we got better performance without memory mapped IO. We had the same problems your are describing, what happens is that Windows is rather aggressive about swapping out memory when all the memory is used, and it starts swapping out unused parts of the heap, which

Re: Decreasing the number of nodes in the ring

2012-08-27 Thread Henrik Schröder
Removetoken should only be used when removing a dead node from a cluster, it's a much slower and more expensive operation since it triggers a repair so that the remaining nodes can figure out which data they should now have. Decommission on the other hand is much simpler, the node that's being

Re: Nodes marked dead…. leap second?

2012-07-02 Thread Henrik Schröder
Bug: https://lkml.org/lkml/2012/6/30/122 Simple fix to reset the leap second flag: date; date `date +%m%d%H%M%C%y.%S`; date; /Henrik On Mon, Jul 2, 2012 at 1:56 PM, Jean Paul Adant jean.paul.ad...@gmail.comwrote: Hi, I did have the same problem with cassandra 1.1.1 on Ubuntu 11.10 I had

Random slow connects.

2012-06-14 Thread Henrik Schröder
Hi everyone, We have problem with our Cassandra cluster, and that is that sometimes it takes several seconds to open a new Thrift connection to the server. We've had this issue when we ran on windows, and we have this issue now that we run on Ubuntu. We've had it with our old networking setup,

Re: Random slow connects.

2012-06-14 Thread Henrik Schröder
connection pool which recycles connections every few hours and only waits a few seconds for a connection before timing out? /Henrik On Thu, Jun 14, 2012 at 4:54 PM, Mina Naguib mina.nag...@bloomdigital.comwrote: On 2012-06-14, at 10:38 AM, Henrik Schröder wrote: Hi everyone, We have

Migrating from a windows cluster to a linux cluster.

2012-05-24 Thread Henrik Schröder
Hey everyone, We're trying to migrate a cassandra cluster from a bunch of Windows machines to a bunch of (newer and more powerful) Linux machines. Our initial plan was to simply bootstrap the Linux servers into the cluster one by one, and then decommission the old servers one by one. However,

Re: Migrating from a windows cluster to a linux cluster.

2012-05-24 Thread Henrik Schröder
the datafiles, it should have all the data associated with that token, and on joining the cluster it should just pop in at the right place, but with a new ip address. And then we repeat that for each server. Will this work? Or is there a better way? /Henrik On Thu, May 24, 2012 at 7:41 PM, Henrik

Re: Migrating from a windows cluster to a linux cluster.

2012-05-24 Thread Henrik Schröder
On Thu, May 24, 2012 at 8:07 PM, Brandon Williams dri...@gmail.com wrote: Are there any other ways of doing the migration? What happens if we join the new servers without bootstrapping and run repair? Are there any other ugly hacks or workaround we can do? We're not looking to run a mixed

Re: Migrating from a windows cluster to a linux cluster.

2012-05-24 Thread Henrik Schröder
On Thu, May 24, 2012 at 9:28 PM, Brandon Williams dri...@gmail.com wrote: That sounds fine, with the caveat that you can't run sstableloader from a machine running Cassandra before 1.1, so copying the sstables manually (assuming both clusters are the same size and have the same tokens) might

Re: Question regarding major compaction.

2012-05-02 Thread Henrik Schröder
On Tue, May 1, 2012 at 6:07 PM, Rob Coli rc...@palominodb.com wrote: The primary differences, as I understand it, are that the index performance and bloom filter false positive rate for your One Big File are worse. First, you are more likely to get a bloom filter false positive due to the

Re: Question regarding major compaction.

2012-05-02 Thread Henrik Schröder
On Tue, May 1, 2012 at 9:06 PM, Edward Capriolo edlinuxg...@gmail.comwrote: Also there are some tickets in JIRA to impose a max sstable size and some other related optimizations that I think got stuck behind levelDB in coolness factor. Not every use case is good for leveled so adding more

Re: Question regarding major compaction.

2012-05-02 Thread Henrik Schröder
On Tue, May 1, 2012 at 10:00 PM, Oleg Proudnikov ol...@cloudorange.comwrote: There is this note regarding major compaction in the tuning guide: once you run a major compaction, automatic minor compactions are no longer triggered frequently forcing you to manually run major compactions on a

Re: Question regarding major compaction.

2012-05-01 Thread Henrik Schröder
But what's the difference between doing an extra read from that One Big File, than doing an extra read from whatever SSTable happen to be largest in the course of automatic minor compaction? We have a pretty update-heavy application, and doing a major compaction can remove up to 30% of the used

Re: cassandra and .net

2012-04-10 Thread Henrik Schröder
In your code you are using BufferedTransport, but in the Cassandra logs you're getting errors when it tries to use FramedTransport. If I remember correctly, BufferedTransport is gone, so you should only use FramedTransport. Like this: TTransport transport = new TFramedTransport(new TSocket(host,

Huge amount of empty files in data directory.

2012-03-01 Thread Henrik Schröder
seeing this behaviour in older versions of Cassandra, shouldn't it delete temp files while running? Is it possible to force it to delete temp files while running? Is this fixed in a later version? Or do we have to periodically restart servers to clean up the datadirectories? /Henrik Schröder

Re: Huge amount of empty files in data directory.

2012-03-01 Thread Henrik Schröder
Great, thanks! /Henrik On Thu, Mar 1, 2012 at 13:08, Sylvain Lebresne sylv...@datastax.com wrote: It's a bug, namely: https://issues.apache.org/jira/browse/CASSANDRA-3616 You'd want to upgrade. -- Sylvain On Thu, Mar 1, 2012 at 1:01 PM, Henrik Schröder skro...@gmail.com wrote: Hi

Re: List all keys with RandomPartitioner

2012-02-22 Thread Henrik Schröder
I had to port that piece of code to C#, and it's just a few lines of code, so just write your own. Here's the original so you can see what it does: http://git-wip-us.apache.org/repos/asf?p=cassandra.git;a=blob_plain;f=src/java/org/apache/cassandra/utils/FBUtilities.java;hb=refs/heads/trunk

Re: Moving existing nodes to a different network.

2011-12-11 Thread Henrik Schröder
I'm running Cassandra 1.0.1 if that makes any difference. /Henrik On Sun, Dec 11, 2011 at 13:16, Henrik Schröder skro...@gmail.com wrote: I have an existing cluster of four Cassandra nodes. The machines have both an internal and an external IP, and originally I set them up to use

Re: Cassandra 1.0.0 - Node Load Bug

2011-10-24 Thread Henrik Schröder
Cassandra fixes it. I hope this helps a little bit at least. /Henrik Schröder On Thu, Oct 20, 2011 at 18:53, Dan Hendry dan.hendry.j...@gmail.com wrote: I have been playing around with Cassandra 1.0.0 in our test environment it seems pretty sweet so far. I have however come across what appears

Re: Migrating all rows from 0.6.13 to 0.7.5 over thrift?

2011-05-06 Thread Henrik Schröder
Uhm, having a program that can talk to 0.6 and 0.7 servers at the same time is not the hard problem, it took way less than five minutes to copy in both generated clients in the same project and rename the C# namespaces. Two apps and write to disk inbetween? Maven? That's crazy talk. :-D What I

Re: Migrating all rows from 0.6.13 to 0.7.5 over thrift?

2011-05-06 Thread Henrik Schröder
I'll see if I can make some example broken files this weekend. /Henrik Schröder On Fri, May 6, 2011 at 02:10, aaron morton aa...@thelastpickle.com wrote: The difficulty is the different thrift clients between 0.6 and 0.7. If you want to roll your own solution I would consider: - write

Re: Migrating all rows from 0.6.13 to 0.7.5 over thrift?

2011-05-06 Thread Henrik Schröder
datafiles in 0.6 that are uncleanable in 0.7 so that you all can repeat this and hopefully fix it. /Henrik Schröder On Sat, May 7, 2011 at 00:35, Jeremy Hanna jeremy.hanna1...@gmail.comwrote: If you're able, go into the #cassandra channel on freenode (IRC) and talk to driftx or jbellis

Re: Unicode key encoding problem when upgrading from 0.6.13 to 0.7.5

2011-05-05 Thread Henrik Schröder
://issues.apache.org/jira/browse/CASSANDRA-2367 On May 3, 2011, at 5:15 PM, Henrik Schröder wrote: Hey everyone, We did some tests before upgrading our Cassandra cluster from 0.6 to 0.7, just to make sure that the change in how keys are encoded wouldn't cause us any dataloss. Unfortunately it seems that rows

Re: Unicode key encoding problem when upgrading from 0.6.13 to 0.7.5

2011-05-05 Thread Henrik Schröder
is still there after all. /Henrik Schröder

Re: Unicode key encoding problem when upgrading from 0.6.13 to 0.7.5

2011-05-05 Thread Henrik Schröder
, cleanup throws the error, etc. /Henrik Schröder On Thu, May 5, 2011 at 13:57, aaron morton aa...@thelastpickle.com wrote: The hard core way to fix the data is export to json with sstable2json, hand edit, and then json2sstable it back. Also to confirm, this only happens when data is written

Re: Unicode key encoding problem when upgrading from 0.6.13 to 0.7.5

2011-05-05 Thread Henrik Schröder
); } } The string key is new String(ByteBufferUtil.getArray(key.key), UTF-8) If you find one that you don't like just skip it. This way compaction goes through but obviously you'll loose data. On May 5, 2011, at 1:12 PM, Henrik Schröder wrote: Yeah, I've seen that one, and I'm

Migrating all rows from 0.6.13 to 0.7.5 over thrift?

2011-05-05 Thread Henrik Schröder
in assuming that if we use the consistencylevel ALL we'll get all rows? /Henrik Schröder

Re: Unicode key encoding problem when upgrading from 0.6.13 to 0.7.5

2011-05-04 Thread Henrik Schröder
with identical keys, except one of the rows doesn't really work? So, now what? And how did I end up in this state? /Henrik Schröder On Tue, May 3, 2011 at 22:10, aaron morton aa...@thelastpickle.com wrote: Can you provide some details of the data returned from you do the = get_range

Re: Unicode key encoding problem when upgrading from 0.6.13 to 0.7.5

2011-05-04 Thread Henrik Schröder
to figure out how to remove the rows that get corrupted by upgrading, hopefully it can be solved. /Henrik Schröder On Tue, May 3, 2011 at 17:19, Henrik Schröder skro...@gmail.com wrote: The way we solved this problem is that it turned out we had only a few hundred rows with unicode keys, so we

Unicode key encoding problem when upgrading from 0.6.13 to 0.7.5

2011-05-03 Thread Henrik Schröder
? Is there a way to avoid this and upgrade from 0.6 to 0.7 and not lose any rows? I would also really like to know which byte-array I should send in to get back that second row, there's gotta be some key that can be used to get it, the row is still there after all. /Henrik Schröder

Re: Unicode key encoding problem when upgrading from 0.6.13 to 0.7.5

2011-05-03 Thread Henrik Schröder
to be a problem in the future? Is there a chance that the good duplicate is cleaned out in favour of the bad duplicate so that we suddnely lose those rows again? /Henrik Schröder

Re: Range scan performance in 0.6.0 beta2

2010-03-29 Thread Henrik Schröder
On Fri, Mar 26, 2010 at 14:47, Jonathan Ellis jbel...@gmail.com wrote: On Fri, Mar 26, 2010 at 7:40 AM, Henrik Schröder skro...@gmail.com wrote: For each indexvalue we insert a row where the key is indexid + : + indexvalue encoded as hex string, and the row contains only one column, where

Re: Range scan performance in 0.6.0 beta2

2010-03-26 Thread Henrik Schröder
So all the values for an entire index will be in one row? That doesn't sound good. You really want to put each index [and each table] in its own CF, but until we can do that dynamically (0.7) you could at least make the index row keys a tuple of (indexid, indexvalue) and the column names

Re: Range scan performance in 0.6.0 beta2

2010-03-25 Thread Henrik Schröder
On Thu, Mar 25, 2010 at 15:17, Sylvain Lebresne sylv...@yakaz.com wrote: I don't know If that could play any role, but if ever you have disabled the assertions when running cassandra (that is, you removed the -ea line in cassandra.in.sh), there was a bug in 0.6beta2 that will make read in