Hi Paul, Concerning large rows which are not compacting, I've probably managed to reproduce your problem. I suppose you're using collections, but also TTLs ?
Anyway, I opened an issue here : https://issues.apache.org/jira/browse/CASSANDRA-5799 Hope this helps 2013/7/24 Christopher Wirt <chris.w...@struq.com> > Hi Paul,**** > > ** ** > > Sorry to hear you’re having a low point.**** > > ** ** > > We ended up not using the collection features of 1.2. **** > > Instead storing a compressed string containing the map and handling client > side.**** > > ** ** > > We only have fixed schema short rows so no experience with large row > compaction.**** > > ** ** > > File descriptors have never got that high for us. But, if you only have a > couple physical nodes with loads of data and small ss-tables maybe they > could get that high?**** > > ** ** > > Only time I’ve had file descriptors get out of hand was then compaction > got slightly confused with a new schema when I dropped and recreated > instead of truncating. > https://issues.apache.org/jira/browse/CASSANDRA-4857 restarting the node > fixed the issue.**** > > ** ** > > ** ** > > From my limited experience I think Cassandra is a dangerous choice for an > young limited funding/experience start-up expecting to scale fast. We are a > fairly mature start-up with funding. We’ve just spent 3-5 months moving > from Mongo to Cassandra. It’s been expensive and painful getting Cassandra > to read like Mongo, but we’ve made it J**** > > ** ** > > ** ** > > ** ** > > ** ** > > *From:* Paul Ingalls [mailto:paulinga...@gmail.com] > *Sent:* 24 July 2013 06:01 > *To:* user@cassandra.apache.org > *Subject:* disappointed**** > > ** ** > > I want to check in. I'm sad, mad and afraid. I've been trying to get a > 1.2 cluster up and working with my data set for three weeks with no > success. I've been running a 1.1 cluster for 8 months now with no hiccups, > but for me at least 1.2 has been a disaster. I had high hopes for > leveraging the new features of 1.2, specifically vnodes and collections. > But at this point I can't release my system into production, and will > probably need to find a new back end. As a small startup, this could be > catastrophic. I'm mostly mad at myself. I took a risk moving to the new > tech. I forgot sometimes when you gamble, you lose.**** > > ** ** > > First, the performance of 1.2.6 was horrible when using collections. I > wasn't able to push through 500k rows before the cluster became unusable. > With a lot of digging, and way too much time, I discovered I was hitting a > bug that had just been fixed, but was unreleased. This scared me, because > the release was already at 1.2.6 and I would have expected something as > https://issues.apache.org/jira/browse/CASSANDRA-5677 would have been > addressed long before. But gamely I grabbed the latest code from the 1.2 > branch, built it and I was finally able to get past half a million rows. > **** > > ** ** > > But, then I hit ~4 million rows, and a multitude of problems. Even with > the fix above, I was still seeing a ton of compactions failing, > specifically the ones for large rows. Not a single large row will compact, > they all assert with the wrong size. Worse, and this is what kills the > whole thing, I keep hitting a wall with open files, even after dumping the > whole DB, dropping vnodes and trying again. Seriously, 650k open file > descriptors? When it hits this limit, the whole DB craps out and is > basically unusable. This isn't that many rows. I have close to a half a > billion in 1.1…**** > > ** ** > > I'm now at a standstill. I figure I have two options unless someone here > can help me. Neither of them involve 1.2. I can either go back to 1.1 and > remove the features that collections added to my service, or I find another > data backend that has similar performance characteristics to cassandra but > allows collections type behavior in a scalable manner. Cause as far as I > can tell, 1.2 doesn't scale. Which makes me sad, I was proud of what I > accomplished with 1.1….**** > > ** ** > > Does anyone know why there are so many open file descriptors? Any ideas > on why a large row won't compact?**** > > ** ** > > Paul**** > -- Fabien Rousseau * * <aur...@yakaz.com>www.yakaz.com