Number of cores: 2x6Cores x 2(HT). I do agree with you that the the hardware is certainly overestimated for just one Cassandra, but we got a very good price since we ordered several 10s of the same nodes for a different project. That's why we use for multiple cassandra instances.
Jirka H. On 02/12/2015 04:18 PM, Eric Stevens wrote: > > each node has 256G of memory, 24x1T drives, 2x Xeon CPU > > I don't have first hand experience running Cassandra on such massive > hardware, but it strikes me that these machines are dramatically > oversized to be good candidates for Cassandra (though I wonder how > many cores are in those CPUs; I'm guessing closer to 18 than 2 based > on the other hardware). > > A larger cluster of smaller hardware would be a much better shape for > Cassandra. Or several clusters of smaller hardware since you're > running multiple instances on this hardware - best practices have one > instance per host no matter the hardware size. > > On Thu, Feb 12, 2015 at 12:36 AM, Jiri Horky <ho...@avast.com > <mailto:ho...@avast.com>> wrote: > > Hi Chris, > > On 02/09/2015 04:22 PM, Chris Lohfink wrote: >> - number of tombstones - how can I reliably find it out? >> https://github.com/spotify/cassandra-opstools >> https://github.com/cloudian/support-tools > thanks. >> >> If not getting much compression it may be worth trying to disable >> it, it may contribute but its very unlikely that its the cause of >> the gc pressure itself. >> >> 7000 sstables but STCS? Sounds like compactions couldn't keep >> up. Do you have a lot of pending compactions (nodetool)? You >> may want to increase your compaction throughput (nodetool) to see >> if you can catch up a little, it would cause a lot of heap >> overhead to do reads with that many. May even need to take more >> drastic measures if it cant catch back up. > I am sorry, I was wrong. We actually do use LCS (the switch was > done recently). There are almost none pending compaction. We have > increased the size sstable to 768M, so it should help as as well. > >> >> May also be good to check `nodetool cfstats` for very wide >> partitions. > There are basically none, this is fine. > > It seems that the problem really comes from having so much data in > so many sstables, so > org.apache.cassandra.io.compress.CompressedRandomAccessReader > classes consumes more memory than 0.75*HEAP_SIZE, which triggers > the CMS over and over. > > We have turned off the compression and so far, the situation seems > to be fine. > > Cheers > Jirka H. > > >> >> Theres a good chance if under load and you have over 8gb heap >> your GCs could use tuning. The bigger the nodes the more manual >> tweaking it will require to get the most out of >> them https://issues.apache.org/jira/browse/CASSANDRA-8150 also >> has some ideas. >> >> Chris >> >> On Mon, Feb 9, 2015 at 2:00 AM, Jiri Horky <ho...@avast.com >> <mailto:ho...@avast.com>> wrote: >> >> Hi all, >> >> thank you all for the info. >> >> To answer the questions: >> - we have 2 DCs with 5 nodes in each, each node has 256G of >> memory, 24x1T drives, 2x Xeon CPU - there are multiple >> cassandra instances running for different project. The node >> itself is powerful enough. >> - there 2 keyspaces, one with 3 replicas per DC, one with 1 >> replica per DC (because of amount of data and because it >> serves more or less like a cache) >> - there are about 4k/s Request-response, 3k/s Read and 2k/s >> Mutation requests - numbers are sum of all nodes >> - we us STCS (LCS would be quite IO have for this amount of >> data) >> - number of tombstones - how can I reliably find it out? >> - the biggest CF (3.6T per node) has 7000 sstables >> >> Now, I understand that the best practice for Cassandra is to >> run "with the minimum size of heap which is enough" which for >> this case we thought is about 12G - there is always 8G >> consumbed by the SSTable readers. Also, I though that high >> number of tombstones create pressure in the new space (which >> can then cause pressure in old space as well), but this is >> not what we are seeing. We see continuous GC activity in Old >> generation only. >> >> Also, I noticed that the biggest CF has Compression factor of >> 0.99 which basically means that the data come compressed >> already. Do you think that turning off the compression should >> help with memory consumption? >> >> Also, I think that tuning CMSInitiatingOccupancyFraction=75 >> might help here, as it seems that 8G is something that >> Cassandra needs for bookkeeping this amount of data and that >> this was sligtly above the 75% limit which triggered the CMS >> again and again. >> >> I will definitely have a look at the presentation. >> >> Regards >> Jiri Horky >> >> >> On 02/08/2015 10:32 PM, Mark Reddy wrote: >>> Hey Jiri, >>> >>> While I don't have any experience running 4TB nodes (yet), I >>> would recommend taking a look at a presentation by Arron >>> Morton on large >>> nodes: >>> http://planetcassandra.org/blog/cassandra-community-webinar-videoslides-large-nodes-with-cassandra-by-aaron-morton/ >>> to see if you can glean anything from that. >>> >>> I would note that at the start of his talk he mentions that >>> in version 1.2 we can now talk about nodes around 1 - 3 TB >>> in size, so if you are storing anything more than that you >>> are getting into very specialised use cases. >>> >>> If you could provide us with some more information about >>> your cluster setup (No. of CFs, read/write patterns, do you >>> delete / update often, etc.) that may help in getting you to >>> a better place. >>> >>> >>> Regards, >>> Mark >>> >>> On 8 February 2015 at 21:10, Kevin Burton >>> <bur...@spinn3r.com <mailto:bur...@spinn3r.com>> wrote: >>> >>> Do you have a lot of individual tables? Or lots of >>> small compactions? >>> >>> I think the general consensus is that (at least for >>> Cassandra), 8GB heaps are ideal. >>> >>> If you have lots of small tables it’s a known >>> anti-pattern (I believe) because the Cassandra internals >>> could do a better job on handling the in memory metadata >>> representation. >>> >>> I think this has been improved in 2.0 and 2.1 though so >>> the fact that you’re on 1.2.18 could exasperate the >>> issue. You might want to consider an upgrade (though >>> that has its own issues as well). >>> >>> On Sun, Feb 8, 2015 at 12:44 PM, Jiri Horky >>> <ho...@avast.com <mailto:ho...@avast.com>> wrote: >>> >>> Hi all, >>> >>> we are seeing quite high GC pressure (in old space >>> by CMS GC Algorithm) >>> on a node with 4TB of data. It runs C* 1.2.18 with >>> 12G of heap memory >>> (2G for new space). The node runs fine for couple of >>> days when the GC >>> activity starts to raise and reaches about 15% of >>> the C* activity which >>> causes dropped messages and other problems. >>> >>> Taking a look at heap dump, there is about 8G used >>> by SSTableReader >>> classes in >>> >>> org.apache.cassandra.io.compress.CompressedRandomAccessReader. >>> >>> Is this something expected and we have just reached >>> the limit of how >>> many data a single Cassandra instance can handle or >>> it is possible to >>> tune it better? >>> >>> Regards >>> Jiri Horky >>> >>> >>> >>> >>> -- >>> Founder/CEO Spinn3r.com <http://Spinn3r.com> >>> Location: *San Francisco, CA* >>> blog:* *http://burtonator.wordpress.com >>> … or check out my Google+ profile >>> <https://plus.google.com/102718274791889610666/posts> >>> <http://spinn3r.com> >>> >>> >> >> > >