New question #217773 on Graphite: https://answers.launchpad.net/graphite/+question/217773
Is there a readily available method for calculating the storage requirements for a solution to be deployed on EC2? Suppose I had 1,000 instances and 10 counters. How would I figure out the storage requirements and IO requirements? Since it probably changes at 100,000 instances or 1000 counters, how does the formula change? Is there any guidance beyond try x and then tweak and try y? I realize that there is probably not a straightforward "you need an instance type of x with y GB available", but how can I estimate the correct size? Assume I want some level of detail on the 1,000 instances -- how would I estimate this storage? Adding more volumes isn't the issue so much as trying to figure out the IO requirements and ability to add/remove nodes, handle failure, etc. It's also very unclear to me how to distribute the data files. I think that this is what Ceres is planning on addressing, but everything I see on that version appears a bit dated,so I don't know if that is the right way to go and what the implications are of such a decision. Basically, given where the project is now, if you were planning on a potentially extremely large scale data set on EC2 (assume that an intermediate layer can do some level of near-real-time aggregation, the Graphite wouldn't be network bound and would get batched data), what would you do? -- You received this question notification because you are a member of graphite-dev, which is an answer contact for Graphite. _______________________________________________ Mailing list: https://launchpad.net/~graphite-dev Post to : graphite-dev@lists.launchpad.net Unsubscribe : https://launchpad.net/~graphite-dev More help : https://help.launchpad.net/ListHelp