I've seen several folks build their own storage clusters and use Nexenta. We use EMC VMAX.
> -----Original Message----- > From: David Nalley [mailto:da...@gnsa.us] > Sent: Thursday, March 14, 2013 12:52 PM > To: cloudstack-users@incubator.apache.org > Subject: Re: What's everyone using for primary storage? > > On Thu, Mar 14, 2013 at 11:27 AM, Kirk Jantzer <kirk.jant...@gmail.com> > wrote: > > For my testing, because of available resources, I chose local storage. > > However, shared storage (like GlusterFS or the like) is appealing. > > What are you using? Why did you chose what you're using? How many > > instances are you running? etc. > > Thanks!! > > So I think it depends on workload and use case. > If your applications are highly fault tolerant and largely stateless, you > simply > may not care about HA. Besides you'll almost certainly get better > performance with local storage. > If you are running a test/dev environment, you can probably tolerate > instance failure, so why use shared storage. > If people are going to come scream at you and threaten to take money away > if something fails, perhaps you want something a bit more robust. > > The best of both worlds (with plenty of compromises too) is distributed, > replicated shared storage like Ceph RBD. (GlusterFS 3.4, with all of the work > that IBM has done around KVM is promising, but yet to be released. Early > versions were robust, but had problems at any scale providing decent IO) > Sheepdog is also promising and I keep hearing there are patches incoming for > Sheepdog support. Of course these are all KVM-only for the moment. > > There are also plenty of people who do both local and shared storage. > With higher internal costs for deploying to shared storage with the > assumption that folks would do it for things that need a higher level of > resiliency or less tolerance for failure. > > For shared storage, I've seen everything from NFS running on Linux to things > like Isilon, Netapp, and EMC - again the choice depending on the tolerance > for failure. > > --David