I've been happy with glusterfs. for a sharedmountpoint kind of
storage. Works great.

(Note: invested in Infiniband to make things fast)

On Thu, Mar 14, 2013 at 11:51 AM, Musayev, Ilya <imusa...@webmd.net> wrote:
> I've seen several folks build their own storage clusters and use Nexenta.
>
> We use EMC VMAX.
>
>> -----Original Message-----
>> From: David Nalley [mailto:da...@gnsa.us]
>> Sent: Thursday, March 14, 2013 12:52 PM
>> To: cloudstack-users@incubator.apache.org
>> Subject: Re: What's everyone using for primary storage?
>>
>> On Thu, Mar 14, 2013 at 11:27 AM, Kirk Jantzer <kirk.jant...@gmail.com>
>> wrote:
>> > For my testing, because of available resources, I chose local storage.
>> > However, shared storage (like GlusterFS or the like) is appealing.
>> > What are you using? Why did you chose what you're using? How many
>> > instances are you running? etc.
>> > Thanks!!
>>
>> So I think it depends on workload and use case.
>> If your applications are highly fault tolerant and largely stateless, you 
>> simply
>> may not care about HA. Besides you'll almost certainly get better
>> performance with local storage.
>> If you are running a test/dev environment, you can probably tolerate
>> instance failure, so why use shared storage.
>> If people are going to come scream at you and threaten to take money away
>> if something fails, perhaps you want something a bit more robust.
>>
>> The best of both worlds (with plenty of compromises too) is distributed,
>> replicated shared storage like Ceph RBD. (GlusterFS 3.4, with all of the work
>> that IBM has done around KVM is promising, but yet to be released. Early
>> versions were robust, but had problems at any scale providing decent IO)
>> Sheepdog is also promising and I keep hearing there are patches incoming for
>> Sheepdog support. Of course these are all KVM-only for the moment.
>>
>> There are also plenty of people who do both local and shared storage.
>> With higher internal costs for deploying to shared storage with the
>> assumption that folks would do it for things that need a higher level of
>> resiliency or less tolerance for failure.
>>
>> For shared storage, I've seen everything from NFS running on Linux to things
>> like Isilon, Netapp, and EMC - again the choice depending on the tolerance
>> for failure.
>>
>> --David
>
>

Reply via email to