On Thu, Dec 20, 2012 at 9:37 AM, JuanFra Rodriguez Cardoso < [email protected]> wrote:
> Yes, I'm really agree with Diego. > It would be a good choice for submitting a blueprint with this storage > feature based on tenants. > > According to current quotas control, it limits the: > > - > > Number of volumes which may be created > - > > Total size of all volumes within a project as measured in GB > - > > Number of instances which may be launched > - > > Number of processor cores which may be allocated > - Publicly accessible IP addresses > > > Another new feature related to shared storage we had thought about, it's > to include an option for choosing if an instance has to be replicated or > not, i.e. in a MooseFS scenario, to indicate goal (number of replicas). > It's useful for example in testing or demo projects, where HA is not > required. > > Regards, > > JuanFra. > > 2012/12/20 Diego Parrilla Santamaría <[email protected]> > >> mmm... not sure if the concept of oVirt multiple storage domains is >> something that can be implemented in Nova as it is, but I would like to >> share my thoughts because it's something that -from my point of view- >> matters. >> >> If you want to change the folder where the nova instances are stored you >> have to modify the option in nova-compute.conf 'instances_path': >> >> If you look at that folder (/var/lib/nova/instances/ by default) you will >> see a structure like this: >> >> drwxrwxr-x 2 nova nova 73 Dec 4 12:16 _base >> drwxrwxr-x 2 nova nova 5 Oct 16 13:34 instance-00000002 >> ... >> drwxrwxr-x 2 nova nova 5 Nov 26 17:38 instance-0000005c >> drwxrwxr-x 2 nova nova 6 Dec 11 15:38 instance-00000065 >> >> If you have a shared storage for that folder, then your fstab entry looks >> like this one: >> 10.15.100.3:/volumes/vol1/zone1/instances /var/lib/nova/instances nfs >> defaults 0 0 >> >> So, I think that it could be possible to implement something like >> 'storage domains', but tenant/project oriented. Instead of having multiple >> generic mountpoints, each tenant would have a private mountpoint for >> his/her instances. So the /var/lib/nova/instances could look like this >> sample: >> >> /instances >> +/tenantID1 >> ++/instance-0000X >> ++/instance-0000Y >> ++/instance-0000Z >> +/tenantID2 >> ++/instance-0000A >> ++/instance-0000B >> ++/instance-0000C >> ... >> +/tenantIDN >> ++/instance-0000A >> ++/instance-0000B >> ++/instance-0000C >> >> And in the /etc/fstab something like this sample too: >> 10.15.100.3:/volumes/vol1/zone1/instances/tenantID1 >> /var/lib/nova/instances/tenantID1 nfs defaults 0 0 >> 10.15.100.3:/volumes/vol1/zone1/instances/tenantID2 >> /var/lib/nova/instances/tenantID2 nfs >> defaults 0 0 >> ... >> 10.15.100.3:/volumes/vol1/zone1/instances/tenantIDN >> /var/lib/nova/instances/tenantIDN nfs defaults 0 0 >> >> With this approach, we could have something like per tenant QoS on shared >> storage to resell differente storage capabilities on a tenant basis. >> >> I would love to hear feedback, drawback, improvements... >> >> Cheers >> Diego >> >> -- >> Diego Parrilla >> <http://www.stackops.com/>*CEO* >> *www.stackops.com | * [email protected]** | +34 649 94 43 29 | >> skype:diegoparrilla* >> * <http://www.stackops.com/> >> * >> >> * >> >> >> >> >> On Thu, Dec 20, 2012 at 4:32 PM, Andrew Holway <[email protected]>wrote: >> >>> Good plan. >>> >>> >>> https://blueprints.launchpad.net/openstack-ci/+spec/multiple-storage-domains >>> >>> >>> On Dec 20, 2012, at 4:25 PM, David Busby wrote: >>> >>> > I may of course be entirely wrong :) which would be cool if this is >>> achievable / on the roadmap. >>> > >>> > At the very least if this is not already in discussion I'd raise it on >>> launchpad as a potential feature. >>> > >>> > >>> > >>> > >>> > On Thu, Dec 20, 2012 at 3:19 PM, Andrew Holway <[email protected]> >>> wrote: >>> > Ah shame. You can specify different storage domains in oVirt. >>> > >>> > On Dec 20, 2012, at 4:16 PM, David Busby wrote: >>> > >>> > > Hi Andrew, >>> > > >>> > > An interesting idea, but I am unaware if nova supports storage >>> affinity in any way, it does support host affinity iirc, as a kludge you >>> could have say some nova compute nodes using your "slow mount" and reserve >>> the "fast mount" nodes as required, perhaps even defining separate zones >>> for deployment? >>> > > >>> > > Cheers >>> > > >>> > > David >>> > > >>> > > >>> > > >>> > > >>> > > >>> > > On Thu, Dec 20, 2012 at 2:53 PM, Andrew Holway < >>> [email protected]> wrote: >>> > > Hi David, >>> > > >>> > > It is for nova. >>> > > >>> > > Im not sure I understand. I want to be able to say to openstack; >>> "openstack, please install this instance (A) on this mountpoint and please >>> install this instance (B) on this other mountpoint." I am planning on >>> having two NFS / Gluster based stores, a fast one and a slow one. >>> > > >>> > > I probably will not want to say please every time :) >>> > > >>> > > Thanks, >>> > > >>> > > Andrew >>> > > >>> > > On Dec 20, 2012, at 3:42 PM, David Busby wrote: >>> > > >>> > > > Hi Andrew, >>> > > > >>> > > > Is this for glance or nova ? >>> > > > >>> > > > For nova change: >>> > > > >>> > > > state_path = /var/lib/nova >>> > > > lock_path = /var/lib/nova/tmp >>> > > > >>> > > > in your nova.conf >>> > > > >>> > > > For glance I'm unsure, may be easier to just mount gluster right >>> onto /var/lib/glance (similarly could do the same for /var/lib/nova). >>> > > > >>> > > > And just my £0.02 I've had no end of problems getting gluster to >>> "play nice" on small POC clusters (3 - 5 nodes, I've tried nfs tried >>> glusterfs, tried 2 replica N distribute setups with many a random glusterfs >>> death), as such I have opted for using ceph. >>> > > > >>> > > > ceph's rados can also be used with cinder from the brief reading >>> I've been doing into it. >>> > > > >>> > > > >>> > > > Cheers >>> > > > >>> > > > David >>> > > > >>> > > > >>> > > > >>> > > > >>> > > > >>> > > > On Thu, Dec 20, 2012 at 1:53 PM, Andrew Holway < >>> [email protected]> wrote: >>> > > > Hi, >>> > > > >>> > > > If I have /nfs1mount and /nfs2mount or /nfs1mount and >>> /glustermount can I control where openstack puts the disk files? >>> > > > >>> > > > Thanks, >>> > > > >>> > > > Andrew >>> > > > >>> > > > _______________________________________________ >>> > > > Mailing list: https://launchpad.net/~openstack >>> > > > Post to : [email protected] >>> > > > Unsubscribe : https://launchpad.net/~openstack >>> > > > More help : https://help.launchpad.net/ListHelp >>> > > > >>> > > >>> > > >>> > > >>> > >>> > >>> > >>> >>> >>> >>> _______________________________________________ >>> Mailing list: https://launchpad.net/~openstack >>> Post to : [email protected] >>> Unsubscribe : https://launchpad.net/~openstack >>> More help : https://help.launchpad.net/ListHelp >>> >> >> >> _______________________________________________ >> Mailing list: https://launchpad.net/~openstack >> Post to : [email protected] >> Unsubscribe : https://launchpad.net/~openstack >> More help : https://help.launchpad.net/ListHelp >> >> > > _______________________________________________ > Mailing list: https://launchpad.net/~openstack > Post to : [email protected] > Unsubscribe : https://launchpad.net/~openstack > More help : https://help.launchpad.net/ListHelp > > For the sake of simplicity could you just use the existing features and add a compute node with instances_patch set to the *other* back-ends? Then utilize the existing compute scheduler parameters to determine which compute node and thus which storage back-end is used for the instance? John
_______________________________________________ Mailing list: https://launchpad.net/~openstack Post to : [email protected] Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp

