On Sat, Nov 21, 2015, at 13:59, Dan Kenigsberg wrote:
> On Fri, Nov 20, 2015 at 01:54:35PM +0100, Giuseppe Ragusa wrote:
> > Hi all,
> > I go on with my wishlist, derived from both solitary mumblings and 
> > community talks at the the first Italian oVirt Meetup.
> > 
> > I offer to help in coding (work/family schedules permitting) but keep in 
> > mind that I'm a sysadmin with mainly C and bash-scripting skills (but 
> > hoping to improve my less-than-newbie Python too...)
> > 
> > I've sent separate wishlist messages for oVirt Node and Engine.
> > 
> > VDSM:
> > 
> > *) allow VDSM to configure/manage Samba, CTDB and Ganesha (specifically, 
> > I'm thinking of the GlusterFS integration); there are related wishlist 
> > items on configuring/managing Samba/CTDB/Ganesha on the Engine and on oVirt 
> > Node
> 
> I'd apreciate a more detailed feature definition. Vdsm (and ovirt) try
> to configure only thing that are needed for their own usage. What do you
> want to control? When? You're welcome to draf a feature page prior to
> coding the fix ;-)

I was thinking of adding CIFS/NFSv4 functionality to an hyperconverged cluster 
(GlusterFS/oVirt) which would have separate volumes for virtual machines 
storage (one volume for the Engine and one for other vms, with no CIFS/NFSv4 
capabilities offered) and for data shares (directly accessible by clients on 
LAN and obviously from local vms too).

Think of it as a 3-node HA NetApp+VMware killer ;-)

The UI idea (but that would be the Engine part, I understand) was along the 
lines of single-check enabling CIFS and/or NFSv4 sharing for a GlusterFS data 
volume, then optionally adding any further specific options (hosts allowed, 
users/groups for read/write access, network recycle_bin etc.); global Samba 
(domain/workgroup membership etc.) and CTDB (IPs/interfaces) configuration 
parameters would be needed too.

I have no experience on a GaneshaNFS clustered/HA configuration with GlusterFS, 
but (from superficial skimming through docs) it seems that it was not possible 
at all before 2.2 and now it needs a full Pacemaker/Corosync setup too 
(contrary to the IBM-GPFS-backed case), so that could be a problem.

This VDSM wishlist item was driven by the idea that all actions (and so future 
GlusterFS/Samba/CTDB too) performed by the Engine through the hosts/nodes were 
somehow "mediated" by VDSM and its API, but if this is not the case, then I 
retire my suggestion here and I will try to pursue it only on the Engine/Node 
side ;)

Many thanks for your attention.

Regards,
Giuseppe

> > *) add Open vSwitch direct support (not Neutron-mediated); there are 
> > related wishlist items on configuring/managing Open vSwitch on oVirt Node 
> > and on the Engine
> 
> That's on our immediate roadmap. Soon, vdsm-hook-ovs would be ready for
> testing.
> 
> > 
> > *) add DRBD9 as a supported Storage Domain type; there are related wishlist 
> > items on configuring/managing DRBD9 on the Engine and on oVirt Node
> > 
> > *) allow VDSM to configure/manage containers (maybe extend it by use of the 
> > LXC libvirt driver, similarly to the experimental work that has been put up 
> > to allow Xen vm management); there are related wishlist items on 
> > configuring/managing containers on the Engine and on oVirt Node
> > 
> > *) add a VDSM_remote mode (for lack of a better name, but mainly inspired 
> > by pacemaker_remote) to be used inside a guest by the above mentioned 
> > container support (giving to the Engine the required visibility on the 
> > managed containers, but excluding the "virtual node" from power management 
> > and other unsuitable actions)
> > 
> > Regards,
> > Giuseppe
> > 
> > _______________________________________________
> > Users mailing list
> > [email protected]
> > http://lists.ovirt.org/mailman/listinfo/users
> 
_______________________________________________
Users mailing list
[email protected]
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to