Thanks much, Dave and Pasi, for your feedback. It certainly makes sense to proceed in that order, viz. dealing with thin provisioning of VHD before pursuing a pool-wide lvmthin implementation. Thanks very much for taking this into consideration!
-=Tobias

On 1/12/2015 2:57 AM, Dave Scott wrote:
On 12 Jan 2015, at 09:29, Pasi Kärkkäinen <pa...@iki.fi> wrote:

On Fri, Jan 09, 2015 at 03:59:26PM -0700, Tobias Kreidl wrote:
   So sorry, I meant of course to write that the utility is lvmthin (not
   thinlvm).

Hello,

I *think* the lvmthin stuff is currently designed for single-host systems,
where the storage is used by a single dedicated host only. That's the usual 
model for LVM aswell.
That’s right.

If multiple hosts access the same LVM volumes one needs to make sure the LVM 
metadata changes are properly synchronized across all hosts,
so all the hosts sharing the same physical LUNs/PVs/VGs/LVs have exactly the 
same idea of the LVM settings, and I believe XAPI does this for normal (thick) 
LVM volumes.

But it gets much more difficult with thin volumes, because the "metadata" can 
change on *every* write IO (when one needs to allocate more blocks from the thin pool?).

Note I haven't looked at the internals of LVM thinprovisioning, but I assume 
it'll be more difficult to get working in shared-LUN/multi-host environments 
compared to normal thick-LVM.

I guess there would have to be some "free-blocks-per-host-for-thin-volumes", so 
each host would allocate new blocks from its dedicated pool, without the risk of 
corrupting the LVM PVs or the volumes.. and avoiding the need of syncing metadata on 
every write IO?
LVM’s thin provisioning uses the device mapper “dm-thin” target, which works as 
you describe. I think for local, non-shared LVM it looks really good. I’ve got 
a very experimental storage plugin which can use it (ezlvm[1]). Also I think 
Fedora 21 can boot from thin-provisioned LVM which would let you do things like 
snapshot (and revert) the whole dom0 filesystem.

I’ve thought a little about how we could use it in future for shared LVM. 
Potentially we could make the “thin pool” work for per-host allocation by a 
ballooning-like technique. If every host saw the full set of free blocks, but 
most of the blocks were masked off by a fake disk (like the balloon driver’s 
balloon) then each host would be able to allocate locally from the same block 
address space. This would make it easier to move volumes between hosts. I think 
each host would need its own private copy of the LVM metadata, which would 
function like the journal in this “thinlvhd” design. When I tried to make this 
work a while ago I hit a problem when I tried to modify the “balloon disk”: 
when I reloaded the thin pool metadata and resumed the device mapper device it 
had clearly cached some of the data structure in memory because it immediately 
corrupted itself. This is probably fixable but requires more experimentation :)

So for the shared case I think we should work on “thinlvhd” (i.e. thin 
provisioning via tapdisk) first but play with dm-thin in the background :)

Cheers,
Dave

[1] https://github.com/xapi-project/ezlvm



Just random thoughts :)

-- Pasi

   On 1/9/2015 3:54 PM, Tobias Kreidl wrote:

     As of around CentOS 6.5 (and for sure in RHEL 6.4) that there is a
     thinlvm lvmthin utility that seems to take care of supporting
     thinly-provisioned LVM volumes. There is an associated snapshot design
     based on LVM thin provisioning that even supports the ability to do
     snapshots of snapshots, etc. down the chain.

     Unfortunately, from a cursory look, it doesn't look like a back port to
     CentOS 5 would be that easy or even possible, but that it is integrated
     into both CentOS 6 and CentOS 7 gives some hope for a possibly
     standardized support of it in a future XenServer release, doesn't it?
     The other big question would be how readily something like this could be
     integrated into XenServer.

     -=Tobias




_______________________________________________
Xen-api mailing list
Xen-api@lists.xen.org
http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api

Reply via email to