On Tue, Feb 20, 2018 at 8:35 PM, Rafael Lopez <[email protected]> wrote:
>> There is also work-in-progress for online
>> image migration [1] that will allow you to keep using the image while
>> it's being migrated to a new destination image.
>
>
> Hi Jason,
>
> Is there any recommended method/workaround for live rbd migration in
> luminous? eg. snapshot/copy or export/import then export/import-diff?
> We are looking at options for moving large RBDs (100T) to a new pool with
> minimal downtime.
>
> I was thinking we might be able to configure/hack rbd mirroring to mirror to
> a pool on the same cluster but I gather from the OP and your post that this
> is not really possible?

No, it's not really possible currently and we have no plans to add
such support since it would not be of any long-term value. If you are
using RBD with a kernel block device, you could temporarily wrap two
mapped RBD volumes (original and new) under a md RAID1 with the
original as the primary -- and then do a RAID repair. If using
QEMU+librbd, you could use its built-in block live migration feature
(I've never played w/ this before and I believe you would need to use
the QEMU monitor instead of libvirt to configure).

> --
> Rafael Lopez
> Research Devops Engineer
> Monash University eResearch Centre
>
>



-- 
Jason
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to