Hi,

Yes, you are right, the idea is cloning a snapshot taken from the base
image...

And yes, I'm working with the current RC of luminous.

In this scenario: base image (raw format)  + snapshot + snapshot clones
(for end user Windows 10 vdi). Does tiering ssd+hdd may help?

Thanks a lot


El 18 ago. 2017 4:05, "David Turner" <drakonst...@gmail.com> escribió:

Do you mean a lot of snapshots or creating a lot of clones from a snapshot?
I can agree to the pain of crating a lot of snapshots of rbds in ceph. I'm
assuming that you mean to say that you will have a template rbd with a
version snapshot that you clone each time you need to let someone log in.
Is that what you're planning?

On Thu, Aug 17, 2017, 9:51 PM Christian Balzer <ch...@gol.com> wrote:

>
> Hello,
>
> On Fri, 18 Aug 2017 03:31:56 +0200 Oscar Segarra wrote:
>
> > Hi Christian,
> >
> > Thanks a lot for helping...
> >
> > Have you read:
> > http://docs.ceph.com/docs/master/rbd/rbd-openstack/
> >
> > So just from the perspective of qcow2, you seem to be doomed.
> > --> Sorry, I've talking about RAW + QCOW2 when I meant RBD images and RBD
> > snapshots...
> >
> I tested Snapshots with Hammer and the release before it, found them
> immensely painful (resource intensive) and avoided them since.
> That said, there are supposedly quite some improvements in recent versions
> (I suppose you'll deploy with Luminous), as well as more (and
> working) control knobs to reduce the impact of snapshot operations.
>
> > A sufficiently large cache tier should help there immensely and the base
> image
> > should be in cache (RAM, pagecache on the OSD servers really) all the
> time
> > anyway.
> > --> If talking about RBD images and RBD snapshots... it helps immensely
> as
> > well?
> >
> No experience, so nothing conclusive and authoritative from my end.
> If the VMs write/read alot of the same data (as in 4MB RBD objects),
> cache-tiering should help again.
> But promoting and demoting things through it when dealing with snapshots
> and deletions of them might be a pain.
>
> Christian
>
> > Sizing this and specifying the correct type of SSDs/NVMes for the
> cache-tier
> > is something that only you can answer based on existing data or
> sufficiently
> > detailed and realistic tests.
> > --> Yes, the problem is that I have to buy a HW and for Windows 10 VDI...
> > and I cannot make realistic tests previously :( but I will work on this
> > line...
> >
> > Thanks a lot again!
> >
> >
> >
> > 2017-08-18 3:14 GMT+02:00 Christian Balzer <ch...@gol.com>:
> >
> > >
> > > Hello,
> > >
> > > On Thu, 17 Aug 2017 23:56:49 +0200 Oscar Segarra wrote:
> > >
> > > > Hi David,
> > > >
> > > > Thanks a lot again for your quick answer...
> > > >
> > > > *The rules in the CRUSH map will always be followed.  It is not
> possible
> > > > for Ceph to go against that and put data into a root that shouldn't
> have
> > > > it.*
> > > > --> I will work on your proposal of creating two roots in the CRUSH
> > > map...
> > > > just one question more:
> > > > --> Regarding to space consumption, with this proposal, is it
> possible to
> > > > know how many disk space is it free in each pool?
> > > >
> > > > *The problem with a cache tier is that Ceph is going to need to
> promote
> > > and
> > > > evict stuff all the time (not free).  A lot of people that want to
> use
> > > SSD
> > > > cache tiering for RBDs end up with slower performance because of
> this.
> > > > Christian Balzer is the expert on Cache Tiers for RBD usage.  His
> primary
> > > > stance is that it's most likely a bad idea, but there are definite
> cases
> > > > where it's perfect.*
> > > > --> Christian, is there any advice for VDI --> BASE IMAGE (raw) +
> 1000
> > > > linked clones (qcow2)
> > > >
> > > Have you read:
> > > http://docs.ceph.com/docs/master/rbd/rbd-openstack/
> > >
> > > So just from the perspective of qcow2, you seem to be doomed.
> > >
> > > Windows always appears to be very chatty when it comes to I/O and also
> > > paging/swapping seemingly w/o need, rhyme or reason.
> > > A sufficiently large cache tier should help there immensely and the
> base
> > > image should be in cache (RAM, pagecache on the OSD servers really)
> all the
> > > time anyway.
> > > Sizing this and specifying the correct type of SSDs/NVMes for the
> > > cache-tier is something that only you can answer based on existing
> data or
> > > sufficiently detailed and realistic tests.
> > >
> > > Christian
> > >
> > > > Thanks a lot.
> > > >
> > > >
> > > > 2017-08-17 22:42 GMT+02:00 David Turner <drakonst...@gmail.com>:
> > > >
> > > > > The rules in the CRUSH map will always be followed.  It is not
> possible
> > > > > for Ceph to go against that and put data into a root that shouldn't
> > > have it.
> > > > >
> > > > > The problem with a cache tier is that Ceph is going to need to
> promote
> > > and
> > > > > evict stuff all the time (not free).  A lot of people that want to
> use
> > > SSD
> > > > > cache tiering for RBDs end up with slower performance because of
> this.
> > > > > Christian Balzer is the expert on Cache Tiers for RBD usage.  His
> > > primary
> > > > > stance is that it's most likely a bad idea, but there are definite
> > > cases
> > > > > where it's perfect.
> > > > >
> > > > >
> > > > > On Thu, Aug 17, 2017 at 4:20 PM Oscar Segarra <
> oscar.sega...@gmail.com
> > > >
> > > > > wrote:
> > > > >
> > > > >> Hi David,
> > > > >>
> > > > >> Thanks a lot for your quick answer!
> > > > >>
> > > > >> *If I'm understanding you correctly, you want to have 2 different
> > > roots
> > > > >> that pools can be made using.  The first being entirely SSD
> storage.
> > > The
> > > > >> second being HDD Storage with an SSD cache tier on top of it.  *
> > > > >> --> Yes, this is what I mean.
> > > > >>
> > > > >> https://www.sebastien-han.fr/blog/2014/08/25/ceph-mix-sata-
> > > > >> and-ssd-within-the-same-box/
> > > > >> --> I'm not an expert in CRUSH rules... Whit this configuration,
> it is
> > > > >> guaranteed that objects stored in ssd pool do not "go" to the hdd
> > > disks?
> > > > >>
> > > > >> *The above guide explains how to set up the HDD root and the SSD
> root.
> > > > >> After that all you do is create a pool on the HDD root for RBDs, a
> > > pool on
> > > > >> the SSD root for a cache tier to use with the HDD pool, and then
> a a
> > > pool
> > > > >> on the SSD root for RBDs.  There aren't actually a lot of use
> cases
> > > out
> > > > >> there where using an SSD cache tier on top of an HDD RBD pool is
> what
> > > you
> > > > >> really want.  I would recommend testing this thoroughly and
> comparing
> > > your
> > > > >> performance to just a standard HDD pool for RBDs without a cache
> > > tier.*
> > > > >> --> I'm working on a VDI solution where there are BASE IMAGES
> (raw)
> > > and
> > > > >> qcow2 linked clones... where I expect not all VDIs to be powered
> on
> > > at the
> > > > >> same time and perform a configuration to avoid problems related to
> > > login
> > > > >> storm. (1000 hosts)
> > > > >> --> Do you think it is not a good idea? do you know what does
> usually
> > > > >> people configure for this kind of scenarios?
> > > > >>
> > > > >> Thanks a lot.
> > > > >>
> > > > >>
> > > > >> 2017-08-17 21:31 GMT+02:00 David Turner <drakonst...@gmail.com>:
> > > > >>
> > > > >>> If I'm understanding you correctly, you want to have 2 different
> > > roots
> > > > >>> that pools can be made using.  The first being entirely SSD
> > > storage.  The
> > > > >>> second being HDD Storage with an SSD cache tier on top of it.
> > > > >>>
> > > > >>> https://www.sebastien-han.fr/blog/2014/08/25/ceph-mix-sata-
> > > > >>> and-ssd-within-the-same-box/
> > > > >>>
> > > > >>> The above guide explains how to set up the HDD root and the SSD
> root.
> > > > >>> After that all you do is create a pool on the HDD root for RBDs,
> a
> > > pool on
> > > > >>> the SSD root for a cache tier to use with the HDD pool, and then
> a a
> > > pool
> > > > >>> on the SSD root for RBDs.  There aren't actually a lot of use
> cases
> > > out
> > > > >>> there where using an SSD cache tier on top of an HDD RBD pool is
> > > what you
> > > > >>> really want.  I would recommend testing this thoroughly and
> > > comparing your
> > > > >>> performance to just a standard HDD pool for RBDs without a cache
> > > tier.
> > > > >>>
> > > > >>> On Thu, Aug 17, 2017 at 3:18 PM Oscar Segarra <
> > > oscar.sega...@gmail.com>
> > > > >>> wrote:
> > > > >>>
> > > > >>>> Hi,
> > > > >>>>
> > > > >>>> Sorry guys, during theese days I'm asking a lot about how to
> > > distribute
> > > > >>>> my data.
> > > > >>>>
> > > > >>>> I have two kinds of VM:
> > > > >>>>
> > > > >>>> 1.- Management VMs (linux) --> Full SSD dedicated disks
> > > > >>>> 2.- Windows VM --> SSD + HHD (with tiering).
> > > > >>>>
> > > > >>>> I'm working on installing two clusters on the same host but I'm
> > > > >>>> encountering lots of problems as named clusters look not be
> fully
> > > supported.
> > > > >>>>
> > > > >>>> In the same cluster, Is there any way to distribute my VMs as I
> > > like?
> > > > >>>>
> > > > >>>> Thanks a lot!
> > > > >>>>
> > > > >>>> Ó.
> > > > >>>> _______________________________________________
> > > > >>>> ceph-users mailing list
> > > > >>>> ceph-users@lists.ceph.com
> > > > >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > > > >>>>
> > > > >>>
> > > > >>
> > >
> > >
> > > --
> > > Christian Balzer        Network/Systems Engineer
> > > ch...@gol.com           Rakuten Communications
> > >
>
>
> --
> Christian Balzer        Network/Systems Engineer
> ch...@gol.com           Rakuten Communications
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to