Hi,

For VDI (Windows 10) use case... is there any document about the
recommended configuration with rbd?

Thanks a lot!

2017-08-18 15:40 GMT+02:00 Oscar Segarra <oscar.sega...@gmail.com>:

> Hi,
>
> Yes, you are right, the idea is cloning a snapshot taken from the base
> image...
>
> And yes, I'm working with the current RC of luminous.
>
> In this scenario: base image (raw format)  + snapshot + snapshot clones
> (for end user Windows 10 vdi). Does tiering ssd+hdd may help?
>
> Thanks a lot
>
>
> El 18 ago. 2017 4:05, "David Turner" <drakonst...@gmail.com> escribió:
>
> Do you mean a lot of snapshots or creating a lot of clones from a
> snapshot? I can agree to the pain of crating a lot of snapshots of rbds in
> ceph. I'm assuming that you mean to say that you will have a template rbd
> with a version snapshot that you clone each time you need to let someone
> log in. Is that what you're planning?
>
> On Thu, Aug 17, 2017, 9:51 PM Christian Balzer <ch...@gol.com> wrote:
>
>>
>> Hello,
>>
>> On Fri, 18 Aug 2017 03:31:56 +0200 Oscar Segarra wrote:
>>
>> > Hi Christian,
>> >
>> > Thanks a lot for helping...
>> >
>> > Have you read:
>> > http://docs.ceph.com/docs/master/rbd/rbd-openstack/
>> >
>> > So just from the perspective of qcow2, you seem to be doomed.
>> > --> Sorry, I've talking about RAW + QCOW2 when I meant RBD images and
>> RBD
>> > snapshots...
>> >
>> I tested Snapshots with Hammer and the release before it, found them
>> immensely painful (resource intensive) and avoided them since.
>> That said, there are supposedly quite some improvements in recent versions
>> (I suppose you'll deploy with Luminous), as well as more (and
>> working) control knobs to reduce the impact of snapshot operations.
>>
>> > A sufficiently large cache tier should help there immensely and the
>> base image
>> > should be in cache (RAM, pagecache on the OSD servers really) all the
>> time
>> > anyway.
>> > --> If talking about RBD images and RBD snapshots... it helps immensely
>> as
>> > well?
>> >
>> No experience, so nothing conclusive and authoritative from my end.
>> If the VMs write/read alot of the same data (as in 4MB RBD objects),
>> cache-tiering should help again.
>> But promoting and demoting things through it when dealing with snapshots
>> and deletions of them might be a pain.
>>
>> Christian
>>
>> > Sizing this and specifying the correct type of SSDs/NVMes for the
>> cache-tier
>> > is something that only you can answer based on existing data or
>> sufficiently
>> > detailed and realistic tests.
>> > --> Yes, the problem is that I have to buy a HW and for Windows 10
>> VDI...
>> > and I cannot make realistic tests previously :( but I will work on this
>> > line...
>> >
>> > Thanks a lot again!
>> >
>> >
>> >
>> > 2017-08-18 3:14 GMT+02:00 Christian Balzer <ch...@gol.com>:
>> >
>> > >
>> > > Hello,
>> > >
>> > > On Thu, 17 Aug 2017 23:56:49 +0200 Oscar Segarra wrote:
>> > >
>> > > > Hi David,
>> > > >
>> > > > Thanks a lot again for your quick answer...
>> > > >
>> > > > *The rules in the CRUSH map will always be followed.  It is not
>> possible
>> > > > for Ceph to go against that and put data into a root that shouldn't
>> have
>> > > > it.*
>> > > > --> I will work on your proposal of creating two roots in the CRUSH
>> > > map...
>> > > > just one question more:
>> > > > --> Regarding to space consumption, with this proposal, is it
>> possible to
>> > > > know how many disk space is it free in each pool?
>> > > >
>> > > > *The problem with a cache tier is that Ceph is going to need to
>> promote
>> > > and
>> > > > evict stuff all the time (not free).  A lot of people that want to
>> use
>> > > SSD
>> > > > cache tiering for RBDs end up with slower performance because of
>> this.
>> > > > Christian Balzer is the expert on Cache Tiers for RBD usage.  His
>> primary
>> > > > stance is that it's most likely a bad idea, but there are definite
>> cases
>> > > > where it's perfect.*
>> > > > --> Christian, is there any advice for VDI --> BASE IMAGE (raw) +
>> 1000
>> > > > linked clones (qcow2)
>> > > >
>> > > Have you read:
>> > > http://docs.ceph.com/docs/master/rbd/rbd-openstack/
>> > >
>> > > So just from the perspective of qcow2, you seem to be doomed.
>> > >
>> > > Windows always appears to be very chatty when it comes to I/O and also
>> > > paging/swapping seemingly w/o need, rhyme or reason.
>> > > A sufficiently large cache tier should help there immensely and the
>> base
>> > > image should be in cache (RAM, pagecache on the OSD servers really)
>> all the
>> > > time anyway.
>> > > Sizing this and specifying the correct type of SSDs/NVMes for the
>> > > cache-tier is something that only you can answer based on existing
>> data or
>> > > sufficiently detailed and realistic tests.
>> > >
>> > > Christian
>> > >
>> > > > Thanks a lot.
>> > > >
>> > > >
>> > > > 2017-08-17 22:42 GMT+02:00 David Turner <drakonst...@gmail.com>:
>> > > >
>> > > > > The rules in the CRUSH map will always be followed.  It is not
>> possible
>> > > > > for Ceph to go against that and put data into a root that
>> shouldn't
>> > > have it.
>> > > > >
>> > > > > The problem with a cache tier is that Ceph is going to need to
>> promote
>> > > and
>> > > > > evict stuff all the time (not free).  A lot of people that want
>> to use
>> > > SSD
>> > > > > cache tiering for RBDs end up with slower performance because of
>> this.
>> > > > > Christian Balzer is the expert on Cache Tiers for RBD usage.  His
>> > > primary
>> > > > > stance is that it's most likely a bad idea, but there are definite
>> > > cases
>> > > > > where it's perfect.
>> > > > >
>> > > > >
>> > > > > On Thu, Aug 17, 2017 at 4:20 PM Oscar Segarra <
>> oscar.sega...@gmail.com
>> > > >
>> > > > > wrote:
>> > > > >
>> > > > >> Hi David,
>> > > > >>
>> > > > >> Thanks a lot for your quick answer!
>> > > > >>
>> > > > >> *If I'm understanding you correctly, you want to have 2 different
>> > > roots
>> > > > >> that pools can be made using.  The first being entirely SSD
>> storage.
>> > > The
>> > > > >> second being HDD Storage with an SSD cache tier on top of it.  *
>> > > > >> --> Yes, this is what I mean.
>> > > > >>
>> > > > >> https://www.sebastien-han.fr/blog/2014/08/25/ceph-mix-sata-
>> > > > >> and-ssd-within-the-same-box/
>> > > > >> --> I'm not an expert in CRUSH rules... Whit this configuration,
>> it is
>> > > > >> guaranteed that objects stored in ssd pool do not "go" to the hdd
>> > > disks?
>> > > > >>
>> > > > >> *The above guide explains how to set up the HDD root and the SSD
>> root.
>> > > > >> After that all you do is create a pool on the HDD root for RBDs,
>> a
>> > > pool on
>> > > > >> the SSD root for a cache tier to use with the HDD pool, and then
>> a a
>> > > pool
>> > > > >> on the SSD root for RBDs.  There aren't actually a lot of use
>> cases
>> > > out
>> > > > >> there where using an SSD cache tier on top of an HDD RBD pool is
>> what
>> > > you
>> > > > >> really want.  I would recommend testing this thoroughly and
>> comparing
>> > > your
>> > > > >> performance to just a standard HDD pool for RBDs without a cache
>> > > tier.*
>> > > > >> --> I'm working on a VDI solution where there are BASE IMAGES
>> (raw)
>> > > and
>> > > > >> qcow2 linked clones... where I expect not all VDIs to be powered
>> on
>> > > at the
>> > > > >> same time and perform a configuration to avoid problems related
>> to
>> > > login
>> > > > >> storm. (1000 hosts)
>> > > > >> --> Do you think it is not a good idea? do you know what does
>> usually
>> > > > >> people configure for this kind of scenarios?
>> > > > >>
>> > > > >> Thanks a lot.
>> > > > >>
>> > > > >>
>> > > > >> 2017-08-17 21:31 GMT+02:00 David Turner <drakonst...@gmail.com>:
>> > > > >>
>> > > > >>> If I'm understanding you correctly, you want to have 2 different
>> > > roots
>> > > > >>> that pools can be made using.  The first being entirely SSD
>> > > storage.  The
>> > > > >>> second being HDD Storage with an SSD cache tier on top of it.
>> > > > >>>
>> > > > >>> https://www.sebastien-han.fr/blog/2014/08/25/ceph-mix-sata-
>> > > > >>> and-ssd-within-the-same-box/
>> > > > >>>
>> > > > >>> The above guide explains how to set up the HDD root and the SSD
>> root.
>> > > > >>> After that all you do is create a pool on the HDD root for
>> RBDs, a
>> > > pool on
>> > > > >>> the SSD root for a cache tier to use with the HDD pool, and
>> then a a
>> > > pool
>> > > > >>> on the SSD root for RBDs.  There aren't actually a lot of use
>> cases
>> > > out
>> > > > >>> there where using an SSD cache tier on top of an HDD RBD pool is
>> > > what you
>> > > > >>> really want.  I would recommend testing this thoroughly and
>> > > comparing your
>> > > > >>> performance to just a standard HDD pool for RBDs without a cache
>> > > tier.
>> > > > >>>
>> > > > >>> On Thu, Aug 17, 2017 at 3:18 PM Oscar Segarra <
>> > > oscar.sega...@gmail.com>
>> > > > >>> wrote:
>> > > > >>>
>> > > > >>>> Hi,
>> > > > >>>>
>> > > > >>>> Sorry guys, during theese days I'm asking a lot about how to
>> > > distribute
>> > > > >>>> my data.
>> > > > >>>>
>> > > > >>>> I have two kinds of VM:
>> > > > >>>>
>> > > > >>>> 1.- Management VMs (linux) --> Full SSD dedicated disks
>> > > > >>>> 2.- Windows VM --> SSD + HHD (with tiering).
>> > > > >>>>
>> > > > >>>> I'm working on installing two clusters on the same host but I'm
>> > > > >>>> encountering lots of problems as named clusters look not be
>> fully
>> > > supported.
>> > > > >>>>
>> > > > >>>> In the same cluster, Is there any way to distribute my VMs as I
>> > > like?
>> > > > >>>>
>> > > > >>>> Thanks a lot!
>> > > > >>>>
>> > > > >>>> Ó.
>> > > > >>>> _______________________________________________
>> > > > >>>> ceph-users mailing list
>> > > > >>>> ceph-users@lists.ceph.com
>> > > > >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> > > > >>>>
>> > > > >>>
>> > > > >>
>> > >
>> > >
>> > > --
>> > > Christian Balzer        Network/Systems Engineer
>> > > ch...@gol.com           Rakuten Communications
>> > >
>>
>>
>> --
>> Christian Balzer        Network/Systems Engineer
>> ch...@gol.com           Rakuten Communications
>>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to