https://ceph.com/docs/master/rados/operations/crush-map/#placing-different-pools-on-different-osds

See "rule ssd-primary"
 31 окт. 2013 г. 17:29 пользователь "Alexis GÜNST HORN" <
[email protected]> написал:

> Hello to all,
>
> Here is my ceph osd tree output :
>
>
> # id    weight  type name       up/down reweight
> -1      20      root example
> -12     20              drive ssd
> -22     20                      datacenter ssd-dc1
> -104    10                              room ssd-dc1-A
> -502    10                                      host A-ceph-osd-1
> 0       1                                               osd.0   up      1
>                                                         ...
> 9       1                                               osd.9   up      1
> -103    10                              room ssd-dc1-B
> -501    10                                      host B-ceph-osd-1
> 10      1                                               osd.10  up      1
>                                                         ...
> 19      1                                               osd.19  up      1
> -11     0               drive hdd
> -21     0                       datacenter hdd-dc1
> -102    0                               room hdd-dc1-A
> -503    0                                       host A-ceph-osd-2
> 20      0                                               osd.20  up      1
>                                                         ...
> 27      0                                               osd.27  up      1
> -505    0                                       host A-ceph-osd-3
> 35      0                                               osd.35  up      1
>                                                         ...
> 42      0                                               osd.42  up      1
> -101    0                               room hdd-dc1-B
> -504    0                                       host B-ceph-osd-2
> 28      0                                               osd.28  up      1
>                                                         ...
> 34      0                                               osd.34  up      1
> -506    0                                       host B-ceph-osd-3
> 43      0                                               osd.43  up      1
>                                                         ...
> 50      0                                               osd.50  up      1
>
>
>
>
>
> So, in a word,
> I've 2 sorts of OSD hosts : full SSD, and full standard HDD.
> Hosts are splited in 2 rooms : A and B.
>
> I've created 2 pools :
> vm-sdd and vm-hdd
>
> My goal is this one :
> for pool vm-sdd, replicate data between room A and B on SSD hosts, so
> here between A-ceph-osd-1 and B-ceph-osd-1
>
> for pool vm-hdd, to the same, but only with HDD hosts.
>
> Here is my rules in my crush map :
>
> rule vm-sdd {
>         ruleset 3
>         type replicated
>         min_size 1
>         max_size 10
>         step take example
>         step emit
>         step take ssd
>         step chooseleaf firstn 0 type room
>         step emit
> }
> rule vm-hdd {
>         ruleset 4
>         type replicated
>         min_size 1
>         max_size 10
>         step take example
>         step emit
>         step take hdd
>         step chooseleaf firstn 0 type room
>         step emit
> }
>
>
> But  but... when i do :
> ceph osd pool set vm-hddd crush_ruleset 4
>
> it does not work, i get all my PGs stuck.
> Do I have missed something ?
>
> Do you have an idea ?
> Thanks a lot for your help.
>
>
>
>
>
>
> Best Regards - Cordialement
> Alexis
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to