Nevermind ...a bit of reading was enough to point me to "osd_crush_update_on_start": "true"
Thanks Steven On Wed, 31 Oct 2018 at 10:31, Steven Vacaroaia <[email protected]> wrote: > Hi, > I have created a separate root for my ssd drives > All works well but a reboot ( or restart of the services) wipes out all my > changes > > How can I persist changes to crush rules ? > > here are some details > > Initial / default - this is what I am getting after a restart / reboot > If I just do that on one server, the crush rules specific that that server > will be reverted > The new root ( ssds) will persist though > > ceph osd tree > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > -16 0 root ssds > -17 0 host osd01-ssd > -18 0 host osd02-ssd > -19 0 host osd03-ssd > -20 0 host osd04-ssd > -1 32.63507 root default > -3 8.15877 host osd01 > 4 hdd 1.85789 osd.4 up 1.00000 1.00000 > 5 hdd 1.85789 osd.5 up 1.00000 1.00000 > 6 hdd 1.85789 osd.6 up 1.00000 1.00000 > 7 hdd 1.85789 osd.7 up 1.00000 1.00000 > 0 ssd 0.72719 osd.0 up 1.00000 1.00000 > -5 8.15877 host osd02 > 8 hdd 1.85789 osd.8 up 1.00000 1.00000 > 9 hdd 1.85789 osd.9 up 1.00000 1.00000 > 10 hdd 1.85789 osd.10 up 1.00000 1.00000 > 11 hdd 1.85789 osd.11 up 1.00000 1.00000 > 1 ssd 0.72719 osd.1 up 1.00000 1.00000 > -7 8.15877 host osd03 > 12 hdd 1.85789 osd.12 up 1.00000 1.00000 > 13 hdd 1.85789 osd.13 up 1.00000 1.00000 > 14 hdd 1.85789 osd.14 up 1.00000 1.00000 > 15 hdd 1.85789 osd.15 up 1.00000 1.00000 > 2 ssd 0.72719 osd.2 up 1.00000 1.00000 > -9 8.15877 host osd04 > 16 hdd 1.85789 osd.16 up 1.00000 1.00000 > 17 hdd 1.85789 osd.17 up 1.00000 1.00000 > 18 hdd 1.85789 osd.18 up 1.00000 1.00000 > 19 hdd 1.85789 osd.19 up 1.00000 1.00000 > 3 ssd 0.72719 osd.3 up 1.00000 1.00000 > > > changes made > > ceph osd crush add 0 0.72719 root=ssds > > ceph osd crush set osd.0 0.72719 root=ssds host=osd01-ssd > > ceph osd crush add 1 0.72719 root=ssds > > ceph osd crush set osd.1 0.72719 root=ssds host=osd02-ssd > > ceph osd crush add 2 0.72719 root=ssds > > ceph osd crush set osd.2 0.72719 root=ssds host=osd03-ssd > > ceph osd crush add 3 0.72719 root=ssds > > ceph osd crush set osd.3 0.72719 root=ssds host=osd04-ssd > > > ceph osd tree > > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > > -16 2.90875 root ssds > > -17 0.72719 host osd01-ssd > > 0 ssd 0.72719 osd.0 up 1.00000 1.00000 > > -18 0.72719 host osd02-ssd > > 1 ssd 0.72719 osd.1 up 1.00000 1.00000 > > -19 0.72719 host osd03-ssd > > 2 ssd 0.72719 osd.2 up 1.00000 1.00000 > > -20 0.72719 host osd04-ssd > > 3 ssd 0.72719 osd.3 up 1.00000 1.00000 > > -1 29.72632 root default > > -3 7.43158 host osd01 > > 4 hdd 1.85789 osd.4 up 1.00000 1.00000 > > 5 hdd 1.85789 osd.5 up 1.00000 1.00000 > > 6 hdd 1.85789 osd.6 up 1.00000 1.00000 > > 7 hdd 1.85789 osd.7 up 1.00000 1.00000 > > -5 7.43158 host osd02 > > 8 hdd 1.85789 osd.8 up 1.00000 1.00000 > > 9 hdd 1.85789 osd.9 up 1.00000 1.00000 > > 10 hdd 1.85789 osd.10 up 1.00000 1.00000 > > 11 hdd 1.85789 osd.11 up 1.00000 1.00000 > > -7 7.43158 host osd03 > > 12 hdd 1.85789 osd.12 up 1.00000 1.00000 > > 13 hdd 1.85789 osd.13 up 1.00000 1.00000 > > 14 hdd 1.85789 osd.14 up 1.00000 1.00000 > > 15 hdd 1.85789 osd.15 up 1.00000 1.00000 > > -9 7.43158 host osd04 > > 16 hdd 1.85789 osd.16 up 1.00000 1.00000 > > 17 hdd 1.85789 osd.17 up 1.00000 1.00000 > > 18 hdd 1.85789 osd.18 up 1.00000 1.00000 > > 19 hdd 1.85789 osd.19 up 1.00000 1.00000 > > >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
