The method I have used is to 1) edit ceph.conf, 2) use ceph-deploy config
push, 3) restart monitors
Example:
roger@desktop:~/ceph-cluster$ vi ceph.conf # make ceph.conf change
roger@desktop:~/ceph-cluster$ ceph-deploy --overwrite-conf config push
nuc{1..3}
[ceph_deploy.conf][DEBUG ] found configuration file at:
/home/roger/.cephdeploy.conf
[ceph_deploy.cli][INFO ] Invoked (1.5.38): /usr/bin/ceph-deploy
--overwrite-conf config push nuc1 nuc2 nuc3
[ceph_deploy.cli][INFO ] ceph-deploy options:
[ceph_deploy.cli][INFO ] username : None
[ceph_deploy.cli][INFO ] verbose : False
[ceph_deploy.cli][INFO ] overwrite_conf : True
[ceph_deploy.cli][INFO ] subcommand : push
[ceph_deploy.cli][INFO ] quiet : False
[ceph_deploy.cli][INFO ] cd_conf :
<ceph_deploy.conf.cephdeploy.Conf instance at 0x7fbb393de3b0>
[ceph_deploy.cli][INFO ] cluster : ceph
[ceph_deploy.cli][INFO ] client : ['nuc1', 'nuc2',
'nuc3']
[ceph_deploy.cli][INFO ] func : <function config
at 0x7fbb396299b0>
[ceph_deploy.cli][INFO ] ceph_conf : None
[ceph_deploy.cli][INFO ] default_release : False
[ceph_deploy.config][DEBUG ] Pushing config to nuc1
[nuc1][DEBUG ] connection detected need for sudo
[nuc1][DEBUG ] connected to host: nuc1
[nuc1][DEBUG ] detect platform information from remote host
[nuc1][DEBUG ] detect machine type
[nuc1][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph_deploy.config][DEBUG ] Pushing config to nuc2
[nuc2][DEBUG ] connection detected need for sudo
[nuc2][DEBUG ] connected to host: nuc2
[nuc2][DEBUG ] detect platform information from remote host
[nuc2][DEBUG ] detect machine type
[nuc2][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph_deploy.config][DEBUG ] Pushing config to nuc3
[nuc3][DEBUG ] connection detected need for sudo
[nuc3][DEBUG ] connected to host: nuc3
[nuc3][DEBUG ] detect platform information from remote host
[nuc3][DEBUG ] detect machine type
[nuc3][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
roger@desktop:~/ceph-cluster$ ssh nuc1
roger@nuc1:~$ sudo systemctl restart ceph-mon.target
roger@nuc1:~$
...etc.
On Mon, Jul 24, 2017 at 11:34 AM moftah moftah <[email protected]> wrote:
> Hi
>
> I am having hard time finding documentation on what is the correct way to
> upgrade ceph.conf in running cluster.
>
> The change i want to introduce is this
> osd crush update on start = false
>
> i tried to do it through the tell utility like this
> ceph tell osd.82 injectargs --no-osd-crush-update-on-start
>
> the answer was
> osd_crush_update_on_start = 'false' (unchangeable)
>
> Now it seems i need to reboot somthing to get this new config alive
> I find it strange that i have to reboot all OSDs processes int he system
> just to update the config
>
> is there a procedure for this
> and can i just reboot the mon process on the mon nodes ?
>
> Thanks
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com