Hi Eugen, Thanks for your help! Everything went well and the cluster came back without any problems.
Thanks, Gagan On Thu, Aug 7, 2025 at 8:03 PM gagan tiwari <gagan.tiw...@mathisys-india.com> wrote: > Hi Eugen, > Thanks for clarifying! I will just set the noout > flag and power off all servers. Hopefully, it will work fine. > > Thanks, > Gagan > > On Thu, Aug 7, 2025 at 5:59 PM Eugen Block <ebl...@nde.ag> wrote: > >> No, that's not what I meant. Of course you need the noout flag, that's >> the first point on the list I sent. Other flags are not required as >> pointed out by croit. >> I have never disabled cephfs before shutting it down as mentioned by >> Gerard, so I can't comment on that. >> >> The procedure I quoted from SUSE is still working fine, that's how we >> power off our (cephadm) based clusters as well. >> >> Zitat von gagan tiwari <gagan.tiw...@mathisys-india.com>: >> >> > Hi Eugen, >> > I had deployed the cluster via cephadm. so, it's >> > container based. Just fyi.. >> > >> > So, you meant I don't need to run any commands like "ceph osd set >> noout" , >> > "ceph orch stop osd.<ID>" ,etc. >> > >> > I just need to power off all nodes starting clients nodes , osd nodes , >> mon >> > , mds nodes and power them on after maintenance activity is over. >> > >> > Thanks, >> > Gagan >> > >> > On Wed, Aug 6, 2025 at 11:53 PM Eugen Block <ebl...@nde.ag> wrote: >> > >> >> You don't need to stop all the OSDs (or other daemons) manually, just >> >> shut down the servers (most likely you have services colocated). When >> >> they boot again, the Ceph daemons will also start automatically >> >> (they're handled by systemd). I can check tomorrow which steps exactly >> >> our shutdown procedure consists of when we have planned power outages >> >> etc. >> >> >> >> Zitat von gagan tiwari <gagan.tiw...@mathisys-india.com>: >> >> >> >> > HI Eugen, >> >> > We have 80 osds in the cluster. >> >> > >> >> > So, to stop them I will need to run the command *ceph orch stop >> >> osd.<ID>* >> >> > for all 80 osds one by one. >> >> > Is there any way to stop all of them in one command ? >> >> > >> >> > Also, once all nodes will be back after power maintenance activity , >> >> will I >> >> > need to start all daemons mon , mds , osd,etc >> >> > ? or will they be up automatically once servers will be up ? >> >> > >> >> > Thanks, >> >> > Gagan >> >> > >> >> > On Wed, Aug 6, 2025 at 4:45 PM Eugen Block <ebl...@nde.ag> wrote: >> >> > >> >> >> Although SUSE discontinued their product, the procedure is still >> correct >> >> >> [0]: >> >> >> >> >> >> 1. Tell the Ceph cluster not to mark OSDs as out: >> >> >> >> >> >> ceph osd set noout >> >> >> >> >> >> 2. Stop daemons and nodes in the following order: >> >> >> >> >> >> Storage clients >> >> >> >> >> >> Gateways, for example NFS Ganesha or Object Gateway >> >> >> >> >> >> Metadata Server >> >> >> >> >> >> Ceph OSD >> >> >> >> >> >> Ceph Manager >> >> >> >> >> >> Ceph Monitor >> >> >> >> >> >> 3. If required, perform maintenance tasks. >> >> >> >> >> >> 4. Start the nodes and servers in the reverse order of the shutdown >> >> >> process: >> >> >> >> >> >> Ceph Monitor >> >> >> >> >> >> Ceph Manager >> >> >> >> >> >> Ceph OSD >> >> >> >> >> >> Metadata Server >> >> >> >> >> >> Gateways, for example NFS Ganesha or Object Gateway >> >> >> >> >> >> Storage clients >> >> >> >> >> >> 5. Remove the noout flag: >> >> >> >> >> >> ceph osd unset noout >> >> >> >> >> >> [0] >> >> >> >> >> >> >> >> >> https://documentation.suse.com/en-us/ses/7.1/html/ses-all/storage-salt-cluster.html#sec-salt-cluster-reboot >> >> >> >> >> >> Zitat von gagan tiwari <gagan.tiw...@mathisys-india.com>: >> >> >> >> >> >> > Hi Guys, >> >> >> > I have recently set-up a production ceph cluster >> >> which >> >> >> > consists of 3 monitor nodes and 7 Osd nodes. >> >> >> > >> >> >> > There is power maintenance activity scheduled at the data centre >> >> coming >> >> >> > weekend and due to that I need to power off all the devices. >> >> >> > >> >> >> > Can you please advise me on the safest way to power off all >> servers? >> >> >> > >> >> >> > Should I power off all 7 OSD servers one by one followed by all 3 >> >> monitor >> >> >> > nodes or vice versa? >> >> >> > >> >> >> > Thanks, >> >> >> > Gagan >> >> >> > _______________________________________________ >> >> >> > ceph-users mailing list -- ceph-users@ceph.io >> >> >> > To unsubscribe send an email to ceph-users-le...@ceph.io >> >> >> >> >> >> >> >> >> _______________________________________________ >> >> >> ceph-users mailing list -- ceph-users@ceph.io >> >> >> To unsubscribe send an email to ceph-users-le...@ceph.io >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io