On Tue, Jan 08, 2019 at 01:29:51PM -0600, Bryan K. Walton wrote: > On Tue, Jan 08, 2019 at 10:55:09AM -0600, Ken Gaillot wrote: > > > > FYI pcmk_off_action="off" is the default > > > > If you want the cluster to request an "off" command instead of a > > "reboot" when fencing a node, set the stonith-action cluster property > > to "off". > > Awesome! Thank you, Ken. I don't know how I've missed this, up to now. > Setting this property is exactly what I needed.
I swear I had this working the other day. I'm still struggling with this, apparently. I've set the default stonith-action to off: [root@storage1 ~]# pcs config | grep -i stonith-action stonith-action: off [root@storage1 ~]# pcs config | grep -i stonith-enabled stonith-enabled: true But when I run "pcs stonith fence storage2" (from my storage1 node, the fabric ports are getting sucessfully disabled, and then re-enabled: Here are the logs that show stonith-ng issuing "off" commands (successfully), and then following up with "on" commands: Jan 10 13:31:55 storage1 stonith-ng[43051]: notice: Client stonith_admin.44835.f958d69c wants to fence (reboot) 'storage2' with device '(any)' Jan 10 13:31:55 storage1 stonith-ng[43051]: notice: Requesting peer fencing (off) of storage2 Jan 10 13:31:55 storage1 stonith-ng[43051]: notice: fenceStorage2-millipede can fence (reboot) storage2: static-list Jan 10 13:31:55 storage1 stonith-ng[43051]: notice: fenceStorage2-centipede can fence (reboot) storage2: static-list Jan 10 13:31:56 storage1 stonith-ng[43051]: notice: Operation 'off' [44836] (call 2 from stonith_admin.44835) for host 'storage2' with device 'fenceStorage2-centipede' returned: 0 (OK) Jan 10 13:31:56 storage1 stonith-ng[43051]: notice: Call to fenceStorage2-centipede for 'storage2 off' on behalf of stonith_admin.44835@storage1: OK (0) Jan 10 13:31:57 storage1 stonith-ng[43051]: notice: Operation 'off' [44930] (call 2 from stonith_admin.44835) for host 'storage2' with device 'fenceStorage2-millipede' returned: 0 (OK) Jan 10 13:31:57 storage1 stonith-ng[43051]: notice: Call to fenceStorage2-millipede for 'storage2 off' on behalf of stonith_admin.44835@storage1: OK (0) Jan 10 13:31:58 storage1 stonith-ng[43051]: notice: Operation 'on' [44936] (call 2 from stonith_admin.44835) for host 'storage2' with device 'fenceStorage2-centipede' returned: 0 (OK) Jan 10 13:31:58 storage1 stonith-ng[43051]: notice: Call to fenceStorage2-centipede for 'storage2 on' on behalf of stonith_admin.44835@storage1: OK (0) Jan 10 13:32:00 storage1 stonith-ng[43051]: notice: Operation 'on' [44942] (call 2 from stonith_admin.44835) for host 'storage2' with device 'fenceStorage2-millipede' returned: 0 (OK) Jan 10 13:32:00 storage1 stonith-ng[43051]: notice: Call to fenceStorage2-millipede for 'storage2 on' on behalf of stonith_admin.44835@storage1: OK (0) Jan 10 13:32:00 storage1 stonith-ng[43051]: notice: Operation reboot of storage2 by storage1 for [email protected]: OK Jan 10 13:32:00 storage1 crmd[43055]: notice: Peer storage2 was terminated (reboot) by storage1 on behalf of stonith_admin.44835: OK Any ideas what I'm doing wrong? I'd be happy to provide more mogs, if desired. Thanks! Bryan _______________________________________________ Users mailing list: [email protected] https://lists.clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
