Re: [Pacemaker] Fencing dependency between bare metal host and its VMs guest
В Mon, 10 Nov 2014 10:07:18 +0100 Tomasz Kontusz пишет: > I think the suggestion was to put shooting the host in the fencing path of a > VM. This way if you can't get the host to fence the VM (as the host is > already dead) you just check if the host was fenced. > Exactly. One thing I do not know how it will behave in case of multiple VMs on the same host. I.e. will pacemaker try to fence host for every VM or recognize that all VMs are dead after the first time agent is invoked. > Daniel Dehennin napisał: > >Andrei Borzenkov writes: > > > > > >[...] > > > >>> Now I have one issue, when the bare metal host on which the VM is > >>> running die, the VM is lost and can not be fenced. > >>> > >>> Is there a way to make pacemaker ACK the fencing of the VM running > >on a > >>> host when the host is fenced itself? > >>> > >> > >> Yes, you can define multiple stonith agents and priority between > >them. > >> > >> http://clusterlabs.org/wiki/Fencing_topology > > > >Hello, > > > >If I understand correctly, fencing topology is the way to have several > >fencing devices for a node and try them consecutively until one works. > > > >In my configuration, I group the VM stonith agents with the > >corresponding VM resource, to make them move together[1]. > > > >Here is my use case: > > > >1. Resource ONE-Frontend-Group runs on nebula1 > >2. nebula1 is fenced > >3. node one-fronted can not be fenced > > > >Is there a way to say that the life on node one-frontend is related to > >the state of resource ONE-Frontend? > > > >In which case when the node nebula1 is fenced, pacemaker should be > >aware that > >resource ONE-Frontend is not running any more, so node one-frontend is > >OFFLINE and not UNCLEAN. > > > >Regards. > > > >Footnotes: > >[1] > >http://oss.clusterlabs.org/pipermail/pacemaker/2014-October/022671.html > > > >-- > >Daniel Dehennin > >Récupérer ma clef GPG: gpg --recv-keys 0xCC1E9E5B7A6FE2DF > >Fingerprint: 3E69 014E 5C23 50E8 9ED6 2AAD CC1E 9E5B 7A6F E2DF > > > > > > > > > > > >___ > >Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > >http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > > >Project Home: http://www.clusterlabs.org > >Getting started: > >http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > >Bugs: http://bugs.clusterlabs.org > ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
Re: [Pacemaker] Fencing dependency between bare metal host and its VMs guest
I think the suggestion was to put shooting the host in the fencing path of a VM. This way if you can't get the host to fence the VM (as the host is already dead) you just check if the host was fenced. Daniel Dehennin napisał: >Andrei Borzenkov writes: > > >[...] > >>> Now I have one issue, when the bare metal host on which the VM is >>> running die, the VM is lost and can not be fenced. >>> >>> Is there a way to make pacemaker ACK the fencing of the VM running >on a >>> host when the host is fenced itself? >>> >> >> Yes, you can define multiple stonith agents and priority between >them. >> >> http://clusterlabs.org/wiki/Fencing_topology > >Hello, > >If I understand correctly, fencing topology is the way to have several >fencing devices for a node and try them consecutively until one works. > >In my configuration, I group the VM stonith agents with the >corresponding VM resource, to make them move together[1]. > >Here is my use case: > >1. Resource ONE-Frontend-Group runs on nebula1 >2. nebula1 is fenced >3. node one-fronted can not be fenced > >Is there a way to say that the life on node one-frontend is related to >the state of resource ONE-Frontend? > >In which case when the node nebula1 is fenced, pacemaker should be >aware that >resource ONE-Frontend is not running any more, so node one-frontend is >OFFLINE and not UNCLEAN. > >Regards. > >Footnotes: >[1] >http://oss.clusterlabs.org/pipermail/pacemaker/2014-October/022671.html > >-- >Daniel Dehennin >Récupérer ma clef GPG: gpg --recv-keys 0xCC1E9E5B7A6FE2DF >Fingerprint: 3E69 014E 5C23 50E8 9ED6 2AAD CC1E 9E5B 7A6F E2DF > > > > > >___ >Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >http://oss.clusterlabs.org/mailman/listinfo/pacemaker > >Project Home: http://www.clusterlabs.org >Getting started: >http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >Bugs: http://bugs.clusterlabs.org -- Wysłane za pomocą K-9 Mail.___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
Re: [Pacemaker] Fencing dependency between bare metal host and its VMs guest
Andrei Borzenkov writes: [...] >> Now I have one issue, when the bare metal host on which the VM is >> running die, the VM is lost and can not be fenced. >> >> Is there a way to make pacemaker ACK the fencing of the VM running on a >> host when the host is fenced itself? >> > > Yes, you can define multiple stonith agents and priority between them. > > http://clusterlabs.org/wiki/Fencing_topology Hello, If I understand correctly, fencing topology is the way to have several fencing devices for a node and try them consecutively until one works. In my configuration, I group the VM stonith agents with the corresponding VM resource, to make them move together[1]. Here is my use case: 1. Resource ONE-Frontend-Group runs on nebula1 2. nebula1 is fenced 3. node one-fronted can not be fenced Is there a way to say that the life on node one-frontend is related to the state of resource ONE-Frontend? In which case when the node nebula1 is fenced, pacemaker should be aware that resource ONE-Frontend is not running any more, so node one-frontend is OFFLINE and not UNCLEAN. Regards. Footnotes: [1] http://oss.clusterlabs.org/pipermail/pacemaker/2014-October/022671.html -- Daniel Dehennin Récupérer ma clef GPG: gpg --recv-keys 0xCC1E9E5B7A6FE2DF Fingerprint: 3E69 014E 5C23 50E8 9ED6 2AAD CC1E 9E5B 7A6F E2DF node $id="1084811078" nebula1 node $id="1084811079" nebula2 node $id="1084811080" nebula3 node $id="108488" quorum \ attributes standby="on" node $id="108489" one-frontend primitive ONE-Datastores ocf:heartbeat:Filesystem \ params device="/dev/one-fs/datastores" directory="/var/lib/one/datastores" fstype="gfs2" \ op start interval="0" timeout="90" \ op stop interval="0" timeout="100" \ op monitor interval="20" timeout="40" primitive ONE-Frontend ocf:heartbeat:VirtualDomain \ params config="/var/lib/one/datastores/one/one.xml" \ op start interval="0" timeout="90" \ op stop interval="0" timeout="100" \ utilization cpu="1" hv_memory="1024" primitive ONE-vg ocf:heartbeat:LVM \ params volgrpname="one-fs" \ op start interval="0" timeout="30" \ op stop interval="0" timeout="30" \ op monitor interval="60" timeout="30" primitive Quorum-Node ocf:heartbeat:VirtualDomain \ params config="/var/lib/libvirt/qemu/pcmk/quorum.xml" \ op start interval="0" timeout="90" \ op stop interval="0" timeout="100" \ utilization cpu="1" hv_memory="1024" primitive Stonith-ONE-Frontend stonith:external/libvirt \ params hostlist="one-frontend" hypervisor_uri="qemu:///system" pcmk_host_list="one-frontend" pcmk_host_check="static-list" \ op monitor interval="30m" primitive Stonith-Quorum-Node stonith:external/libvirt \ params hostlist="quorum" hypervisor_uri="qemu:///system" pcmk_host_list="quorum" pcmk_host_check="static-list" \ op monitor interval="30m" primitive Stonith-nebula1-IPMILAN stonith:external/ipmi \ params hostname="nebula1-ipmi" ipaddr="XXX.XXX.XXX.XXX" interface="lanplus" userid="USER" passwd="PASSWORD1" passwd_method="env" priv="operator" pcmk_host_list="nebula1" pcmk_host_check="static-list" \ op monitor interval="30m" \ meta target-role="Started" primitive Stonith-nebula2-IPMILAN stonith:external/ipmi \ params hostname="nebula2-ipmi" ipaddr="YYY.YYY.YYY.YYY" interface="lanplus" userid="USER" passwd="PASSWORD2" passwd_method="env" priv="operator" pcmk_host_list="nebula2" pcmk_host_check="static-list" \ op monitor interval="30m" \ meta target-role="Started" primitive Stonith-nebula3-IPMILAN stonith:external/ipmi \ params hostname="nebula3-ipmi" ipaddr="ZZZ.ZZZ.ZZZ.ZZZ" interface="lanplus" userid="USER" passwd="PASSWORD3" passwd_method="env" priv="operator" pcmk_host_list="nebula3" pcmk_host_check="static-list" \ op monitor interval="30m" \ meta target-role="Started" primitive clvm ocf:lvm2:clvmd \ op start interval="0" timeout="90" \ op stop interval="0" timeout="100" \ op monitor interval="60" timeout="90" primitive dlm ocf:pacemaker:controld \ op start interval="0" timeout="90" \ op stop interval="0" timeout="100" \ op monitor interval="60" timeout="60" group ONE-Frontend-Group Stonith-ONE-Frontend ONE-Frontend \ meta target-role="Started" group ONE-Storage dlm clvm ONE-vg ONE-Datastores group Quorum-Node-Group Stonith-Quorum-Node Quorum-Node \ meta target-role="Started" clone ONE-Storage-Clone ONE-Storage \ meta interleave="true" target-role="Started" location Nebula1-does-not-fence-itslef Stonith-nebula1-IPMILAN \ rule $id="Nebula1-does-not-fence-itslef-rule" 50: #uname eq nebula2 \ rule $id="Nebula1-does-not-fence-itslef-rule-0" 40: #uname eq nebula3 location Nebula2-does-not-fence-itslef Stonith-nebula2-IPMILAN \ rule $id="Nebula2-does-not-fence-itslef-rule" 50: #una
Re: [Pacemaker] Fencing dependency between bare metal host and its VMs guest
В Fri, 07 Nov 2014 17:46:40 +0100 Daniel Dehennin пишет: > Hello, > > As I finally manage to integrate my VM to corosync and my dlm/clvm/GFS2 > are running on it. > > Now I have one issue, when the bare metal host on which the VM is > running die, the VM is lost and can not be fenced. > > Is there a way to make pacemaker ACK the fencing of the VM running on a > host when the host is fenced itself? > Yes, you can define multiple stonith agents and priority between them. http://clusterlabs.org/wiki/Fencing_topology > Regards. > signature.asc Description: PGP signature ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
[Pacemaker] Fencing dependency between bare metal host and its VMs guest
Hello, As I finally manage to integrate my VM to corosync and my dlm/clvm/GFS2 are running on it. Now I have one issue, when the bare metal host on which the VM is running die, the VM is lost and can not be fenced. Is there a way to make pacemaker ACK the fencing of the VM running on a host when the host is fenced itself? Regards. -- Daniel Dehennin Récupérer ma clef GPG: gpg --recv-keys 0xCC1E9E5B7A6FE2DF Fingerprint: 3E69 014E 5C23 50E8 9ED6 2AAD CC1E 9E5B 7A6F E2DF signature.asc Description: PGP signature ___ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org