IPMI fencing is a very common type, and it shouldn't be so hard to get working. Easiest is to test it out first on the command line, outside pacemaker. Run:
fence_ipmilan -a <ipmi_ip> -l <ipmi_user> -p <password> -o status If that doesn't work, you may need to use lanplus or similar. See 'man fence_ipmilan'. Once you can use the above command to query the power status of the nodes, you're 95% of the way there. Fencing can not be put on the back burner, as you've now seen. Without it, things can and will go very wrong. On 21/09/15 09:34 AM, Jason Gress wrote: > Thank you for comment. I attempted to use iDRAC/IPMI STONITH, and after > spending over a day, I had to put it on the backburner for timeline > reasons. For whatever reason, I could not get IPMI to talk, and the > iDRAC5 plugin was not working either for reasons I don't understand. > > Is that what you had in mind, or is there another method/configuration for > fencing DRBD? > > Thank you for your advice, > > Jason > > On 9/20/15, 9:40 PM, "Digimer" <li...@alteeve.ca> wrote: > >> On 20/09/15 09:18 PM, Jason Gress wrote: >>> I had seemed to cause a split brain attempting to repair this. But that >> >> Use fencing! Voila, no more split-brains. >> >>> wasn't the issue. You can't have any colocation requirements for DRBD >>> resources; that's what killed me. This line did it: >>> >>> ms_drbd_vmfs with ClusterIP (score:INFINITY) >>> (id:colocation-ms_drbd_vmfs-ClusterIP-INFINITY) >>> >>> Do NOT do this! >>> >>> Jason >>> >>> From: Jason Gress <jgr...@accertify.com <mailto:jgr...@accertify.com>> >>> Reply-To: Cluster Labs - All topics related to open-source clustering >>> welcomed <users@clusterlabs.org <mailto:users@clusterlabs.org>> >>> Date: Friday, September 18, 2015 at 3:03 PM >>> To: Cluster Labs - All topics related to open-source clustering welcomed >>> <users@clusterlabs.org <mailto:users@clusterlabs.org>> >>> Subject: Re: [ClusterLabs] Pacemaker/pcs & DRBD not demoting secondary >>> node to Slave (always Stopped) >>> >>> Well, it almost worked. I was able to modify the existing cluster per >>> your command, and it worked great. >>> >>> Today, I made two more clusters via the exact same process (I >>> used/modified my notes as I was building and fixing the first one >>> yesterday) and now it's doing the same thing, despite having your >>> improved master slave rule. Here's the config: >>> >>> [root@fx201-1a ~]# pcs config --full >>> Cluster Name: fx201-vmcl >>> Corosync Nodes: >>> fx201-1a.zwo fx201-1b.zwo >>> Pacemaker Nodes: >>> fx201-1a.zwo fx201-1b.zwo >>> >>> Resources: >>> Resource: ClusterIP (class=ocf provider=heartbeat type=IPaddr2) >>> Attributes: ip=10.XX.XX.XX cidr_netmask=24 >>> Operations: start interval=0s timeout=20s >>> (ClusterIP-start-timeout-20s) >>> stop interval=0s timeout=20s (ClusterIP-stop-timeout-20s) >>> monitor interval=15s (ClusterIP-monitor-interval-15s) >>> Master: ms_drbd_vmfs >>> Meta Attrs: master-max=1 master-node-max=1 clone-max=2 >>> clone-node-max=1 notify=true >>> Resource: drbd_vmfs (class=ocf provider=linbit type=drbd) >>> Attributes: drbd_resource=vmfs >>> Operations: start interval=0s timeout=240 >>> (drbd_vmfs-start-timeout-240) >>> promote interval=0s timeout=90 >>> (drbd_vmfs-promote-timeout-90) >>> demote interval=0s timeout=90 >>> (drbd_vmfs-demote-timeout-90) >>> stop interval=0s timeout=100 (drbd_vmfs-stop-timeout-100) >>> monitor interval=29s role=Master >>> (drbd_vmfs-monitor-interval-29s-role-Master) >>> monitor interval=31s role=Slave >>> (drbd_vmfs-monitor-interval-31s-role-Slave) >>> Resource: vmfsFS (class=ocf provider=heartbeat type=Filesystem) >>> Attributes: device=/dev/drbd0 directory=/exports/vmfs fstype=xfs >>> Operations: start interval=0s timeout=60 (vmfsFS-start-timeout-60) >>> stop interval=0s timeout=60 (vmfsFS-stop-timeout-60) >>> monitor interval=20 timeout=40 >>> (vmfsFS-monitor-interval-20) >>> Resource: nfs-server (class=systemd type=nfs-server) >>> Operations: monitor interval=60s (nfs-server-monitor-interval-60s) >>> >>> Stonith Devices: >>> Fencing Levels: >>> >>> Location Constraints: >>> Ordering Constraints: >>> promote ms_drbd_vmfs then start vmfsFS (kind:Mandatory) >>> (id:order-ms_drbd_vmfs-vmfsFS-mandatory) >>> start vmfsFS then start nfs-server (kind:Mandatory) >>> (id:order-vmfsFS-nfs-server-mandatory) >>> start ClusterIP then start nfs-server (kind:Mandatory) >>> (id:order-ClusterIP-nfs-server-mandatory) >>> Colocation Constraints: >>> ms_drbd_vmfs with ClusterIP (score:INFINITY) >>> (id:colocation-ms_drbd_vmfs-ClusterIP-INFINITY) >>> vmfsFS with ms_drbd_vmfs (score:INFINITY) (with-rsc-role:Master) >>> (id:colocation-vmfsFS-ms_drbd_vmfs-INFINITY) >>> nfs-server with vmfsFS (score:INFINITY) >>> (id:colocation-nfs-server-vmfsFS-INFINITY) >>> nfs-server with ClusterIP (score:INFINITY) >>> (id:colocation-nfs-server-ClusterIP-INFINITY) >>> >>> Cluster Properties: >>> cluster-infrastructure: corosync >>> cluster-name: fx201-vmcl >>> dc-version: 1.1.13-a14efad >>> have-watchdog: false >>> stonith-enabled: false >>> >>> [root@fx201-1a ~]# pcs status --full >>> Cluster name: fx201-vmcl >>> Last updated: Fri Sep 18 15:02:16 2015Last change: Fri Sep 18 14:44:33 >>> 2015 by root via crm_attribute on fx201-1b.zwo >>> Stack: corosync >>> Current DC: fx201-1a.zwo (1) (version 1.1.13-a14efad) - partition with >>> quorum >>> 2 nodes and 5 resources configured >>> >>> Online: [ fx201-1a.zwo (1) fx201-1b.zwo (2) ] >>> >>> Full list of resources: >>> >>> ClusterIP(ocf::heartbeat:IPaddr2):Started fx201-1a.zwo >>> Master/Slave Set: ms_drbd_vmfs [drbd_vmfs] >>> drbd_vmfs(ocf::linbit:drbd):Master fx201-1a.zwo >>> drbd_vmfs(ocf::linbit:drbd):Stopped >>> Masters: [ fx201-1a.zwo ] >>> Stopped: [ fx201-1b.zwo ] >>> vmfsFS(ocf::heartbeat:Filesystem):Started fx201-1a.zwo >>> nfs-server(systemd:nfs-server):Started fx201-1a.zwo >>> >>> PCSD Status: >>> fx201-1a.zwo: Online >>> fx201-1b.zwo: Online >>> >>> Daemon Status: >>> corosync: active/enabled >>> pacemaker: active/enabled >>> pcsd: active/enabled >>> >>> This is so strange... The master/slave rule fixed my other two clusters, >>> but not this one. >>> >>> Thank you all for your advice, >>> >>> Jason >>> >>> From: Jason Gress <jgr...@accertify.com <mailto:jgr...@accertify.com>> >>> Reply-To: Cluster Labs - All topics related to open-source clustering >>> welcomed <users@clusterlabs.org <mailto:users@clusterlabs.org>> >>> Date: Thursday, September 17, 2015 at 7:25 PM >>> To: Cluster Labs - All topics related to open-source clustering welcomed >>> <users@clusterlabs.org <mailto:users@clusterlabs.org>> >>> Subject: Re: [ClusterLabs] Pacemaker/pcs & DRBD not demoting secondary >>> node to Slave (always Stopped) >>> >>> That was Exactly what I needed. Thank you so much! >>> >>> Jason >>> >>> From: Luke Pascoe <l...@osnz.co.nz <mailto:l...@osnz.co.nz>> >>> Reply-To: Cluster Labs - All topics related to open-source clustering >>> welcomed <users@clusterlabs.org <mailto:users@clusterlabs.org>> >>> Date: Thursday, September 17, 2015 at 7:08 PM >>> To: Cluster Labs - All topics related to open-source clustering welcomed >>> <users@clusterlabs.org <mailto:users@clusterlabs.org>> >>> Subject: Re: [ClusterLabs] Pacemaker/pcs & DRBD not demoting secondary >>> node to Slave (always Stopped) >>> >>> pcs resource create drbd_iscsivg0 ocf:linbit:drbd drbd_resource=iscsivg0 >>> op monitor interval="29s" role="Master" op monitor interval="31s" >>> role="Slave" >>> >>> Luke Pascoe >>> >>> >>> * >>> * >>> >>> *E* l...@osnz.co.nz <mailto:l...@osnz.co.nz> >>> *P* +64 (9) 296 2961 >>> *M* +64 (27) 426 6649 >>> *W* www.osnz.co.nz <http://www.osnz.co.nz/> >>> >>> 24 Wellington St >>> Papakura >>> Auckland, 2110 >>> New Zealand >>> >>> >>> On 18 September 2015 at 12:02, Jason Gress <jgr...@accertify.com >>> <mailto:jgr...@accertify.com>> wrote: >>> >>> That may very well be it. Would you be so kind as to show me the >>> pcs command to create that config? I generated my configuration >>> with these commands, and I'm not sure how to get the additional >>> monitor options in there: >>> >>> pcs resource create drbd_vmfs ocf:linbit:drbd drbd_resource=vmfs op >>> monitor interval=30s >>> pcs resource master ms_drbd_vmfs drbd_vmfs master-max=1 >>> master-node-max=1 clone-max=2 clone-node-max=1 notify=true >>> >>> Thank you very much for your help, and sorry for the newbie >>> question! >>> >>> Jason >>> >>> From: Luke Pascoe <l...@osnz.co.nz <mailto:l...@osnz.co.nz>> >>> Reply-To: Cluster Labs - All topics related to open-source >>> clustering welcomed <users@clusterlabs.org >>> <mailto:users@clusterlabs.org>> >>> Date: Thursday, September 17, 2015 at 6:54 PM >>> >>> To: Cluster Labs - All topics related to open-source clustering >>> welcomed <users@clusterlabs.org <mailto:users@clusterlabs.org>> >>> Subject: Re: [ClusterLabs] Pacemaker/pcs & DRBD not demoting >>> secondary node to Slave (always Stopped) >>> >>> The only difference in the DRBD resource between yours and mine that >>> I can see is the monitoring parameters (mine works nicely, but is >>> Centos 6). Here's mine: >>> >>> Master: ms_drbd_iscsicg0 >>> Meta Attrs: master-max=1 master-node-max=1 clone-max=2 >>> clone-node-max=1 notify=true >>> Resource: drbd_iscsivg0 (class=ocf provider=linbit type=drbd) >>> Attributes: drbd_resource=iscsivg0 >>> Operations: start interval=0s timeout=240 >>> (drbd_iscsivg0-start-timeout-240) >>> promote interval=0s timeout=90 >>> (drbd_iscsivg0-promote-timeout-90) >>> demote interval=0s timeout=90 >>> (drbd_iscsivg0-demote-timeout-90) >>> stop interval=0s timeout=100 >>> (drbd_iscsivg0-stop-timeout-100) >>> monitor interval=29s role=Master >>> (drbd_iscsivg0-monitor-interval-29s-role-Master) >>> monitor interval=31s role=Slave >>> (drbd_iscsivg0-monitor-interval-31s-role-Slave) >>> >>> What mechanism are you using to fail over? Check your constraints >>> after you do it and make sure it hasn't added one which stops the >>> slave clone from starting on the "failed" node. >>> >>> >>> Luke Pascoe >>> >>> >>> * >>> * >>> >>> *E* l...@osnz.co.nz <mailto:l...@osnz.co.nz> >>> *P* +64 (9) 296 2961 <tel:%2B64%20%289%29%20296%202961> >>> *M* +64 (27) 426 6649 >>> *W* www.osnz.co.nz <http://www.osnz.co.nz/> >>> >>> 24 Wellington St >>> Papakura >>> Auckland, 2110 >>> New Zealand >>> >>> >>> On 18 September 2015 at 11:40, Jason Gress <jgr...@accertify.com >>> <mailto:jgr...@accertify.com>> wrote: >>> >>> Looking more closely, according to page 64 >>> (http://clusterlabs.org/doc/Cluster_from_Scratch.pdf) it does >>> indeed appear that 1 is the correct number. (I just realized >>> that it's page 64 of the "book", but page 76 of the pdf.) >>> >>> Thank you again, >>> >>> Jason >>> >>> From: Jason Gress <jgr...@accertify.com >>> <mailto:jgr...@accertify.com>> >>> Reply-To: Cluster Labs - All topics related to open-source >>> clustering welcomed <users@clusterlabs.org >>> <mailto:users@clusterlabs.org>> >>> Date: Thursday, September 17, 2015 at 6:36 PM >>> To: Cluster Labs - All topics related to open-source clustering >>> welcomed <users@clusterlabs.org <mailto:users@clusterlabs.org>> >>> Subject: Re: [ClusterLabs] Pacemaker/pcs & DRBD not demoting >>> secondary node to Slave (always Stopped) >>> >>> I can't say whether or not you are right or wrong (you may be >>> right!) but I followed the Cluster From Scratch tutorial >>> closely, and it only had a clone-node-max=1 there. (Page 106 of >>> the pdf, for the curious.) >>> >>> Thanks, >>> >>> Jason >>> >>> From: Luke Pascoe <l...@osnz.co.nz <mailto:l...@osnz.co.nz>> >>> Reply-To: Cluster Labs - All topics related to open-source >>> clustering welcomed <users@clusterlabs.org >>> <mailto:users@clusterlabs.org>> >>> Date: Thursday, September 17, 2015 at 6:29 PM >>> To: Cluster Labs - All topics related to open-source clustering >>> welcomed <users@clusterlabs.org <mailto:users@clusterlabs.org>> >>> Subject: Re: [ClusterLabs] Pacemaker/pcs & DRBD not demoting >>> secondary node to Slave (always Stopped) >>> >>> I may be wrong, but shouldn't "clone-node-max" be 2 on >>> the ms_drbd_vmfs resource? >>> >>> Luke Pascoe >>> >>> >>> * >>> * >>> >>> *E* l...@osnz.co.nz <mailto:l...@osnz.co.nz> >>> *P* +64 (9) 296 2961 <tel:%2B64%20%289%29%20296%202961> >>> *M* +64 (27) 426 6649 >>> *W* www.osnz.co.nz <http://www.osnz.co.nz/> >>> >>> 24 Wellington St >>> Papakura >>> Auckland, 2110 >>> New Zealand >>> >>> >>> On 18 September 2015 at 11:02, Jason Gress <jgr...@accertify.com >>> <mailto:jgr...@accertify.com>> wrote: >>> >>> I have a simple DRBD + filesystem + NFS configuration that >>> works properly when I manually start/stop DRBD, but will not >>> start the DRBD slave resource properly on failover or >>> recovery. I cannot ever get the Master/Slave set to say >>> anything but 'Stopped'. I am running ?CentOS 7.1 with the >>> latest packages as of today: >>> >>> [root@fx201-1a log]# rpm -qa | grep -e pcs -e pacemaker -e >>> drbd >>> pacemaker-cluster-libs-1.1.12-22.el7_1.4.x86_64 >>> pacemaker-1.1.12-22.el7_1.4.x86_64 >>> pcs-0.9.137-13.el7_1.4.x86_64 >>> pacemaker-libs-1.1.12-22.el7_1.4.x86_64 >>> drbd84-utils-8.9.3-1.1.el7.elrepo.x86_64 >>> pacemaker-cli-1.1.12-22.el7_1.4.x86_64 >>> kmod-drbd84-8.4.6-1.el7.elrepo.x86_64 >>> >>> Here is my pcs config output: >>> >>> [root@fx201-1a log]# pcs config >>> Cluster Name: fx201-vmcl >>> Corosync Nodes: >>> fx201-1a.ams fx201-1b.ams >>> Pacemaker Nodes: >>> fx201-1a.ams fx201-1b.ams >>> >>> Resources: >>> Resource: ClusterIP (class=ocf provider=heartbeat >>> type=IPaddr2) >>> Attributes: ip=10.XX.XX.XX cidr_netmask=24 >>> Operations: start interval=0s timeout=20s >>> (ClusterIP-start-timeout-20s) >>> stop interval=0s timeout=20s >>> (ClusterIP-stop-timeout-20s) >>> monitor interval=15s >>> (ClusterIP-monitor-interval-15s) >>> Master: ms_drbd_vmfs >>> Meta Attrs: master-max=1 master-node-max=1 clone-max=2 >>> clone-node-max=1 notify=true >>> Resource: drbd_vmfs (class=ocf provider=linbit type=drbd) >>> Attributes: drbd_resource=vmfs >>> Operations: start interval=0s timeout=240 >>> (drbd_vmfs-start-timeout-240) >>> promote interval=0s timeout=90 >>> (drbd_vmfs-promote-timeout-90) >>> demote interval=0s timeout=90 >>> (drbd_vmfs-demote-timeout-90) >>> stop interval=0s timeout=100 >>> (drbd_vmfs-stop-timeout-100) >>> monitor interval=30s >>> (drbd_vmfs-monitor-interval-30s) >>> Resource: vmfsFS (class=ocf provider=heartbeat >>> type=Filesystem) >>> Attributes: device=/dev/drbd0 directory=/exports/vmfs >>> fstype=xfs >>> Operations: start interval=0s timeout=60 >>> (vmfsFS-start-timeout-60) >>> stop interval=0s timeout=60 >>> (vmfsFS-stop-timeout-60) >>> monitor interval=20 timeout=40 >>> (vmfsFS-monitor-interval-20) >>> Resource: nfs-server (class=systemd type=nfs-server) >>> Operations: monitor interval=60s >>> (nfs-server-monitor-interval-60s) >>> >>> Stonith Devices: >>> Fencing Levels: >>> >>> Location Constraints: >>> Ordering Constraints: >>> promote ms_drbd_vmfs then start vmfsFS (kind:Mandatory) >>> (id:order-ms_drbd_vmfs-vmfsFS-mandatory) >>> start vmfsFS then start nfs-server (kind:Mandatory) >>> (id:order-vmfsFS-nfs-server-mandatory) >>> start ClusterIP then start nfs-server (kind:Mandatory) >>> (id:order-ClusterIP-nfs-server-mandatory) >>> Colocation Constraints: >>> ms_drbd_vmfs with ClusterIP (score:INFINITY) >>> (id:colocation-ms_drbd_vmfs-ClusterIP-INFINITY) >>> vmfsFS with ms_drbd_vmfs (score:INFINITY) >>> (with-rsc-role:Master) >>> (id:colocation-vmfsFS-ms_drbd_vmfs-INFINITY) >>> nfs-server with vmfsFS (score:INFINITY) >>> (id:colocation-nfs-server-vmfsFS-INFINITY) >>> >>> Cluster Properties: >>> cluster-infrastructure: corosync >>> cluster-name: fx201-vmcl >>> dc-version: 1.1.13-a14efad >>> have-watchdog: false >>> last-lrm-refresh: 1442528181 >>> stonith-enabled: false >>> >>> And status: >>> >>> [root@fx201-1a log]# pcs status --full >>> Cluster name: fx201-vmcl >>> Last updated: Thu Sep 17 17:55:56 2015Last change: Thu Sep >>> 17 17:18:10 2015 by root via crm_attribute on fx201-1b.ams >>> Stack: corosync >>> Current DC: fx201-1b.ams (2) (version 1.1.13-a14efad) - >>> partition with quorum >>> 2 nodes and 5 resources configured >>> >>> Online: [ fx201-1a.ams (1) fx201-1b.ams (2) ] >>> >>> Full list of resources: >>> >>> ClusterIP(ocf::heartbeat:IPaddr2):Started fx201-1a.ams >>> Master/Slave Set: ms_drbd_vmfs [drbd_vmfs] >>> drbd_vmfs(ocf::linbit:drbd):Master fx201-1a.ams >>> drbd_vmfs(ocf::linbit:drbd):Stopped >>> Masters: [ fx201-1a.ams ] >>> Stopped: [ fx201-1b.ams ] >>> vmfsFS(ocf::heartbeat:Filesystem):Started fx201-1a.ams >>> nfs-server(systemd:nfs-server):Started fx201-1a.ams >>> >>> PCSD Status: >>> fx201-1a.ams: Online >>> fx201-1b.ams: Online >>> >>> Daemon Status: >>> corosync: active/enabled >>> pacemaker: active/enabled >>> pcsd: active/enabled >>> >>> If I do a failover, after manually confirming that the DRBD >>> data is synchronized completely, it does work, but then >>> never reconnects the secondary side, and in order to get the >>> resource synchronized again, I have to manually correct it, >>> ad infinitum. I have tried standby/unstandby, pcs resource >>> debug-start (with undesirable results), and so on. >>> >>> Here are some relevant log messages from pacemaker.log: >>> >>> Sep 17 17:48:10 [13954] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> crmd: info: >>> crm_timer_popped:PEngine Recheck Timer (I_PE_CALC) just >>> popped (900000ms) >>> Sep 17 17:48:10 [13954] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> crmd: notice: >>> do_state_transition:State transition S_IDLE -> >>> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED >>> origin=crm_timer_popped ] >>> Sep 17 17:48:10 [13954] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> crmd: info: >>> do_state_transition:Progressed to state S_POLICY_ENGINE >>> after C_TIMER_POPPED >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> process_pe_message:Input has not changed since last time, >>> not saving to disk >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> determine_online_status:Node fx201-1b.ams is online >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> determine_online_status:Node fx201-1a.ams is online >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> determine_op_status:Operation monitor found resource >>> drbd_vmfs:0 active in master mode on fx201-1b.ams >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> determine_op_status:Operation monitor found resource >>> drbd_vmfs:0 active on fx201-1a.ams >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> native_print:ClusterIP(ocf::heartbeat:IPaddr2):Started >>> fx201-1a.ams >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> clone_print:Master/Slave Set: ms_drbd_vmfs [drbd_vmfs] >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> short_print: Masters: [ fx201-1a.ams ] >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> short_print: Stopped: [ fx201-1b.ams ] >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> native_print:vmfsFS(ocf::heartbeat:Filesystem):Started >>> fx201-1a.ams >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> native_print:nfs-server(systemd:nfs-server):Started >>> fx201-1a.ams >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> native_color:Resource drbd_vmfs:1 cannot run anywhere >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> master_color:Promoting drbd_vmfs:0 (Master fx201-1a.ams) >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> master_color:ms_drbd_vmfs: Promoted 1 instances of a >>> possible 1 to master >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> LogActions:Leave ClusterIP(Started fx201-1a.ams) >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> LogActions:Leave drbd_vmfs:0(Master fx201-1a.ams) >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> LogActions:Leave drbd_vmfs:1(Stopped) >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> LogActions:Leave vmfsFS(Started fx201-1a.ams) >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: info: >>> LogActions:Leave nfs-server(Started fx201-1a.ams) >>> Sep 17 17:48:10 [5662] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> pengine: notice: >>> process_pe_message:Calculated Transition 16: >>> /var/lib/pacemaker/pengine/pe-input-61.bz2 >>> Sep 17 17:48:10 [13954] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> crmd: info: >>> do_state_transition:State transition S_POLICY_ENGINE -> >>> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE >>> origin=handle_response ] >>> Sep 17 17:48:10 [13954] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> crmd: info: >>> do_te_invoke:Processing graph 16 >>> (ref=pe_calc-dc-1442530090-97) derived from >>> /var/lib/pacemaker/pengine/pe-input-61.bz2 >>> Sep 17 17:48:10 [13954] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> crmd: notice: >>> run_graph:Transition 16 (Complete=0, Pending=0, Fired=0, >>> Skipped=0, Incomplete=0, >>> Source=/var/lib/pacemaker/pengine/pe-input-61.bz2): Complete >>> Sep 17 17:48:10 [13954] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> crmd: info: >>> do_log:FSA: Input I_TE_SUCCESS from notify_crmd() received >>> in state S_TRANSITION_ENGINE >>> Sep 17 17:48:10 [13954] fx201-1b.ams.accertify.net >>> <http://fx201-1b.ams.accertify.net> crmd: notice: >>> do_state_transition:State transition S_TRANSITION_ENGINE -> >>> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL >>> origin=notify_crmd ] >>> >>> Thank you all for your help, >>> >>> Jason >>> >>> "This message and any attachments may contain confidential >>> information. If you >>> have received this message in error, any use or >>> distribution is prohibited. >>> Please notify us by reply e-mail if you have mistakenly >>> received this message, >>> and immediately and permanently delete it and any >>> attachments. Thank you." >>> >>> >>> _______________________________________________ >>> Users mailing list: Users@clusterlabs.org >>> <mailto:Users@clusterlabs.org> >>> http://clusterlabs.org/mailman/listinfo/users >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: >>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >>> >>> >>> "This message and any attachments may contain confidential >>> information. If you >>> have received this message in error, any use or distribution >>> is prohibited. >>> Please notify us by reply e-mail if you have mistakenly >>> received this message, >>> and immediately and permanently delete it and any attachments. >>> Thank you." >>> >>> "This message and any attachments may contain confidential >>> information. If you >>> have received this message in error, any use or distribution >>> is prohibited. >>> Please notify us by reply e-mail if you have mistakenly >>> received this message, >>> and immediately and permanently delete it and any attachments. >>> Thank you." >>> >>> >>> _______________________________________________ >>> Users mailing list: Users@clusterlabs.org >>> <mailto:Users@clusterlabs.org> >>> http://clusterlabs.org/mailman/listinfo/users >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: >>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >>> >>> >>> "This message and any attachments may contain confidential >>> information. If you >>> have received this message in error, any use or distribution is >>> prohibited. >>> Please notify us by reply e-mail if you have mistakenly received >>> this message, >>> and immediately and permanently delete it and any attachments. >>> Thank you." >>> >>> >>> _______________________________________________ >>> Users mailing list: Users@clusterlabs.org >>> <mailto:Users@clusterlabs.org> >>> http://clusterlabs.org/mailman/listinfo/users >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: >>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >>> >>> >>> "This message and any attachments may contain confidential information. >>> If you >>> have received this message in error, any use or distribution is >>> prohibited. >>> Please notify us by reply e-mail if you have mistakenly received this >>> message, >>> and immediately and permanently delete it and any attachments. Thank >>> you." >>> >>> "This message and any attachments may contain confidential information. >>> If you >>> have received this message in error, any use or distribution is >>> prohibited. >>> Please notify us by reply e-mail if you have mistakenly received this >>> message, >>> and immediately and permanently delete it and any attachments. Thank >>> you." >>> >>> >>> "This message and any attachments may contain confidential information. >>> If you >>> have received this message in error, any use or distribution is >>> prohibited. >>> Please notify us by reply e-mail if you have mistakenly received this >>> message, >>> and immediately and permanently delete it and any attachments. Thank >>> you." >>> >>> >>> >>> _______________________________________________ >>> Users mailing list: Users@clusterlabs.org >>> http://clusterlabs.org/mailman/listinfo/users >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >>> >> >> >> -- >> Digimer >> Papers and Projects: https://alteeve.ca/w/ >> What if the cure for cancer is trapped in the mind of a person without >> access to education? >> >> _______________________________________________ >> Users mailing list: Users@clusterlabs.org >> http://clusterlabs.org/mailman/listinfo/users >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org > > > > > "This message and any attachments may contain confidential information. If you > have received this message in error, any use or distribution is prohibited. > Please notify us by reply e-mail if you have mistakenly received this message, > and immediately and permanently delete it and any attachments. Thank you." > > _______________________________________________ > Users mailing list: Users@clusterlabs.org > http://clusterlabs.org/mailman/listinfo/users > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > -- Digimer Papers and Projects: https://alteeve.ca/w/ What if the cure for cancer is trapped in the mind of a person without access to education? _______________________________________________ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org