On 09/04/2013, at 7:31 PM, Oriol Mula-Valls <oriol.mula-va...@ic3.cat> wrote:
> Thanks Andrew I've managed to set up the system and currently I have it > working but still on testing. > > I have configure external/ipmi as fencing device and then I force a reboot > doing a echo b > /proc/sysrq-trigger. The fencing is working properly as the > node is shut off and the VM migrated. However, as soon as I turn on the > fenced now and the OS has started the surviving is shut down. Is it normal or > am I doing something wrong? Can you clarify "turn on the fenced"? > > On the other hand I've seen that in case I completely lose power fencing > obviously fails. Would SBD stonith solve this issue? > > Kind regards, > Oriol > > On 08/04/13 04:11, Andrew Beekhof wrote: >> >> On 03/04/2013, at 9:15 PM, Oriol Mula-Valls<oriol.mula-va...@ic3.cat> wrote: >> >>> Hi, >>> >>> I've started with Linux HA about one year ago. Currently I'm facing a new >>> project in which I have to set up two nodes with high available virtual >>> machines. I have used as a starting point the Digimer's tutorial >>> (https://alteeve.ca/w/2-Node_Red_Hat_KVM_Cluster_Tutorial). >>> >>> To deploy this new infrastructure I have two Fujitsu Primergy Rx100S7. Both >>> machines have 8GB of RAM and 2x500GB HD. I started creating a software >>> RAID1 with the internal drives and installing Debian 7.0 (Wheezy). Apart >>> from the O.S. partition I have created 3 more partitions, one for the >>> shared storage between both machines with OCFS2 and the two other will be >>> used as PVs to create LVs to support the VMs (one for the VMs that will be >>> primary on node1 an the other for primary machines on node2). These 3 >>> partitions are replicated using DRBD. >>> >>> The shared storage folder contains: >>> * ISO images needed when provisioning VMs >>> * scripts used to call virt-install which handles the creation of our VMs. >>> * XML definition files which define the emulated hardware backing the VMs >>> * old copies of the XML definition files. >>> >>> I have more or less done the configuration for the OCFS2 fs and I was about >>> to start the configuration of cLVM for one of the VGs but I have some >>> doubts. I have one dlm for the OCFS2 filesystem, should I create another >>> for cLVM RA? >> >> No, there should only ever be one dlm resource (cloned like you have it) >> >>> >>> This is the current configuration: >>> node node1 >>> node node2 >>> primitive p_dlm_controld ocf:pacemaker:controld \ >>> op start interval="0" timeout="90" \ >>> op stop interval="0" timeout="100" \ >>> op monitor interval="10" >>> primitive p_drbd_shared ocf:linbit:drbd \ >>> params drbd_resource="shared" \ >>> op monitor interval="10" role="Master" timeout="20" \ >>> op monitor interval="20" role="Slave" timeout="20" \ >>> op start interval="0" timeout="240s" \ >>> op stop interval="0" timeout="120s" >>> primitive p_drbd_vm_1 ocf:linbit:drbd \ >>> params drbd_resource="vm_1" \ >>> op monitor interval="10" role="Master" timeout="20" \ >>> op monitor interval="20" role="Slave" timeout="20" \ >>> op start interval="0" timeout="240s" \ >>> op stop interval="0" timeout="120s" >>> primitive p_fs_shared ocf:heartbeat:Filesystem \ >>> params device="/dev/drbd/by-res/shared" directory="/shared" >>> fstype="ocfs2" \ >>> meta target-role="Started" \ >>> op monitor interval="10" >>> primitive p_ipmi_node1 stonith:external/ipmi \ >>> params hostname="node1" userid="admin" passwd="xxx" ipaddr="10.0.0.2" >>> interface="lanplus" >>> primitive p_ipmi_node2 stonith:external/ipmi \ >>> params hostname="node2" userid="admin" passwd="xxx" ipaddr="10.0.0.3" >>> interface="lanplus" >>> primitive p_libvirtd lsb:libvirt-bin \ >>> op monitor interval="120s" \ >>> op start interval="0" \ >>> op stop interval="0" >>> primitive p_o2cb ocf:pacemaker:o2cb \ >>> op start interval="0" timeout="90" \ >>> op stop interval="0" timeout="100" \ >>> op monitor interval="10" \ >>> meta target-role="Started" >>> group g_shared p_dlm_controld p_o2cb p_fs_shared >>> ms ms_drbd_shared p_drbd_shared \ >>> meta master-max="2" clone-max="2" notify="true" >>> ms ms_drbd_vm_1 p_drbd_vm_1 \ >>> meta master-max="2" clone-max="2" notify="true" >>> clone cl_libvirtd p_libvirtd \ >>> meta globally-unique="false" interlave="true" >>> clone cl_shared g_shared \ >>> meta interleave="true" >>> location l_ipmi_node1 p_ipmi_node1 -inf: node1 >>> location l_ipmi_node2 p_ipmi_node2 -inf: node2 >>> order o_drbd_before_shared inf: ms_drbd_shared:promote cl_shared:start >>> >>> Packages' versions: >>> clvm 2.02.95-7 >>> corosync 1.4.2-3 >>> dlm-pcmk 3.0.12-3.2+deb7u2 >>> drbd8-utils 2:8.3.13-2 >>> libdlm3 3.0.12-3.2+deb7u2 >>> libdlmcontrol3 3.0.12-3.2+deb7u2 >>> ocfs2-tools 1.6.4-1+deb7u1 >>> ocfs2-tools-pacemaker 1.6.4-1+deb7u1 >>> openais 1.1.4-4.1 >>> pacemaker 1.1.7-1 >>> >>> As this is my first serious set up suggestions are more than welcome. >>> >>> Thanks for your help. >>> >>> Oriol >>> -- >>> Oriol Mula Valls >>> Institut Català de Ciències del Clima (IC3) >>> Doctor Trueta 203 - 08005 Barcelona >>> Tel:+34 93 567 99 77 >>> >>> _______________________________________________ >>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >> >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> > > > -- > Oriol Mula Valls > Institut Català de Ciències del Clima (IC3) > Doctor Trueta 203 - 08005 Barcelona > Tel:+34 93 567 99 77 > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org