Hello. The HA-LVM + clvmd is working!!!
The Vg should be tagged as cluster. The major problem I noticed when I checked the cluster is that on the passive system I am able to run vgchange -c n vgxx ; vgchange -a y vgxx ; mount the LV out of the cluster - While this lv is mounted on the active member. Shalom. On Sat, Feb 5, 2011 at 12:24 PM, dOminic <[email protected]> wrote: > Hi, > > If you are you using traditional HA-LVM setup with locking_type = 1 , then > you need to setup tagging / initrd rebuild > Anything went wrong with tagging.... then it won't prevent admin to > activate vg, remove the VG which is activated/used in another node. > > # vgchange -ay domvg > 1 logical volume(s) in volume group "domvg" now active > # lvremove /dev/domvg/domlv > Do you really want to remove active logical volume domlv? [y/n]: y > Logical volume "domlv" successfully removed > > If you are using clvmd variant with locking_type = 3 and cluster is > *active/passive* then clvmd won't let you to remove the VG which is > activate/used in another node. > > # lvremove /dev/domvg/domlv > Do you really want to remove active clustered logical volume domlv? [y/n]: > y > Error locking on node node1.domtest.com: LV domvg/domlv in use: not > deactivating > Unable to deactivate logical volume "domlv" > > - dominic > > On Sat, Feb 5, 2011 at 1:07 PM, שלום קלמר <[email protected]> wrote: > >> Hi. >> >> After reading the manual again ,I think i know what was my setup >> problem. >> >> The VGs should be taged as cluster , ( this action will path the lvm.sh >> check ) & the LVs should be deactivated. >> >> The cluster will activate the LVs as exclusive. (i will check it on >> monday ). >> >> Shalom. >> >> >> On Sat, Feb 5, 2011 at 12:07 AM, שלום קלמר <[email protected]> wrote: >> >>> Hello Dominic. >>> >>> I will be in lab on monday and collect all steps & logs files. >>> >>> btw - Is it redhat recommendation to preferre using HA-LVM with the >>> locking_type=1 method ?? >>> >>> >>> Regards >>> >>> Shalom >>> >>> >>> On Fri, Feb 4, 2011 at 5:23 PM, dOminic <[email protected]> wrote: >>> >>>> Hi Shalom, >>>> >>>> Are you still facing problem implementing HA-LVM with locking_type = 3 >>>> setting ?. If yes, it would be great if you could provide the following >>>> details . >>>> So that others can also check >>>> >>>> * steps you are following along with complete output >>>> * status in "clustat" after making changes in cluster.conf >>>> * attach cluster.conf and /var/log/messages. >>>> >>>> dominic >>>> >>>> On Fri, Feb 4, 2011 at 8:34 PM, שלום קלמר <[email protected]> wrote: >>>> >>>>> Thanks. >>>>> >>>>> This is the old methos and its work great but its hard to maintain such >>>>> cluster. >>>>> >>>>> Shalom. >>>>> >>>>> >>>>> On Fri, Feb 4, 2011 at 4:32 PM, Dominic Geevarghese < >>>>> [email protected]> wrote: >>>>> >>>>>> >>>>>> Hi, >>>>>> >>>>>> I am not sure about the error you are getting but it would be great if >>>>>> you could try the preferred method >>>>>> >>>>>> locking_type = 1 >>>>>> volume_list [ "your-root-vg-name" , "@hostname" ] >>>>>> >>>>>> rebuild initrd >>>>>> >>>>>> add the <lvm> and <fs> resources in cluster.conf , start the cman, >>>>>> rgmanager . >>>>>> >>>>>> >>>>>> Thanks, >>>>>> >>>>>> On Thu, Feb 3, 2011 at 8:02 PM, Corey Kovacs >>>>>> <[email protected]>wrote: >>>>>> >>>>>>> Excellent, >>>>>>> >>>>>>> >>>>>>> Thanks >>>>>>> >>>>>>> -C >>>>>>> >>>>>>> On Thu, Feb 3, 2011 at 10:38 AM, שלום קלמר <[email protected]> >>>>>>> wrote: >>>>>>> > >>>>>>> > >>>>>>> > On Thu, Feb 3, 2011 at 12:35 PM, שלום קלמר <[email protected]> >>>>>>> wrote: >>>>>>> >> >>>>>>> >> >>>>>>> >> >>>>>>> >> https://access.redhat.com/kb/docs/DOC-3068 >>>>>>> >> >>>>>>> >> On Thu, Feb 3, 2011 at 11:13 AM, Corey Kovacs < >>>>>>> [email protected]> >>>>>>> >> wrote: >>>>>>> >>> >>>>>>> >>> Is using ha-lvm with clvmd a new capability? It's always been my >>>>>>> >>> understanding that the lvm locking type for using ha-lvm had to >>>>>>> be set >>>>>>> >>> to '1'. >>>>>>> >>> >>>>>>> >>> I'd much rather be using clvmd if it is the way to go. Can you >>>>>>> point >>>>>>> >>> me to the docs you are seeing these instructions in please? >>>>>>> >>> >>>>>>> >>> As for why your config isn't working, clvmd requires that it's >>>>>>> >>> resources are indeed tagged as cluster volumes, so you might try >>>>>>> doing >>>>>>> >>> that and see how it goes. >>>>>>> >>> >>>>>>> >>> -C >>>>>>> >>> >>>>>>> >>> On Thu, Feb 3, 2011 at 7:26 AM, שלום קלמר <[email protected]> >>>>>>> wrote: >>>>>>> >>> > Hello. >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > I followed redhat instruction trying install HA-LVM with clvmd. >>>>>>> ( rhcs >>>>>>> >>> > 5.6 - >>>>>>> >>> > rgmanager 2.0.52-9 ) >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > I can't make it work. >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > lvm.conf- locking_type=3 >>>>>>> >>> > >>>>>>> >>> > clvmd work >>>>>>> >>> > >>>>>>> >>> > Its failed saying HA-LVM is not configured correctly. >>>>>>> >>> > >>>>>>> >>> > The manual said that we should run "lvchange -a n lvxx" edit >>>>>>> the >>>>>>> >>> > cluster.conf & start the service. >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > But From lvm.conf : >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > case $1 in >>>>>>> >>> > >>>>>>> >>> > start) >>>>>>> >>> > >>>>>>> >>> > if ! [[ $(vgs -o attr --noheadings $OCF_RESKEY_vg_name) >>>>>>> =~ >>>>>>> >>> > .....c >>>>>>> >>> > ]]; then >>>>>>> >>> > >>>>>>> >>> > ha_lvm_proper_setup_check || exit 1 >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > If the vg is not taged as cluster than the ha_lvm is looking >>>>>>> for >>>>>>> >>> > volume_list >>>>>>> >>> > in lvm.conf. >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > I am confused- Does the VG should taged as cluster ?? ( BTW - >>>>>>> the old >>>>>>> >>> > fashion HA-LVM is worked with no problems ) >>>>>>> >>> > >>>>>>> >>> > redhat instructions : >>>>>>> >>> > >>>>>>> >>> > To set up HA LVM Failover (using the preferred CLVM variant), >>>>>>> perform >>>>>>> >>> > the >>>>>>> >>> > following steps: >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > 1. Ensure that the parameter locking_type in the global section >>>>>>> >>> > of /etc/lvm/lvm.conf is set to the value '3', that all the >>>>>>> necessary >>>>>>> >>> > LVM >>>>>>> >>> > cluster packages are installed, and the necessary daemons are >>>>>>> started >>>>>>> >>> > (like >>>>>>> >>> > 'clvmd' and the cluster mirror log daemon - if necessary). >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > 2. Create the logical volume and filesystem using standard LVM2 >>>>>>> and >>>>>>> >>> > file >>>>>>> >>> > system commands. For example: >>>>>>> >>> > >>>>>>> >>> > # pvcreate /dev/sd[cde]1 >>>>>>> >>> > >>>>>>> >>> > # vgcreate <volume group name> /dev/sd[cde]1 >>>>>>> >>> > >>>>>>> >>> > # lvcreate -L 10G -n <logical volume name> <volume group name> >>>>>>> >>> > >>>>>>> >>> > # mkfs.ext3 /dev/<volume group name>/<logical volume name> >>>>>>> >>> > >>>>>>> >>> > # lvchange -an <volume group name>/<logical volume name> >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > 3. Edit /etc/cluster/cluster.conf to include the newly created >>>>>>> logical >>>>>>> >>> > volume as a resource in one of your services. Alternatively, >>>>>>> >>> > configuration >>>>>>> >>> > tools such as Conga or system-config-cluster may be used to >>>>>>> create >>>>>>> >>> > these >>>>>>> >>> > entries. Below is a sample resource manager section >>>>>>> >>> > from /etc/cluster/cluster.conf: >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > <rm> <failoverdomains> <failoverdomain name="FD" >>>>>>> >>> > ordered="1" >>>>>>> >>> > restricted="0"> <failoverdomainnode name="neo-01" >>>>>>> >>> > priority="1"/> >>>>>>> >>> > <failoverdomainnode name="neo-02" priority="2"/> >>>>>>> >>> > </failoverdomain> </failoverdomains> <resources> >>>>>>> <lvm >>>>>>> >>> > name="lvm" vg_name="shared_vg" lv_name="ha-lv"/> <fs >>>>>>> name="FS" >>>>>>> >>> > device="/dev/shared_vg/ha-lv" force_fsck="0" force_unmount="1" >>>>>>> >>> > fsid="64050" >>>>>>> >>> > fstype="ext3" mountpoint="/mnt" options="" self_fence="0"/> >>>>>>> >>> > </resources> >>>>>>> >>> > <service autostart="1" domain="FD" name="serv" >>>>>>> recovery="relocate"> >>>>>>> >>> > <lvm ref="lvm"/> <fs ref="FS"/> </service> >>>>>>> </rm> >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > Regards >>>>>>> >>> > >>>>>>> >>> > Shalom. >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > >>>>>>> >>> > -- >>>>>>> >>> > Linux-cluster mailing list >>>>>>> >>> > [email protected] >>>>>>> >>> > https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>> > >>>>>>> >>> >>>>>>> >>> -- >>>>>>> >>> Linux-cluster mailing list >>>>>>> >>> [email protected] >>>>>>> >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> > >>>>>>> > >>>>>>> > -- >>>>>>> > Linux-cluster mailing list >>>>>>> > [email protected] >>>>>>> > https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> > >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> [email protected] >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> [email protected] >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> [email protected] >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> [email protected] >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >> >> -- >> Linux-cluster mailing list >> [email protected] >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > -- > Linux-cluster mailing list > [email protected] > https://www.redhat.com/mailman/listinfo/linux-cluster >
-- Linux-cluster mailing list [email protected] https://www.redhat.com/mailman/listinfo/linux-cluster
