please use pastebin and show your whole logs 2014-12-29 9:06 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>: > by the way.. just to note that.. for a normal testing (manual failover, > rebooting the active node)... the cluster is working fine. I only encounter > this error if I try to poweroff/shutoff the active node. > > On Mon, Dec 29, 2014 at 4:05 PM, Marlon Guao <marlon.g...@gmail.com> wrote: > >> Hi. >> >> >> Dec 29 13:47:16 s1 LVM(vg1)[1601]: WARNING: LVM Volume cluvg1 is not >> available (stopped) >> Dec 29 13:47:16 s1 crmd[1515]: notice: process_lrm_event: Operation >> vg1_monitor_0: not running (node= >> s1, call=23, rc=7, cib-update=40, confirmed=true) >> Dec 29 13:47:16 s1 crmd[1515]: notice: te_rsc_command: Initiating action >> 9: monitor fs1_monitor_0 on >> s1 (local) >> Dec 29 13:47:16 s1 crmd[1515]: notice: te_rsc_command: Initiating action >> 16: monitor vg1_monitor_0 on >> s2 >> Dec 29 13:47:16 s1 Filesystem(fs1)[1618]: WARNING: Couldn't find device >> [/dev/mapper/cluvg1-clulv1]. Ex >> pected /dev/??? to exist >> >> >> from the LVM agent, it checked if the volume is already available.. and >> will raise the above error if not. But, I don't see that it tries to >> activate it before raising the VG. Perhaps, it assumes that the VG is >> already activated... so, I'm not sure who should be activating it (should >> it be LVM?). >> >> >> if [ $rc -ne 0 ]; then >> ocf_log $loglevel "LVM Volume $1 is not available >> (stopped)" >> rc=$OCF_NOT_RUNNING >> else >> case $(get_vg_mode) in >> 1) # exclusive with tagging. >> # If vg is running, make sure the correct tag is >> present. Otherwise we >> # can not guarantee exclusive activation. >> if ! check_tags; then >> ocf_exit_reason "WARNING: >> $OCF_RESKEY_volgrpname is active without the cluster tag, \"$OUR_TAG\"" >> >> On Mon, Dec 29, 2014 at 3:36 PM, emmanuel segura <emi2f...@gmail.com> >> wrote: >> >>> logs? >>> >>> 2014-12-29 6:54 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>: >>> > Hi, >>> > >>> > just want to ask regarding the LVM resource agent on pacemaker/corosync. >>> > >>> > I setup 2 nodes cluster (opensuse13.2 -- my config below). The cluster >>> > works as expected, like doing a manual failover (via crm resource move), >>> > and automatic failover (by rebooting the active node for instance). >>> But, if >>> > i try to just "shutoff" the active node (it's a VM, so I can do a >>> > poweroff). The resources won't be able to failover to the passive node. >>> > when I did an investigation, it's due to an LVM resource not starting >>> > (specifically, the VG). I found out that the LVM resource won't try to >>> > activate the volume group in the passive node. Is this an expected >>> > behaviour? >>> > >>> > what I really expect is that, in the event that the active node be >>> shutoff >>> > (by a power outage for instance), all resources should be failover >>> > automatically to the passive. LVM should re-activate the VG. >>> > >>> > >>> > here's my config. >>> > >>> > node 1: s1 >>> > node 2: s2 >>> > primitive cluIP IPaddr2 \ >>> > params ip=192.168.13.200 cidr_netmask=32 \ >>> > op monitor interval=30s >>> > primitive clvm ocf:lvm2:clvmd \ >>> > params daemon_timeout=30 \ >>> > op monitor timeout=90 interval=30 >>> > primitive dlm ocf:pacemaker:controld \ >>> > op monitor interval=60s timeout=90s on-fail=ignore \ >>> > op start interval=0 timeout=90 >>> > primitive fs1 Filesystem \ >>> > params device="/dev/mapper/cluvg1-clulv1" directory="/data" fstype=btrfs >>> > primitive mariadb mysql \ >>> > params config="/etc/my.cnf" >>> > primitive sbd stonith:external/sbd \ >>> > op monitor interval=15s timeout=60s >>> > primitive vg1 LVM \ >>> > params volgrpname=cluvg1 exclusive=yes \ >>> > op start timeout=10s interval=0 \ >>> > op stop interval=0 timeout=10 \ >>> > op monitor interval=10 timeout=30 on-fail=restart depth=0 >>> > group base-group dlm clvm >>> > group rgroup cluIP vg1 fs1 mariadb \ >>> > meta target-role=Started >>> > clone base-clone base-group \ >>> > meta interleave=true target-role=Started >>> > property cib-bootstrap-options: \ >>> > dc-version=1.1.12-1.1.12.git20140904.266d5c2 \ >>> > cluster-infrastructure=corosync \ >>> > no-quorum-policy=ignore \ >>> > last-lrm-refresh=1419514875 \ >>> > cluster-name=xxx \ >>> > stonith-enabled=true >>> > rsc_defaults rsc-options: \ >>> > resource-stickiness=100 >>> > >>> > -- >>> >>>> import this >>> > _______________________________________________ >>> > Linux-HA mailing list >>> > Linux-HA@lists.linux-ha.org >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha >>> > See also: http://linux-ha.org/ReportingProblems >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> _______________________________________________ >>> Linux-HA mailing list >>> Linux-HA@lists.linux-ha.org >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >>> See also: http://linux-ha.org/ReportingProblems >>> >> >> >> >> -- >> >>> import this >> > > > > -- >>>> import this > _______________________________________________ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems
-- esta es mi vida e me la vivo hasta que dios quiera _______________________________________________ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems