Hi.
Dec 29 13:47:16 s1 LVM(vg1)[1601]: WARNING: LVM Volume cluvg1 is not available (stopped) Dec 29 13:47:16 s1 crmd[1515]: notice: process_lrm_event: Operation vg1_monitor_0: not running (node= s1, call=23, rc=7, cib-update=40, confirmed=true) Dec 29 13:47:16 s1 crmd[1515]: notice: te_rsc_command: Initiating action 9: monitor fs1_monitor_0 on s1 (local) Dec 29 13:47:16 s1 crmd[1515]: notice: te_rsc_command: Initiating action 16: monitor vg1_monitor_0 on s2 Dec 29 13:47:16 s1 Filesystem(fs1)[1618]: WARNING: Couldn't find device [/dev/mapper/cluvg1-clulv1]. Ex pected /dev/??? to exist from the LVM agent, it checked if the volume is already available.. and will raise the above error if not. But, I don't see that it tries to activate it before raising the VG. Perhaps, it assumes that the VG is already activated... so, I'm not sure who should be activating it (should it be LVM?). if [ $rc -ne 0 ]; then ocf_log $loglevel "LVM Volume $1 is not available (stopped)" rc=$OCF_NOT_RUNNING else case $(get_vg_mode) in 1) # exclusive with tagging. # If vg is running, make sure the correct tag is present. Otherwise we # can not guarantee exclusive activation. if ! check_tags; then ocf_exit_reason "WARNING: $OCF_RESKEY_volgrpname is active without the cluster tag, \"$OUR_TAG\"" On Mon, Dec 29, 2014 at 3:36 PM, emmanuel segura <emi2f...@gmail.com> wrote: > logs? > > 2014-12-29 6:54 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>: > > Hi, > > > > just want to ask regarding the LVM resource agent on pacemaker/corosync. > > > > I setup 2 nodes cluster (opensuse13.2 -- my config below). The cluster > > works as expected, like doing a manual failover (via crm resource move), > > and automatic failover (by rebooting the active node for instance). But, > if > > i try to just "shutoff" the active node (it's a VM, so I can do a > > poweroff). The resources won't be able to failover to the passive node. > > when I did an investigation, it's due to an LVM resource not starting > > (specifically, the VG). I found out that the LVM resource won't try to > > activate the volume group in the passive node. Is this an expected > > behaviour? > > > > what I really expect is that, in the event that the active node be > shutoff > > (by a power outage for instance), all resources should be failover > > automatically to the passive. LVM should re-activate the VG. > > > > > > here's my config. > > > > node 1: s1 > > node 2: s2 > > primitive cluIP IPaddr2 \ > > params ip=192.168.13.200 cidr_netmask=32 \ > > op monitor interval=30s > > primitive clvm ocf:lvm2:clvmd \ > > params daemon_timeout=30 \ > > op monitor timeout=90 interval=30 > > primitive dlm ocf:pacemaker:controld \ > > op monitor interval=60s timeout=90s on-fail=ignore \ > > op start interval=0 timeout=90 > > primitive fs1 Filesystem \ > > params device="/dev/mapper/cluvg1-clulv1" directory="/data" fstype=btrfs > > primitive mariadb mysql \ > > params config="/etc/my.cnf" > > primitive sbd stonith:external/sbd \ > > op monitor interval=15s timeout=60s > > primitive vg1 LVM \ > > params volgrpname=cluvg1 exclusive=yes \ > > op start timeout=10s interval=0 \ > > op stop interval=0 timeout=10 \ > > op monitor interval=10 timeout=30 on-fail=restart depth=0 > > group base-group dlm clvm > > group rgroup cluIP vg1 fs1 mariadb \ > > meta target-role=Started > > clone base-clone base-group \ > > meta interleave=true target-role=Started > > property cib-bootstrap-options: \ > > dc-version=1.1.12-1.1.12.git20140904.266d5c2 \ > > cluster-infrastructure=corosync \ > > no-quorum-policy=ignore \ > > last-lrm-refresh=1419514875 \ > > cluster-name=xxx \ > > stonith-enabled=true > > rsc_defaults rsc-options: \ > > resource-stickiness=100 > > > > -- > >>>> import this > > _______________________________________________ > > Linux-HA mailing list > > Linux-HA@lists.linux-ha.org > > http://lists.linux-ha.org/mailman/listinfo/linux-ha > > See also: http://linux-ha.org/ReportingProblems > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > _______________________________________________ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems > -- >>> import this _______________________________________________ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems