please use pastebin and show your whole logs

2014-12-29 9:06 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
> by the way.. just to note that.. for a normal testing (manual failover,
> rebooting the active node)... the cluster is working fine. I only encounter
> this error if I try to poweroff/shutoff the active node.
>
> On Mon, Dec 29, 2014 at 4:05 PM, Marlon Guao <marlon.g...@gmail.com> wrote:
>
>> Hi.
>>
>>
>> Dec 29 13:47:16 s1 LVM(vg1)[1601]: WARNING: LVM Volume cluvg1 is not
>> available (stopped)
>> Dec 29 13:47:16 s1 crmd[1515]:   notice: process_lrm_event: Operation
>> vg1_monitor_0: not running (node=
>> s1, call=23, rc=7, cib-update=40, confirmed=true)
>> Dec 29 13:47:16 s1 crmd[1515]:   notice: te_rsc_command: Initiating action
>> 9: monitor fs1_monitor_0 on
>> s1 (local)
>> Dec 29 13:47:16 s1 crmd[1515]:   notice: te_rsc_command: Initiating action
>> 16: monitor vg1_monitor_0 on
>>  s2
>> Dec 29 13:47:16 s1 Filesystem(fs1)[1618]: WARNING: Couldn't find device
>> [/dev/mapper/cluvg1-clulv1]. Ex
>> pected /dev/??? to exist
>>
>>
>> from the LVM agent, it checked if the volume is already available.. and
>> will raise the above error if not. But, I don't see that it tries to
>> activate it before raising the VG. Perhaps, it assumes that the VG is
>> already activated... so, I'm not sure who should be activating it (should
>> it be LVM?).
>>
>>
>>  if [ $rc -ne 0 ]; then
>>                 ocf_log $loglevel "LVM Volume $1 is not available
>> (stopped)"
>>                 rc=$OCF_NOT_RUNNING
>>         else
>>                 case $(get_vg_mode) in
>>                 1) # exclusive with tagging.
>>                         # If vg is running, make sure the correct tag is
>> present. Otherwise we
>>                         # can not guarantee exclusive activation.
>>                         if ! check_tags; then
>>                                 ocf_exit_reason "WARNING:
>> $OCF_RESKEY_volgrpname is active without the cluster tag, \"$OUR_TAG\""
>>
>> On Mon, Dec 29, 2014 at 3:36 PM, emmanuel segura <emi2f...@gmail.com>
>> wrote:
>>
>>> logs?
>>>
>>> 2014-12-29 6:54 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
>>> > Hi,
>>> >
>>> > just want to ask regarding the LVM resource agent on pacemaker/corosync.
>>> >
>>> > I setup 2 nodes cluster (opensuse13.2 -- my config below). The cluster
>>> > works as expected, like doing a manual failover (via crm resource move),
>>> > and automatic failover (by rebooting the active node for instance).
>>> But, if
>>> > i try to just "shutoff" the active node (it's a VM, so I can do a
>>> > poweroff). The resources won't be able to failover to the passive node.
>>> > when I did an investigation, it's due to an LVM resource not starting
>>> > (specifically, the VG). I found out that the LVM resource won't try to
>>> > activate the volume group in the passive node. Is this an expected
>>> > behaviour?
>>> >
>>> > what I really expect is that, in the event that the active node be
>>> shutoff
>>> > (by a power outage for instance), all resources should be failover
>>> > automatically to the passive. LVM should re-activate the VG.
>>> >
>>> >
>>> > here's my config.
>>> >
>>> > node 1: s1
>>> > node 2: s2
>>> > primitive cluIP IPaddr2 \
>>> > params ip=192.168.13.200 cidr_netmask=32 \
>>> > op monitor interval=30s
>>> > primitive clvm ocf:lvm2:clvmd \
>>> > params daemon_timeout=30 \
>>> > op monitor timeout=90 interval=30
>>> > primitive dlm ocf:pacemaker:controld \
>>> > op monitor interval=60s timeout=90s on-fail=ignore \
>>> > op start interval=0 timeout=90
>>> > primitive fs1 Filesystem \
>>> > params device="/dev/mapper/cluvg1-clulv1" directory="/data" fstype=btrfs
>>> > primitive mariadb mysql \
>>> > params config="/etc/my.cnf"
>>> > primitive sbd stonith:external/sbd \
>>> > op monitor interval=15s timeout=60s
>>> > primitive vg1 LVM \
>>> > params volgrpname=cluvg1 exclusive=yes \
>>> > op start timeout=10s interval=0 \
>>> > op stop interval=0 timeout=10 \
>>> > op monitor interval=10 timeout=30 on-fail=restart depth=0
>>> > group base-group dlm clvm
>>> > group rgroup cluIP vg1 fs1 mariadb \
>>> > meta target-role=Started
>>> > clone base-clone base-group \
>>> > meta interleave=true target-role=Started
>>> > property cib-bootstrap-options: \
>>> > dc-version=1.1.12-1.1.12.git20140904.266d5c2 \
>>> > cluster-infrastructure=corosync \
>>> > no-quorum-policy=ignore \
>>> > last-lrm-refresh=1419514875 \
>>> > cluster-name=xxx \
>>> > stonith-enabled=true
>>> > rsc_defaults rsc-options: \
>>> > resource-stickiness=100
>>> >
>>> > --
>>> >>>> import this
>>> > _______________________________________________
>>> > Linux-HA mailing list
>>> > Linux-HA@lists.linux-ha.org
>>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> > See also: http://linux-ha.org/ReportingProblems
>>>
>>>
>>>
>>> --
>>> esta es mi vida e me la vivo hasta que dios quiera
>>> _______________________________________________
>>> Linux-HA mailing list
>>> Linux-HA@lists.linux-ha.org
>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> See also: http://linux-ha.org/ReportingProblems
>>>
>>
>>
>>
>> --
>> >>> import this
>>
>
>
>
> --
>>>> import this
> _______________________________________________
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems



-- 
esta es mi vida e me la vivo hasta que dios quiera
_______________________________________________
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to