hi,

uploaded it here.

http://susepaste.org/45413433

thanks.

On Mon, Dec 29, 2014 at 5:09 PM, Marlon Guao <marlon.g...@gmail.com> wrote:

> Ok, i attached the log file of one of the nodes.
>
> On Mon, Dec 29, 2014 at 4:42 PM, emmanuel segura <emi2f...@gmail.com>
> wrote:
>
>> please use pastebin and show your whole logs
>>
>> 2014-12-29 9:06 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
>> > by the way.. just to note that.. for a normal testing (manual failover,
>> > rebooting the active node)... the cluster is working fine. I only
>> encounter
>> > this error if I try to poweroff/shutoff the active node.
>> >
>> > On Mon, Dec 29, 2014 at 4:05 PM, Marlon Guao <marlon.g...@gmail.com>
>> wrote:
>> >
>> >> Hi.
>> >>
>> >>
>> >> Dec 29 13:47:16 s1 LVM(vg1)[1601]: WARNING: LVM Volume cluvg1 is not
>> >> available (stopped)
>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice: process_lrm_event: Operation
>> >> vg1_monitor_0: not running (node=
>> >> s1, call=23, rc=7, cib-update=40, confirmed=true)
>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice: te_rsc_command: Initiating
>> action
>> >> 9: monitor fs1_monitor_0 on
>> >> s1 (local)
>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice: te_rsc_command: Initiating
>> action
>> >> 16: monitor vg1_monitor_0 on
>> >>  s2
>> >> Dec 29 13:47:16 s1 Filesystem(fs1)[1618]: WARNING: Couldn't find device
>> >> [/dev/mapper/cluvg1-clulv1]. Ex
>> >> pected /dev/??? to exist
>> >>
>> >>
>> >> from the LVM agent, it checked if the volume is already available.. and
>> >> will raise the above error if not. But, I don't see that it tries to
>> >> activate it before raising the VG. Perhaps, it assumes that the VG is
>> >> already activated... so, I'm not sure who should be activating it
>> (should
>> >> it be LVM?).
>> >>
>> >>
>> >>  if [ $rc -ne 0 ]; then
>> >>                 ocf_log $loglevel "LVM Volume $1 is not available
>> >> (stopped)"
>> >>                 rc=$OCF_NOT_RUNNING
>> >>         else
>> >>                 case $(get_vg_mode) in
>> >>                 1) # exclusive with tagging.
>> >>                         # If vg is running, make sure the correct tag
>> is
>> >> present. Otherwise we
>> >>                         # can not guarantee exclusive activation.
>> >>                         if ! check_tags; then
>> >>                                 ocf_exit_reason "WARNING:
>> >> $OCF_RESKEY_volgrpname is active without the cluster tag, \"$OUR_TAG\""
>> >>
>> >> On Mon, Dec 29, 2014 at 3:36 PM, emmanuel segura <emi2f...@gmail.com>
>> >> wrote:
>> >>
>> >>> logs?
>> >>>
>> >>> 2014-12-29 6:54 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
>> >>> > Hi,
>> >>> >
>> >>> > just want to ask regarding the LVM resource agent on
>> pacemaker/corosync.
>> >>> >
>> >>> > I setup 2 nodes cluster (opensuse13.2 -- my config below). The
>> cluster
>> >>> > works as expected, like doing a manual failover (via crm resource
>> move),
>> >>> > and automatic failover (by rebooting the active node for instance).
>> >>> But, if
>> >>> > i try to just "shutoff" the active node (it's a VM, so I can do a
>> >>> > poweroff). The resources won't be able to failover to the passive
>> node.
>> >>> > when I did an investigation, it's due to an LVM resource not
>> starting
>> >>> > (specifically, the VG). I found out that the LVM resource won't try
>> to
>> >>> > activate the volume group in the passive node. Is this an expected
>> >>> > behaviour?
>> >>> >
>> >>> > what I really expect is that, in the event that the active node be
>> >>> shutoff
>> >>> > (by a power outage for instance), all resources should be failover
>> >>> > automatically to the passive. LVM should re-activate the VG.
>> >>> >
>> >>> >
>> >>> > here's my config.
>> >>> >
>> >>> > node 1: s1
>> >>> > node 2: s2
>> >>> > primitive cluIP IPaddr2 \
>> >>> > params ip=192.168.13.200 cidr_netmask=32 \
>> >>> > op monitor interval=30s
>> >>> > primitive clvm ocf:lvm2:clvmd \
>> >>> > params daemon_timeout=30 \
>> >>> > op monitor timeout=90 interval=30
>> >>> > primitive dlm ocf:pacemaker:controld \
>> >>> > op monitor interval=60s timeout=90s on-fail=ignore \
>> >>> > op start interval=0 timeout=90
>> >>> > primitive fs1 Filesystem \
>> >>> > params device="/dev/mapper/cluvg1-clulv1" directory="/data"
>> fstype=btrfs
>> >>> > primitive mariadb mysql \
>> >>> > params config="/etc/my.cnf"
>> >>> > primitive sbd stonith:external/sbd \
>> >>> > op monitor interval=15s timeout=60s
>> >>> > primitive vg1 LVM \
>> >>> > params volgrpname=cluvg1 exclusive=yes \
>> >>> > op start timeout=10s interval=0 \
>> >>> > op stop interval=0 timeout=10 \
>> >>> > op monitor interval=10 timeout=30 on-fail=restart depth=0
>> >>> > group base-group dlm clvm
>> >>> > group rgroup cluIP vg1 fs1 mariadb \
>> >>> > meta target-role=Started
>> >>> > clone base-clone base-group \
>> >>> > meta interleave=true target-role=Started
>> >>> > property cib-bootstrap-options: \
>> >>> > dc-version=1.1.12-1.1.12.git20140904.266d5c2 \
>> >>> > cluster-infrastructure=corosync \
>> >>> > no-quorum-policy=ignore \
>> >>> > last-lrm-refresh=1419514875 \
>> >>> > cluster-name=xxx \
>> >>> > stonith-enabled=true
>> >>> > rsc_defaults rsc-options: \
>> >>> > resource-stickiness=100
>> >>> >
>> >>> > --
>> >>> >>>> import this
>> >>> > _______________________________________________
>> >>> > Linux-HA mailing list
>> >>> > Linux-HA@lists.linux-ha.org
>> >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> >>> > See also: http://linux-ha.org/ReportingProblems
>> >>>
>> >>>
>> >>>
>> >>> --
>> >>> esta es mi vida e me la vivo hasta que dios quiera
>> >>> _______________________________________________
>> >>> Linux-HA mailing list
>> >>> Linux-HA@lists.linux-ha.org
>> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> >>> See also: http://linux-ha.org/ReportingProblems
>> >>>
>> >>
>> >>
>> >>
>> >> --
>> >> >>> import this
>> >>
>> >
>> >
>> >
>> > --
>> >>>> import this
>> > _______________________________________________
>> > Linux-HA mailing list
>> > Linux-HA@lists.linux-ha.org
>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> > See also: http://linux-ha.org/ReportingProblems
>>
>>
>>
>> --
>> esta es mi vida e me la vivo hasta que dios quiera
>> _______________________________________________
>> Linux-HA mailing list
>> Linux-HA@lists.linux-ha.org
>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> See also: http://linux-ha.org/ReportingProblems
>>
>
>
>
> --
> >>> import this
>



-- 
>>> import this
_______________________________________________
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to