Dec 27 15:38:00 s1 cib[1514]:    error: crm_xml_err: XML Error:
Permission deniedPermission deniedI/O warning : failed to load
external entity "/var/lib/pacemaker/cib/cib.xml"
Dec 27 15:38:00 s1 cib[1514]:    error: write_cib_contents: Cannot
link /var/lib/pacemaker/cib/cib.xml to
/var/lib/pacemaker/cib/cib-0.raw: Operation not permitted (1)

2014-12-29 10:33 GMT+01:00 emmanuel segura <emi2f...@gmail.com>:
> Hi,
>
> You have  a problem with the cluster stonithd:"error: crm_abort:
> crm_glib_handler: Forked child 6186 to record non-fatal assert at
> logging.c:73 "
>
> Try to post your cluster version(packages), maybe someone can tell you
> if this is a known bug or new.
>
>
>
> 2014-12-29 10:29 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
>> ok, sorry for that.. please use this instead.
>>
>> http://pastebin.centos.org/14771/
>>
>> thanks.
>>
>> On Mon, Dec 29, 2014 at 5:25 PM, emmanuel segura <emi2f...@gmail.com> wrote:
>>
>>> Sorry,
>>>
>>> But your paste is empty.
>>>
>>> 2014-12-29 10:19 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
>>> > hi,
>>> >
>>> > uploaded it here.
>>> >
>>> > http://susepaste.org/45413433
>>> >
>>> > thanks.
>>> >
>>> > On Mon, Dec 29, 2014 at 5:09 PM, Marlon Guao <marlon.g...@gmail.com>
>>> wrote:
>>> >
>>> >> Ok, i attached the log file of one of the nodes.
>>> >>
>>> >> On Mon, Dec 29, 2014 at 4:42 PM, emmanuel segura <emi2f...@gmail.com>
>>> >> wrote:
>>> >>
>>> >>> please use pastebin and show your whole logs
>>> >>>
>>> >>> 2014-12-29 9:06 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
>>> >>> > by the way.. just to note that.. for a normal testing (manual
>>> failover,
>>> >>> > rebooting the active node)... the cluster is working fine. I only
>>> >>> encounter
>>> >>> > this error if I try to poweroff/shutoff the active node.
>>> >>> >
>>> >>> > On Mon, Dec 29, 2014 at 4:05 PM, Marlon Guao <marlon.g...@gmail.com>
>>> >>> wrote:
>>> >>> >
>>> >>> >> Hi.
>>> >>> >>
>>> >>> >>
>>> >>> >> Dec 29 13:47:16 s1 LVM(vg1)[1601]: WARNING: LVM Volume cluvg1 is not
>>> >>> >> available (stopped)
>>> >>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice: process_lrm_event:
>>> Operation
>>> >>> >> vg1_monitor_0: not running (node=
>>> >>> >> s1, call=23, rc=7, cib-update=40, confirmed=true)
>>> >>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice: te_rsc_command: Initiating
>>> >>> action
>>> >>> >> 9: monitor fs1_monitor_0 on
>>> >>> >> s1 (local)
>>> >>> >> Dec 29 13:47:16 s1 crmd[1515]:   notice: te_rsc_command: Initiating
>>> >>> action
>>> >>> >> 16: monitor vg1_monitor_0 on
>>> >>> >>  s2
>>> >>> >> Dec 29 13:47:16 s1 Filesystem(fs1)[1618]: WARNING: Couldn't find
>>> device
>>> >>> >> [/dev/mapper/cluvg1-clulv1]. Ex
>>> >>> >> pected /dev/??? to exist
>>> >>> >>
>>> >>> >>
>>> >>> >> from the LVM agent, it checked if the volume is already available..
>>> and
>>> >>> >> will raise the above error if not. But, I don't see that it tries to
>>> >>> >> activate it before raising the VG. Perhaps, it assumes that the VG
>>> is
>>> >>> >> already activated... so, I'm not sure who should be activating it
>>> >>> (should
>>> >>> >> it be LVM?).
>>> >>> >>
>>> >>> >>
>>> >>> >>  if [ $rc -ne 0 ]; then
>>> >>> >>                 ocf_log $loglevel "LVM Volume $1 is not available
>>> >>> >> (stopped)"
>>> >>> >>                 rc=$OCF_NOT_RUNNING
>>> >>> >>         else
>>> >>> >>                 case $(get_vg_mode) in
>>> >>> >>                 1) # exclusive with tagging.
>>> >>> >>                         # If vg is running, make sure the correct
>>> tag
>>> >>> is
>>> >>> >> present. Otherwise we
>>> >>> >>                         # can not guarantee exclusive activation.
>>> >>> >>                         if ! check_tags; then
>>> >>> >>                                 ocf_exit_reason "WARNING:
>>> >>> >> $OCF_RESKEY_volgrpname is active without the cluster tag,
>>> \"$OUR_TAG\""
>>> >>> >>
>>> >>> >> On Mon, Dec 29, 2014 at 3:36 PM, emmanuel segura <
>>> emi2f...@gmail.com>
>>> >>> >> wrote:
>>> >>> >>
>>> >>> >>> logs?
>>> >>> >>>
>>> >>> >>> 2014-12-29 6:54 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>:
>>> >>> >>> > Hi,
>>> >>> >>> >
>>> >>> >>> > just want to ask regarding the LVM resource agent on
>>> >>> pacemaker/corosync.
>>> >>> >>> >
>>> >>> >>> > I setup 2 nodes cluster (opensuse13.2 -- my config below). The
>>> >>> cluster
>>> >>> >>> > works as expected, like doing a manual failover (via crm resource
>>> >>> move),
>>> >>> >>> > and automatic failover (by rebooting the active node for
>>> instance).
>>> >>> >>> But, if
>>> >>> >>> > i try to just "shutoff" the active node (it's a VM, so I can do a
>>> >>> >>> > poweroff). The resources won't be able to failover to the passive
>>> >>> node.
>>> >>> >>> > when I did an investigation, it's due to an LVM resource not
>>> >>> starting
>>> >>> >>> > (specifically, the VG). I found out that the LVM resource won't
>>> try
>>> >>> to
>>> >>> >>> > activate the volume group in the passive node. Is this an
>>> expected
>>> >>> >>> > behaviour?
>>> >>> >>> >
>>> >>> >>> > what I really expect is that, in the event that the active node
>>> be
>>> >>> >>> shutoff
>>> >>> >>> > (by a power outage for instance), all resources should be
>>> failover
>>> >>> >>> > automatically to the passive. LVM should re-activate the VG.
>>> >>> >>> >
>>> >>> >>> >
>>> >>> >>> > here's my config.
>>> >>> >>> >
>>> >>> >>> > node 1: s1
>>> >>> >>> > node 2: s2
>>> >>> >>> > primitive cluIP IPaddr2 \
>>> >>> >>> > params ip=192.168.13.200 cidr_netmask=32 \
>>> >>> >>> > op monitor interval=30s
>>> >>> >>> > primitive clvm ocf:lvm2:clvmd \
>>> >>> >>> > params daemon_timeout=30 \
>>> >>> >>> > op monitor timeout=90 interval=30
>>> >>> >>> > primitive dlm ocf:pacemaker:controld \
>>> >>> >>> > op monitor interval=60s timeout=90s on-fail=ignore \
>>> >>> >>> > op start interval=0 timeout=90
>>> >>> >>> > primitive fs1 Filesystem \
>>> >>> >>> > params device="/dev/mapper/cluvg1-clulv1" directory="/data"
>>> >>> fstype=btrfs
>>> >>> >>> > primitive mariadb mysql \
>>> >>> >>> > params config="/etc/my.cnf"
>>> >>> >>> > primitive sbd stonith:external/sbd \
>>> >>> >>> > op monitor interval=15s timeout=60s
>>> >>> >>> > primitive vg1 LVM \
>>> >>> >>> > params volgrpname=cluvg1 exclusive=yes \
>>> >>> >>> > op start timeout=10s interval=0 \
>>> >>> >>> > op stop interval=0 timeout=10 \
>>> >>> >>> > op monitor interval=10 timeout=30 on-fail=restart depth=0
>>> >>> >>> > group base-group dlm clvm
>>> >>> >>> > group rgroup cluIP vg1 fs1 mariadb \
>>> >>> >>> > meta target-role=Started
>>> >>> >>> > clone base-clone base-group \
>>> >>> >>> > meta interleave=true target-role=Started
>>> >>> >>> > property cib-bootstrap-options: \
>>> >>> >>> > dc-version=1.1.12-1.1.12.git20140904.266d5c2 \
>>> >>> >>> > cluster-infrastructure=corosync \
>>> >>> >>> > no-quorum-policy=ignore \
>>> >>> >>> > last-lrm-refresh=1419514875 \
>>> >>> >>> > cluster-name=xxx \
>>> >>> >>> > stonith-enabled=true
>>> >>> >>> > rsc_defaults rsc-options: \
>>> >>> >>> > resource-stickiness=100
>>> >>> >>> >
>>> >>> >>> > --
>>> >>> >>> >>>> import this
>>> >>> >>> > _______________________________________________
>>> >>> >>> > Linux-HA mailing list
>>> >>> >>> > Linux-HA@lists.linux-ha.org
>>> >>> >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >>> >>> > See also: http://linux-ha.org/ReportingProblems
>>> >>> >>>
>>> >>> >>>
>>> >>> >>>
>>> >>> >>> --
>>> >>> >>> esta es mi vida e me la vivo hasta que dios quiera
>>> >>> >>> _______________________________________________
>>> >>> >>> Linux-HA mailing list
>>> >>> >>> Linux-HA@lists.linux-ha.org
>>> >>> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >>> >>> See also: http://linux-ha.org/ReportingProblems
>>> >>> >>>
>>> >>> >>
>>> >>> >>
>>> >>> >>
>>> >>> >> --
>>> >>> >> >>> import this
>>> >>> >>
>>> >>> >
>>> >>> >
>>> >>> >
>>> >>> > --
>>> >>> >>>> import this
>>> >>> > _______________________________________________
>>> >>> > Linux-HA mailing list
>>> >>> > Linux-HA@lists.linux-ha.org
>>> >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >>> > See also: http://linux-ha.org/ReportingProblems
>>> >>>
>>> >>>
>>> >>>
>>> >>> --
>>> >>> esta es mi vida e me la vivo hasta que dios quiera
>>> >>> _______________________________________________
>>> >>> Linux-HA mailing list
>>> >>> Linux-HA@lists.linux-ha.org
>>> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> >>> See also: http://linux-ha.org/ReportingProblems
>>> >>>
>>> >>
>>> >>
>>> >>
>>> >> --
>>> >> >>> import this
>>> >>
>>> >
>>> >
>>> >
>>> > --
>>> >>>> import this
>>> > _______________________________________________
>>> > Linux-HA mailing list
>>> > Linux-HA@lists.linux-ha.org
>>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> > See also: http://linux-ha.org/ReportingProblems
>>>
>>>
>>>
>>> --
>>> esta es mi vida e me la vivo hasta que dios quiera
>>> _______________________________________________
>>> Linux-HA mailing list
>>> Linux-HA@lists.linux-ha.org
>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> See also: http://linux-ha.org/ReportingProblems
>>>
>>
>>
>>
>> --
>>>>> import this
>> _______________________________________________
>> Linux-HA mailing list
>> Linux-HA@lists.linux-ha.org
>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> See also: http://linux-ha.org/ReportingProblems
>
>
>
> --
> esta es mi vida e me la vivo hasta que dios quiera



-- 
esta es mi vida e me la vivo hasta que dios quiera
_______________________________________________
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to