Dec 27 15:38:00 s1 cib[1514]: error: crm_xml_err: XML Error: Permission deniedPermission deniedI/O warning : failed to load external entity "/var/lib/pacemaker/cib/cib.xml" Dec 27 15:38:00 s1 cib[1514]: error: write_cib_contents: Cannot link /var/lib/pacemaker/cib/cib.xml to /var/lib/pacemaker/cib/cib-0.raw: Operation not permitted (1)
2014-12-29 10:33 GMT+01:00 emmanuel segura <emi2f...@gmail.com>: > Hi, > > You have a problem with the cluster stonithd:"error: crm_abort: > crm_glib_handler: Forked child 6186 to record non-fatal assert at > logging.c:73 " > > Try to post your cluster version(packages), maybe someone can tell you > if this is a known bug or new. > > > > 2014-12-29 10:29 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>: >> ok, sorry for that.. please use this instead. >> >> http://pastebin.centos.org/14771/ >> >> thanks. >> >> On Mon, Dec 29, 2014 at 5:25 PM, emmanuel segura <emi2f...@gmail.com> wrote: >> >>> Sorry, >>> >>> But your paste is empty. >>> >>> 2014-12-29 10:19 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>: >>> > hi, >>> > >>> > uploaded it here. >>> > >>> > http://susepaste.org/45413433 >>> > >>> > thanks. >>> > >>> > On Mon, Dec 29, 2014 at 5:09 PM, Marlon Guao <marlon.g...@gmail.com> >>> wrote: >>> > >>> >> Ok, i attached the log file of one of the nodes. >>> >> >>> >> On Mon, Dec 29, 2014 at 4:42 PM, emmanuel segura <emi2f...@gmail.com> >>> >> wrote: >>> >> >>> >>> please use pastebin and show your whole logs >>> >>> >>> >>> 2014-12-29 9:06 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>: >>> >>> > by the way.. just to note that.. for a normal testing (manual >>> failover, >>> >>> > rebooting the active node)... the cluster is working fine. I only >>> >>> encounter >>> >>> > this error if I try to poweroff/shutoff the active node. >>> >>> > >>> >>> > On Mon, Dec 29, 2014 at 4:05 PM, Marlon Guao <marlon.g...@gmail.com> >>> >>> wrote: >>> >>> > >>> >>> >> Hi. >>> >>> >> >>> >>> >> >>> >>> >> Dec 29 13:47:16 s1 LVM(vg1)[1601]: WARNING: LVM Volume cluvg1 is not >>> >>> >> available (stopped) >>> >>> >> Dec 29 13:47:16 s1 crmd[1515]: notice: process_lrm_event: >>> Operation >>> >>> >> vg1_monitor_0: not running (node= >>> >>> >> s1, call=23, rc=7, cib-update=40, confirmed=true) >>> >>> >> Dec 29 13:47:16 s1 crmd[1515]: notice: te_rsc_command: Initiating >>> >>> action >>> >>> >> 9: monitor fs1_monitor_0 on >>> >>> >> s1 (local) >>> >>> >> Dec 29 13:47:16 s1 crmd[1515]: notice: te_rsc_command: Initiating >>> >>> action >>> >>> >> 16: monitor vg1_monitor_0 on >>> >>> >> s2 >>> >>> >> Dec 29 13:47:16 s1 Filesystem(fs1)[1618]: WARNING: Couldn't find >>> device >>> >>> >> [/dev/mapper/cluvg1-clulv1]. Ex >>> >>> >> pected /dev/??? to exist >>> >>> >> >>> >>> >> >>> >>> >> from the LVM agent, it checked if the volume is already available.. >>> and >>> >>> >> will raise the above error if not. But, I don't see that it tries to >>> >>> >> activate it before raising the VG. Perhaps, it assumes that the VG >>> is >>> >>> >> already activated... so, I'm not sure who should be activating it >>> >>> (should >>> >>> >> it be LVM?). >>> >>> >> >>> >>> >> >>> >>> >> if [ $rc -ne 0 ]; then >>> >>> >> ocf_log $loglevel "LVM Volume $1 is not available >>> >>> >> (stopped)" >>> >>> >> rc=$OCF_NOT_RUNNING >>> >>> >> else >>> >>> >> case $(get_vg_mode) in >>> >>> >> 1) # exclusive with tagging. >>> >>> >> # If vg is running, make sure the correct >>> tag >>> >>> is >>> >>> >> present. Otherwise we >>> >>> >> # can not guarantee exclusive activation. >>> >>> >> if ! check_tags; then >>> >>> >> ocf_exit_reason "WARNING: >>> >>> >> $OCF_RESKEY_volgrpname is active without the cluster tag, >>> \"$OUR_TAG\"" >>> >>> >> >>> >>> >> On Mon, Dec 29, 2014 at 3:36 PM, emmanuel segura < >>> emi2f...@gmail.com> >>> >>> >> wrote: >>> >>> >> >>> >>> >>> logs? >>> >>> >>> >>> >>> >>> 2014-12-29 6:54 GMT+01:00 Marlon Guao <marlon.g...@gmail.com>: >>> >>> >>> > Hi, >>> >>> >>> > >>> >>> >>> > just want to ask regarding the LVM resource agent on >>> >>> pacemaker/corosync. >>> >>> >>> > >>> >>> >>> > I setup 2 nodes cluster (opensuse13.2 -- my config below). The >>> >>> cluster >>> >>> >>> > works as expected, like doing a manual failover (via crm resource >>> >>> move), >>> >>> >>> > and automatic failover (by rebooting the active node for >>> instance). >>> >>> >>> But, if >>> >>> >>> > i try to just "shutoff" the active node (it's a VM, so I can do a >>> >>> >>> > poweroff). The resources won't be able to failover to the passive >>> >>> node. >>> >>> >>> > when I did an investigation, it's due to an LVM resource not >>> >>> starting >>> >>> >>> > (specifically, the VG). I found out that the LVM resource won't >>> try >>> >>> to >>> >>> >>> > activate the volume group in the passive node. Is this an >>> expected >>> >>> >>> > behaviour? >>> >>> >>> > >>> >>> >>> > what I really expect is that, in the event that the active node >>> be >>> >>> >>> shutoff >>> >>> >>> > (by a power outage for instance), all resources should be >>> failover >>> >>> >>> > automatically to the passive. LVM should re-activate the VG. >>> >>> >>> > >>> >>> >>> > >>> >>> >>> > here's my config. >>> >>> >>> > >>> >>> >>> > node 1: s1 >>> >>> >>> > node 2: s2 >>> >>> >>> > primitive cluIP IPaddr2 \ >>> >>> >>> > params ip=192.168.13.200 cidr_netmask=32 \ >>> >>> >>> > op monitor interval=30s >>> >>> >>> > primitive clvm ocf:lvm2:clvmd \ >>> >>> >>> > params daemon_timeout=30 \ >>> >>> >>> > op monitor timeout=90 interval=30 >>> >>> >>> > primitive dlm ocf:pacemaker:controld \ >>> >>> >>> > op monitor interval=60s timeout=90s on-fail=ignore \ >>> >>> >>> > op start interval=0 timeout=90 >>> >>> >>> > primitive fs1 Filesystem \ >>> >>> >>> > params device="/dev/mapper/cluvg1-clulv1" directory="/data" >>> >>> fstype=btrfs >>> >>> >>> > primitive mariadb mysql \ >>> >>> >>> > params config="/etc/my.cnf" >>> >>> >>> > primitive sbd stonith:external/sbd \ >>> >>> >>> > op monitor interval=15s timeout=60s >>> >>> >>> > primitive vg1 LVM \ >>> >>> >>> > params volgrpname=cluvg1 exclusive=yes \ >>> >>> >>> > op start timeout=10s interval=0 \ >>> >>> >>> > op stop interval=0 timeout=10 \ >>> >>> >>> > op monitor interval=10 timeout=30 on-fail=restart depth=0 >>> >>> >>> > group base-group dlm clvm >>> >>> >>> > group rgroup cluIP vg1 fs1 mariadb \ >>> >>> >>> > meta target-role=Started >>> >>> >>> > clone base-clone base-group \ >>> >>> >>> > meta interleave=true target-role=Started >>> >>> >>> > property cib-bootstrap-options: \ >>> >>> >>> > dc-version=1.1.12-1.1.12.git20140904.266d5c2 \ >>> >>> >>> > cluster-infrastructure=corosync \ >>> >>> >>> > no-quorum-policy=ignore \ >>> >>> >>> > last-lrm-refresh=1419514875 \ >>> >>> >>> > cluster-name=xxx \ >>> >>> >>> > stonith-enabled=true >>> >>> >>> > rsc_defaults rsc-options: \ >>> >>> >>> > resource-stickiness=100 >>> >>> >>> > >>> >>> >>> > -- >>> >>> >>> >>>> import this >>> >>> >>> > _______________________________________________ >>> >>> >>> > Linux-HA mailing list >>> >>> >>> > Linux-HA@lists.linux-ha.org >>> >>> >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha >>> >>> >>> > See also: http://linux-ha.org/ReportingProblems >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> -- >>> >>> >>> esta es mi vida e me la vivo hasta que dios quiera >>> >>> >>> _______________________________________________ >>> >>> >>> Linux-HA mailing list >>> >>> >>> Linux-HA@lists.linux-ha.org >>> >>> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >>> >>> >>> See also: http://linux-ha.org/ReportingProblems >>> >>> >>> >>> >>> >> >>> >>> >> >>> >>> >> >>> >>> >> -- >>> >>> >> >>> import this >>> >>> >> >>> >>> > >>> >>> > >>> >>> > >>> >>> > -- >>> >>> >>>> import this >>> >>> > _______________________________________________ >>> >>> > Linux-HA mailing list >>> >>> > Linux-HA@lists.linux-ha.org >>> >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha >>> >>> > See also: http://linux-ha.org/ReportingProblems >>> >>> >>> >>> >>> >>> >>> >>> -- >>> >>> esta es mi vida e me la vivo hasta que dios quiera >>> >>> _______________________________________________ >>> >>> Linux-HA mailing list >>> >>> Linux-HA@lists.linux-ha.org >>> >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >>> >>> See also: http://linux-ha.org/ReportingProblems >>> >>> >>> >> >>> >> >>> >> >>> >> -- >>> >> >>> import this >>> >> >>> > >>> > >>> > >>> > -- >>> >>>> import this >>> > _______________________________________________ >>> > Linux-HA mailing list >>> > Linux-HA@lists.linux-ha.org >>> > http://lists.linux-ha.org/mailman/listinfo/linux-ha >>> > See also: http://linux-ha.org/ReportingProblems >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> _______________________________________________ >>> Linux-HA mailing list >>> Linux-HA@lists.linux-ha.org >>> http://lists.linux-ha.org/mailman/listinfo/linux-ha >>> See also: http://linux-ha.org/ReportingProblems >>> >> >> >> >> -- >>>>> import this >> _______________________________________________ >> Linux-HA mailing list >> Linux-HA@lists.linux-ha.org >> http://lists.linux-ha.org/mailman/listinfo/linux-ha >> See also: http://linux-ha.org/ReportingProblems > > > > -- > esta es mi vida e me la vivo hasta que dios quiera -- esta es mi vida e me la vivo hasta que dios quiera _______________________________________________ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems