On 03/02/2017 07:09 AM, Anne Nicolas wrote:

Le 01/03/2017 à 23:20, Ken Gaillot a écrit :
On 03/01/2017 03:49 PM, Anne Nicolas wrote:
Hi there


I'm testing quite an easy configuration to work on clvm. I'm just
getting crazy as it seems clmd cannot be cloned on other nodes.

clvmd start well on node1 but fails on both node2 and node3.
Your config looks fine, so I'm going to guess there's some local
difference on the nodes.

In pacemaker journalctl I get the following message
Mar 01 16:34:36 node3 pidofproc[27391]: pidofproc: cannot stat /clvmd:
No such file or directory
Mar 01 16:34:36 node3 pidofproc[27392]: pidofproc: cannot stat
/cmirrord: No such file or directory
I have no idea where the above is coming from. pidofproc is an LSB
function, but (given journalctl) I'm assuming you're using systemd. I
don't think anything in pacemaker or resource-agents uses pidofproc (at
least not currently, not sure about the older version you're using).

Thanks for your feedback. I finally checked the RA script and found the
error

in clvm2 RA script on non working nodes I got
# Common variables
DAEMON="${sbindir}/clvmd"
CMIRRORD="${sbindir}/cmirrord"
LVMCONF="${sbindir}/lvmconf"

on working node
DAEMON="/usr/sbin/clvmd"
CMIRRORD="/usr/sbin/cmirrord"

Looks like it was path variables were not interpreted. I just have to
check why I did get those versions
A bugfix for this issue has been released in lvm2 2.02.120-70.1. And, since 
SLE12-SP2
and openSUSE leap42.2, we recommend using 
'/usr/lib/ocf/resource.d/heartbeat/clvm'
instead, which is from 'resource-agents' package.

Eric

THanks again for your answer

Mar 01 16:34:36 node3 lrmd[2174]: notice: finished - rsc:p-clvmd
action:stop call_id:233 pid:27384 exit-code:0 exec-time:45ms queue-time:0ms
Mar 01 16:34:36 node3 crmd[2177]: notice: Operation p-clvmd_stop_0: ok
(node=node3, call=233, rc=0, cib-update=541, confirmed=true)
Mar 01 16:34:36 node3 crmd[2177]: notice: Initiating action 72: stop
p-dlm_stop_0 on node3 (local)
Mar 01 16:34:36 node3 lrmd[2174]: notice: executing - rsc:p-dlm
action:stop call_id:235
Mar 01 16:34:36 node3 crmd[2177]: notice: Initiating action 67: stop
p-dlm_stop_0 on node2

Here is my configuration

node 739312139: node1
node 739312140: node2
node 739312141: node3
primitive admin_addr IPaddr2 \
         params ip=172.17.2.10 \
         op monitor interval=10 timeout=20 \
         meta target-role=Started
primitive p-clvmd ocf:lvm2:clvmd \
         op start timeout=90 interval=0 \
         op stop timeout=100 interval=0 \
         op monitor interval=30 timeout=90
primitive p-dlm ocf:pacemaker:controld \
         op start timeout=90 interval=0 \
         op stop timeout=100 interval=0 \
         op monitor interval=60 timeout=90
primitive stonith-sbd stonith:external/sbd
group g-clvm p-dlm p-clvmd
clone c-clvm g-clvm meta interleave=true
property cib-bootstrap-options: \
         have-watchdog=true \
         dc-version=1.1.13-14.7-6f22ad7 \
         cluster-infrastructure=corosync \
         cluster-name=hacluster \
         stonith-enabled=true \
         placement-strategy=balanced \
         no-quorum-policy=freeze \
         last-lrm-refresh=1488404073
rsc_defaults rsc-options: \
         resource-stickiness=1 \
         migration-threshold=10
op_defaults op-options: \
         timeout=600 \
         record-pending=true

Thanks in advance for your input

Cheers


_______________________________________________
Users mailing list: [email protected]
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org



_______________________________________________
Users mailing list: [email protected]
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to