Since I applied the most recent Pacemaker update last Friday (now
running  pacemaker-1.0.8-2.el5.x86_64 on CentOS 5), I can no longer
enter order directives. I am using the exact same syntax that I used
previously, and the syntax matches some existing directives, but the crm
shell won't take it. Here is an example:

crm(live)configure# primitive VM-cfvmserve ocf:heartbeat:Xen params
xmfile="/etc/xen/cfvmserve" op monitor interval="10" timeout="120"
depth="0" op start interval="0" timeout="60s" op stop interval="0"
timeout="120s" meta allow-migrate="true" target-role="Stopped"
crm(live)configure# order vmnfs-before-cfvmserve inf: vmnfs-cl
VM-cfvmserve
crm(live)configure# verify
ERROR: cib-bootstrap-options: attribute dc-version does not exist
ERROR: cib-bootstrap-options: attribute cluster-infrastructure does not
exist
ERROR: cib-bootstrap-options: attribute last-lrm-refresh does not exist

I don't know why I am getting these errors and I am not sure they are
relevant to the problem I am seeing. But here's what happens later:

crm(live)configure# commit
element rsc_order: validity error : IDREF attribute first references an
unknown ID "vmnfs-cl"

"vmnfs-cl" is a clone resource of a file system mount. That resource is,
and always has been, present. I also have a number of order directives
that reference it that are already in the CIB and are working. Here's a
snippet from "crm configure show":

primitive vmnfs ocf:heartbeat:Filesystem \
        params directory="/vmnfs" device="phantom.ucar.edu:/vol/dsgtest"
fstype="nfs" \
        op start interval="0" timeout="60s" \
        op stop interval="0" timeout="60s"
....
clone vmnfs-cl vmnfs \
        meta target-role="Started"
....
order vmnfs-before-linstall inf: vmnfs-cl VM-linstall

(this last is an example of one that is already present that works).

Lastly, another part of the configure show output addressing the ERROR
messages above:

property $id="cib-bootstrap-options" \
        dc-version="1.0.8-3225fc0d98c8fcd0f7b24f0134e89967136a9b00" \
        cluster-infrastructure="Heartbeat" \
        stonith-enabled="true" \
        last-lrm-refresh="1270484103" \
        default-resource-stickiness=""

This is currently preventing me from being able to add any more virtual
machines. The ones that are already in are working (including proper
failovers and migration). So is this a bug in the new code or is it
something I was doing wrong that is now being flagged that I just
luckily got by with before?

I can send the full configuration if that is deemed necessary but I
would have to sanitize it to remove idrac passwords, local IP addresses,
and so forth, so I won't do that unless it's the only way to figure this
out.

--Greg




_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to