Hi,
Today I was testing my 2 node Xen cluster and noticed a funny thing. I
do have resource stickiness set to infinity however 1 of my VMs (out of
2) called cups was moved back to the rebooted node after joining back to
the cluster. Why is that? Could somebody translate the log for me?
Cib:
<configuration>
<crm_config>
<cluster_property_set id="cibbootstrap">
<attributes>
<nvpair id="cibbootstrap-01" name="transition_idle_timeout"
value="60"/>
<nvpair id="cibbootstrap-02"
name="default_resource_stickiness" value="INFINITY"/>
<nvpair id="cibbootstrap-03"
name="default_resource_failure_stickiness" value="-500"/>
<nvpair id="cibbootstrap-04" name="stonith_enabled"
value="true"/>
<nvpair id="cibbootstrap-05" name="stonith_action"
value="reboot"/>
<nvpair id="cibbootstrap-06" name="symmetric_cluster"
value="true"/>
<nvpair id="cibbootstrap-07" name="no_quorum_policy"
value="stop"/>
<nvpair id="cibbootstrap-08" name="stop_orphan_resources"
value="true"/>
<nvpair id="cibbootstrap-09" name="stop_orphan_actions"
value="true"/>
<nvpair id="cibbootstrap-10" name="is_managed_default"
value="true"/>
</attributes>
</cluster_property_set>
<cluster_property_set id="cib-bootstrap-options">
<attributes>
<nvpair id="cib-bootstrap-options-last-lrm-refresh"
name="last-lrm-refresh" value="1192508059"/>
</attributes>
</cluster_property_set>
</crm_config>
<nodes>
<node id="ec04d3fb-dbad-4176-8a6a-8912aedb3129" uname="apollo"
type="normal"/>
<node id="607b7fd4-298b-4640-b2dc-d06e0a35bcce" uname="artemis"
type="normal"/>
</nodes>
<resources>
<clone id="stonithcloneset" globally_unique="false">
<instance_attributes id="stonithcloneset">
<attributes>
<nvpair id="stonithcloneset-01" name="clone_node_max"
value="1"/>
</attributes>
</instance_attributes>
<primitive id="stonithclone" class="stonith"
type="external/ssh" provider="heartbeat">
<operations>
<op name="monitor" interval="5s" timeout="20s"
prereq="nothing" id="stonithclone-op-01"/>
<op name="start" timeout="20s" prereq="nothing"
id="stonithclone-op-02"/>
</operations>
<instance_attributes id="stonithclone">
<attributes>
<nvpair id="stonithclone-01" name="hostlist"
value="apollo,artemis"/>
</attributes>
</instance_attributes>
</primitive>
</clone>
<clone id="evmsdcloneset" globally_unique="false">
<instance_attributes id="evmsdcloneset">
<attributes>
<nvpair id="evmsdcloneset-01" name="clone_node_max" value="1"/>
<nvpair id="evmsdcloneset_target_role" name="target_role"
value="started"/>
</attributes>
</instance_attributes>
<primitive id="evmsdclone" class="ocf" type="Evmsd"
provider="heartbeat">
<operations>
<op name="monitor" interval="5s" timeout="20s"
id="evmsdclone-op-01"/>
</operations>
<instance_attributes id="evmsdclone:0_instance_attrs">
<attributes>
<nvpair id="evmsdclone:0_target_role" name="target_role"
value="started"/>
</attributes>
</instance_attributes>
</primitive>
</clone>
<clone id="evmscloneset" notify="true" globally_unique="false">
<instance_attributes id="evmscloneset">
<attributes>
<nvpair id="evmscloneset-01" name="clone_node_max" value="1"/>
</attributes>
</instance_attributes>
<primitive id="evmsclone" class="ocf" type="EvmsSCC"
provider="heartbeat"/>
</clone>
<master_slave id="masterslave-drbd0">
<meta_attributes id="ma-masterslave-drbd0">
<attributes>
<nvpair id="ma-masterslave-drbd0-1" name="clone_max"
value="2"/>
<nvpair id="ma-masterslave-drbd0-2" name="clone_node_max"
value="1"/>
<nvpair id="ma-masterslave-drbd0-3" name="master_max"
value="1"/>
<nvpair id="ma-masterslave-drbd0-4" name="master_node_max"
value="1"/>
<nvpair id="ma-masterslave-drbd0-5" name="notify" value="yes"/>
<nvpair id="ma-masterslave-drbd0-6" name="globally_unique"
value="false"/>
</attributes>
</meta_attributes>
<primitive id="drbd0" class="ocf" provider="heartbeat" type="drbd">
<instance_attributes id="ia-drbd0">
<attributes>
<nvpair id="ia-drbd0-1" name="drbd_resource" value="drbd0"/>
</attributes>
</instance_attributes>
</primitive>
</master_slave>
<primitive class="ocf" type="Xen" provider="custom" id="cups">
<operations>
<op name="stop" timeout="60s" id="xen-cups-op-01"/>
<op name="start" timeout="60s" start_delay="0"
id="xen-cups-op-02"/>
<op name="monitor" interval="10s" timeout="60s"
id="xen-cups-op-03"/>
</operations>
<instance_attributes id="cups-instance">
<attributes>
<nvpair id="xen-cups-01" name="xmfile"
value="/etc/xen/vm/cups.xm"/>
</attributes>
</instance_attributes>
<meta_attributes id="cups-meta">
<attributes>
<nvpair id="xen-cups-02" name="allow_migrate" value="true"/>
</attributes>
</meta_attributes>
<instance_attributes id="cups">
<attributes>
<nvpair name="target_role" id="cups-target_role"
value="started"/>
</attributes>
</instance_attributes>
</primitive>
<primitive class="ocf" type="Xen" provider="custom" id="turtle">
<operations>
<op name="stop" timeout="60s" id="xen-turtle-op-01"/>
<op name="start" timeout="60s" start_delay="0"
id="xen-turtle-op-02"/>
<op name="monitor" interval="10s" timeout="60s"
id="xen-turtle-op-03"/>
</operations>
<instance_attributes id="turtle-instance">
<attributes>
<nvpair id="xen-turtle-01" name="xmfile"
value="/etc/xen/vm/turtle.xm"/>
</attributes>
</instance_attributes>
<meta_attributes id="turtle-meta">
<attributes>
<nvpair id="xen-turtle-02" name="allow_migrate" value="true"/>
</attributes>
</meta_attributes>
<instance_attributes id="turtle">
<attributes>
<nvpair name="target_role" id="turtle-target_role"
value="started"/>
</attributes>
</instance_attributes>
</primitive>
<primitive class="ocf" type="Xen" provider="custom" id="lizard">
<operations>
<op name="stop" timeout="60s" id="xen-lizard-op-01"/>
<op name="start" timeout="60s" start_delay="0"
id="xen-lizard-op-02"/>
<op name="monitor" interval="10s" timeout="60s"
id="xen-lizard-op-03"/>
</operations>
<instance_attributes id="lizard-instance">
<attributes>
<nvpair id="xen-lizard-01" name="xmfile"
value="/etc/xen/vm/lizard.xm"/>
<nvpair id="lizard_target_role" name="target_role"
value="started"/>
</attributes>
</instance_attributes>
<instance_attributes id="lizard">
<attributes>
<nvpair name="target_role" id="lizard-target_role"
value="started"/>
</attributes>
</instance_attributes>
</primitive>
<clone id="configpoolcloneset" notify="true" globally_unique="false">
<instance_attributes id="configpoolcloneset">
<attributes>
<nvpair id="configpoolcloneset-01" name="clone_node_max"
value="1"/>
<nvpair id="configpoolcloneset_target_role"
name="target_role" value="started"/>
</attributes>
</instance_attributes>
<primitive id="configpoolclone" class="ocf" type="Filesystem"
provider="heartbeat">
<operations>
<op name="monitor" interval="20s" timeout="60s"
prereq="nothing" id="configpoolclone-op-01"/>
<op name="stop" timeout="60s" prereq="nothing"
id="configpoolclone-op-02"/>
</operations>
<instance_attributes id="configpoolclone">
<attributes>
<nvpair id="configpoolclone-01" name="device"
value="/dev/evms/san2/cfgpool"/>
<nvpair id="configpoolclone-02" name="directory"
value="/etc/xen/vm"/>
<nvpair id="configpoolclone-03" name="fstype" value="ocfs2"/>
<nvpair id="configpoolclone:0_target_role"
name="target_role" value="started"/>
</attributes>
</instance_attributes>
</primitive>
</clone>
</resources>
<constraints>
<rsc_location id="cups_startup_location" rsc="cups">
<rule id="cups-preferred-location" score="INFINITY"
boolean_op="and">
<expression id="cups-preferred-location-01"
attribute="#uname" operation="eq" value="apollo"/>
</rule>
</rsc_location>
<rsc_order id="cups_to_config" from="cups"
to="configpoolcloneset" type="after"/>
<rsc_order id="lizard_to_config" from="lizard"
to="configpoolcloneset" type="after"/>
<rsc_order id="lizard_to_drbd0" from="lizard"
to="masterslave-drbd0" action="start" to_action="promote" type="after"/>
<rsc_colocation id="lizard_to_master_drbd0" from="lizard"
to="masterslave-drbd0" to_role="master" score="INFINITY"/>
<rsc_colocation id="lizard_to_slave_drbd0" from="lizard"
to="masterslave-drbd0" to_role="slave" score="-INFINITY"/>
<rsc_location id="turtle_startup_location" rsc="turtle">
<rule id="turtle-preferred-location" score="INFINITY"
boolean_op="and">
<expression id="turtle-preferred-location-01"
attribute="#uname" operation="eq" value="apollo"/>
</rule>
</rsc_location>
<rsc_order id="turtle_to_config" from="turtle"
to="configpoolclonetset" type="after"/>
<rsc_order id="config_to_evms" from="configpoolcloneset"
to="evmscloneset" type="after"/>
<rsc_order id="evms_to_evmsd" from="evmscloneset"
to="evmsdcloneset" type="after"/>
</constraints>
</configuration>
Log:
tengine[27044]: 2007/10/17_17:15:39 info: process_graph_event: Action
cups_monitor_0 initiated by a different transitioner
pengine[27045]: 2007/10/17_17:15:40 info: native_print: cups
(custom::ocf:Xen): Started apollo
pengine[27045]: 2007/10/17_17:15:40 notice: NoRoleChange: Move resource
cups (apollo -> artemis)
pengine[27045]: 2007/10/17_17:15:40 notice: StopRsc: apollo Stop cups
pengine[27045]: 2007/10/17_17:15:40 WARN: custom_action: Action
cups_stop_0 on apollo is unrunnable (offline)
pengine[27045]: 2007/10/17_17:15:40 notice: StartRsc: artemis Start cups
pengine[27045]: 2007/10/17_17:15:40 notice: RecurringOp: artemis
cups_monitor_10000
pengine[27045]: 2007/10/17_17:15:40 WARN: native_stop_constraints: Stop
of failed resource cups_stop_0 is implicit after apollo is fenced
crmd[5763]: 2007/10/17_17:15:48 info: do_lrm_rsc_op: Performing
op=cups_start_0 key=65:0:f0b3804b-8384-4a6f-af75-724def926045)
tengine[27044]: 2007/10/17_17:15:48 info: send_rsc_command: Initiating
action 65: cups_start_0 on artemis
lrmd[5760]: 2007/10/17_17:15:48 info: RA output: (cups:start:stderr) Error:
lrmd[5760]: 2007/10/17_17:15:48 info: RA output: (cups:start:stderr) No
domain named 'cups'.
lrmd[5760]: 2007/10/17_17:15:57 info: RA output: (cups:start:stdout)
Using config file "/etc/xen/vm/cups.xm".
Started domain cups
crmd[5763]: 2007/10/17_17:15:57 info: process_lrm_event: LRM operation
cups_start_0 (call=31, rc=0) complete
tengine[27044]: 2007/10/17_17:15:57 info: match_graph_event: Action
cups_start_0 (65) confirmed on 607b7fd4-298b-4640-b2dc-d06e0a35bcce
pengine[27045]: 2007/10/17_17:15:57 info: native_print: cups
(custom::ocf:Xen): Started artemis
pengine[27045]: 2007/10/17_17:15:57 notice: NoRoleChange: Leave resource
cups (artemis)
pengine[27045]: 2007/10/17_17:15:57 notice: RecurringOp: artemis
cups_monitor_10000
crmd[5763]: 2007/10/17_17:15:57 info: do_lrm_rsc_op: Performing
op=cups_monitor_10000 key=61:1:f0b3804b-8384-4a6f-af75-724def926045)
tengine[27044]: 2007/10/17_17:15:57 info: send_rsc_command: Initiating
action 61: cups_monitor_10000 on artemis
crmd[5763]: 2007/10/17_17:15:58 info: process_lrm_event: LRM operation
cups_monitor_10000 (call=35, rc=0) complete
tengine[27044]: 2007/10/17_17:15:58 info: match_graph_event: Action
cups_monitor_10000 (61) confirmed on 607b7fd4-298b-4640-b2dc-d06e0a35bcce
pengine[27045]: 2007/10/17_17:15:58 info: native_print: cups
(custom::ocf:Xen): Started artemis
pengine[27045]: 2007/10/17_17:15:58 notice: NoRoleChange: Leave resource
cups (artemis)
pengine[27045]: 2007/10/17_17:18:10 info: native_print: cups
(custom::ocf:Xen): Started artemis
pengine[27045]: 2007/10/17_17:18:10 notice: NoRoleChange: Move resource
cups (artemis -> apollo)
pengine[27045]: 2007/10/17_17:18:10 notice: StopRsc: artemis Stop cups
pengine[27045]: 2007/10/17_17:18:10 notice: StartRsc: apollo Start cups
pengine[27045]: 2007/10/17_17:18:10 notice: RecurringOp: apollo
cups_monitor_10000
pengine[27045]: 2007/10/17_17:18:10 info: native_migrate_reload:
Migrating cups from artemis to apollo
tengine[27044]: 2007/10/17_17:18:10 info: send_rsc_command: Initiating
action 12: cups_monitor_0 on apollo
tengine[27044]: 2007/10/17_17:18:13 info: match_graph_event: Action
cups_monitor_0 (12) confirmed on ec04d3fb-dbad-4176-8a6a-8912aedb3129
pengine[27045]: 2007/10/17_17:18:13 info: native_print: cups
(custom::ocf:Xen): Started artemis
pengine[27045]: 2007/10/17_17:18:13 notice: NoRoleChange: Move resource
cups (artemis -> apollo)
pengine[27045]: 2007/10/17_17:18:13 notice: StopRsc: artemis Stop cups
pengine[27045]: 2007/10/17_17:18:13 notice: StartRsc: apollo Start cups
pengine[27045]: 2007/10/17_17:18:13 notice: RecurringOp: apollo
cups_monitor_10000
pengine[27045]: 2007/10/17_17:18:13 info: native_migrate_reload:
Migrating cups from artemis to apollo
pengine[27045]: 2007/10/17_17:18:22 info: native_print: cups
(custom::ocf:Xen): Started artemis
pengine[27045]: 2007/10/17_17:18:22 notice: NoRoleChange: Move resource
cups (artemis -> apollo)
pengine[27045]: 2007/10/17_17:18:22 notice: StopRsc: artemis Stop cups
pengine[27045]: 2007/10/17_17:18:22 notice: StartRsc: apollo Start cups
pengine[27045]: 2007/10/17_17:18:22 notice: RecurringOp: apollo
cups_monitor_10000
pengine[27045]: 2007/10/17_17:18:22 info: native_migrate_reload:
Migrating cups from artemis to apollo
pengine[27045]: 2007/10/17_17:18:35 info: native_print: cups
(custom::ocf:Xen): Started artemis
pengine[27045]: 2007/10/17_17:18:35 notice: NoRoleChange: Move resource
cups (artemis -> apollo)
pengine[27045]: 2007/10/17_17:18:35 notice: StopRsc: artemis Stop cups
pengine[27045]: 2007/10/17_17:18:35 notice: StartRsc: apollo Start cups
pengine[27045]: 2007/10/17_17:18:35 notice: RecurringOp: apollo
cups_monitor_10000
pengine[27045]: 2007/10/17_17:18:35 info: native_migrate_reload: cups:
start depends on configpoolcloneset_stop_0
pengine[27045]: 2007/10/17_17:18:35 info: native_migrate_reload: cups:
start depends on configpoolcloneset_confirmed-post_notify_start_0
tengine[27044]: 2007/10/17_17:18:35 info: send_rsc_command: Initiating
action 72: cups_stop_0 on artemis
crmd[5763]: 2007/10/17_17:18:35 info: do_lrm_rsc_op: Performing
op=cups_stop_0 key=72:6:f0b3804b-8384-4a6f-af75-724def926045)
crmd[5763]: 2007/10/17_17:18:35 WARN: process_lrm_event: LRM operation
cups_monitor_10000 (call=35, rc=-2) Cancelled
lrmd[5760]: 2007/10/17_17:19:15 info: RA output: (cups:stop:stdout)
Domain cups terminated
crmd[5763]: 2007/10/17_17:19:15 info: process_lrm_event: LRM operation
cups_stop_0 (call=57, rc=0) complete
tengine[27044]: 2007/10/17_17:19:15 info: match_graph_event: Action
cups_stop_0 (72) confirmed on 607b7fd4-298b-4640-b2dc-d06e0a35bcce
tengine[27044]: 2007/10/17_17:19:24 info: send_rsc_command: Initiating
action 73: cups_start_0 on apollo
tengine[27044]: 2007/10/17_17:19:30 info: match_graph_event: Action
cups_start_0 (73) confirmed on ec04d3fb-dbad-4176-8a6a-8912aedb3129
tengine[27044]: 2007/10/17_17:19:30 info: send_rsc_command: Initiating
action 74: cups_monitor_10000 on apollo
tengine[27044]: 2007/10/17_17:19:32 info: match_graph_event: Action
cups_monitor_10000 (74) confirmed on ec04d3fb-dbad-4176-8a6a-8912aedb3129
Thanks,
Ivan
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems