Hi,

Today I was testing my 2 node Xen cluster and noticed a funny thing. I
do have resource stickiness set to infinity however 1 of my VMs (out of
2) called cups was moved back to the rebooted node after joining back to
the cluster. Why is that? Could somebody translate the log for me?

Cib:

   <configuration>
     <crm_config>
       <cluster_property_set id="cibbootstrap">
         <attributes>
           <nvpair id="cibbootstrap-01" name="transition_idle_timeout"
value="60"/>
           <nvpair id="cibbootstrap-02"
name="default_resource_stickiness" value="INFINITY"/>
           <nvpair id="cibbootstrap-03"
name="default_resource_failure_stickiness" value="-500"/>
           <nvpair id="cibbootstrap-04" name="stonith_enabled"
value="true"/>
           <nvpair id="cibbootstrap-05" name="stonith_action"
value="reboot"/>
           <nvpair id="cibbootstrap-06" name="symmetric_cluster"
value="true"/>
           <nvpair id="cibbootstrap-07" name="no_quorum_policy"
value="stop"/>
           <nvpair id="cibbootstrap-08" name="stop_orphan_resources"
value="true"/>
           <nvpair id="cibbootstrap-09" name="stop_orphan_actions"
value="true"/>
           <nvpair id="cibbootstrap-10" name="is_managed_default"
value="true"/>
         </attributes>
       </cluster_property_set>
       <cluster_property_set id="cib-bootstrap-options">
         <attributes>
           <nvpair id="cib-bootstrap-options-last-lrm-refresh"
name="last-lrm-refresh" value="1192508059"/>
         </attributes>
       </cluster_property_set>
     </crm_config>
     <nodes>
       <node id="ec04d3fb-dbad-4176-8a6a-8912aedb3129" uname="apollo"
type="normal"/>
       <node id="607b7fd4-298b-4640-b2dc-d06e0a35bcce" uname="artemis"
type="normal"/>
     </nodes>
     <resources>
       <clone id="stonithcloneset" globally_unique="false">
         <instance_attributes id="stonithcloneset">
           <attributes>
             <nvpair id="stonithcloneset-01" name="clone_node_max"
value="1"/>
           </attributes>
         </instance_attributes>
         <primitive id="stonithclone" class="stonith"
type="external/ssh" provider="heartbeat">
           <operations>
             <op name="monitor" interval="5s" timeout="20s"
prereq="nothing" id="stonithclone-op-01"/>
             <op name="start" timeout="20s" prereq="nothing"
id="stonithclone-op-02"/>
           </operations>
           <instance_attributes id="stonithclone">
             <attributes>
               <nvpair id="stonithclone-01" name="hostlist"
value="apollo,artemis"/>
             </attributes>
           </instance_attributes>
         </primitive>
       </clone>
       <clone id="evmsdcloneset" globally_unique="false">
         <instance_attributes id="evmsdcloneset">
           <attributes>
             <nvpair id="evmsdcloneset-01" name="clone_node_max" value="1"/>
             <nvpair id="evmsdcloneset_target_role" name="target_role"
value="started"/>
           </attributes>
         </instance_attributes>
         <primitive id="evmsdclone" class="ocf" type="Evmsd"
provider="heartbeat">
           <operations>
             <op name="monitor" interval="5s" timeout="20s"
id="evmsdclone-op-01"/>
           </operations>
           <instance_attributes id="evmsdclone:0_instance_attrs">
             <attributes>
               <nvpair id="evmsdclone:0_target_role" name="target_role"
value="started"/>
             </attributes>
           </instance_attributes>
         </primitive>
       </clone>
       <clone id="evmscloneset" notify="true" globally_unique="false">
         <instance_attributes id="evmscloneset">
           <attributes>
             <nvpair id="evmscloneset-01" name="clone_node_max" value="1"/>
           </attributes>
         </instance_attributes>
         <primitive id="evmsclone" class="ocf" type="EvmsSCC"
provider="heartbeat"/>
       </clone>
       <master_slave id="masterslave-drbd0">
         <meta_attributes id="ma-masterslave-drbd0">
           <attributes>
             <nvpair id="ma-masterslave-drbd0-1" name="clone_max"
value="2"/>
             <nvpair id="ma-masterslave-drbd0-2" name="clone_node_max"
value="1"/>
             <nvpair id="ma-masterslave-drbd0-3" name="master_max"
value="1"/>
             <nvpair id="ma-masterslave-drbd0-4" name="master_node_max"
value="1"/>
             <nvpair id="ma-masterslave-drbd0-5" name="notify" value="yes"/>
             <nvpair id="ma-masterslave-drbd0-6" name="globally_unique"
value="false"/>
           </attributes>
         </meta_attributes>
         <primitive id="drbd0" class="ocf" provider="heartbeat" type="drbd">
           <instance_attributes id="ia-drbd0">
             <attributes>
               <nvpair id="ia-drbd0-1" name="drbd_resource" value="drbd0"/>
             </attributes>
           </instance_attributes>
         </primitive>
       </master_slave>
       <primitive class="ocf" type="Xen" provider="custom" id="cups">
         <operations>
           <op name="stop" timeout="60s" id="xen-cups-op-01"/>
           <op name="start" timeout="60s" start_delay="0"
id="xen-cups-op-02"/>
           <op name="monitor" interval="10s" timeout="60s"
id="xen-cups-op-03"/>
         </operations>
         <instance_attributes id="cups-instance">
           <attributes>
             <nvpair id="xen-cups-01" name="xmfile"
value="/etc/xen/vm/cups.xm"/>
           </attributes>
         </instance_attributes>
         <meta_attributes id="cups-meta">
           <attributes>
             <nvpair id="xen-cups-02" name="allow_migrate" value="true"/>
           </attributes>
         </meta_attributes>
         <instance_attributes id="cups">
           <attributes>
             <nvpair name="target_role" id="cups-target_role"
value="started"/>
           </attributes>
         </instance_attributes>
       </primitive>
       <primitive class="ocf" type="Xen" provider="custom" id="turtle">
         <operations>
           <op name="stop" timeout="60s" id="xen-turtle-op-01"/>
           <op name="start" timeout="60s" start_delay="0"
id="xen-turtle-op-02"/>
           <op name="monitor" interval="10s" timeout="60s"
id="xen-turtle-op-03"/>
         </operations>
         <instance_attributes id="turtle-instance">
           <attributes>
             <nvpair id="xen-turtle-01" name="xmfile"
value="/etc/xen/vm/turtle.xm"/>
           </attributes>
         </instance_attributes>
         <meta_attributes id="turtle-meta">
           <attributes>
             <nvpair id="xen-turtle-02" name="allow_migrate" value="true"/>
           </attributes>
         </meta_attributes>
         <instance_attributes id="turtle">
           <attributes>
             <nvpair name="target_role" id="turtle-target_role"
value="started"/>
           </attributes>
         </instance_attributes>
       </primitive>
       <primitive class="ocf" type="Xen" provider="custom" id="lizard">
         <operations>
           <op name="stop" timeout="60s" id="xen-lizard-op-01"/>
           <op name="start" timeout="60s" start_delay="0"
id="xen-lizard-op-02"/>
           <op name="monitor" interval="10s" timeout="60s"
id="xen-lizard-op-03"/>
         </operations>
         <instance_attributes id="lizard-instance">
           <attributes>
             <nvpair id="xen-lizard-01" name="xmfile"
value="/etc/xen/vm/lizard.xm"/>
             <nvpair id="lizard_target_role" name="target_role"
value="started"/>
           </attributes>
         </instance_attributes>
         <instance_attributes id="lizard">
           <attributes>
             <nvpair name="target_role" id="lizard-target_role"
value="started"/>
           </attributes>
         </instance_attributes>
       </primitive>
       <clone id="configpoolcloneset" notify="true" globally_unique="false">
         <instance_attributes id="configpoolcloneset">
           <attributes>
             <nvpair id="configpoolcloneset-01" name="clone_node_max"
value="1"/>
             <nvpair id="configpoolcloneset_target_role"
name="target_role" value="started"/>
           </attributes>
         </instance_attributes>
         <primitive id="configpoolclone" class="ocf" type="Filesystem"
provider="heartbeat">
           <operations>
             <op name="monitor" interval="20s" timeout="60s"
prereq="nothing" id="configpoolclone-op-01"/>
             <op name="stop" timeout="60s" prereq="nothing"
id="configpoolclone-op-02"/>
           </operations>
           <instance_attributes id="configpoolclone">
             <attributes>
               <nvpair id="configpoolclone-01" name="device"
value="/dev/evms/san2/cfgpool"/>
               <nvpair id="configpoolclone-02" name="directory"
value="/etc/xen/vm"/>
               <nvpair id="configpoolclone-03" name="fstype" value="ocfs2"/>
               <nvpair id="configpoolclone:0_target_role"
name="target_role" value="started"/>
             </attributes>
           </instance_attributes>
         </primitive>
       </clone>
     </resources>
     <constraints>
       <rsc_location id="cups_startup_location" rsc="cups">
         <rule id="cups-preferred-location" score="INFINITY"
boolean_op="and">
           <expression id="cups-preferred-location-01"
attribute="#uname" operation="eq" value="apollo"/>
         </rule>
       </rsc_location>
       <rsc_order id="cups_to_config" from="cups"
to="configpoolcloneset" type="after"/>
       <rsc_order id="lizard_to_config" from="lizard"
to="configpoolcloneset" type="after"/>
       <rsc_order id="lizard_to_drbd0" from="lizard"
to="masterslave-drbd0" action="start" to_action="promote" type="after"/>
       <rsc_colocation id="lizard_to_master_drbd0" from="lizard"
to="masterslave-drbd0" to_role="master" score="INFINITY"/>
       <rsc_colocation id="lizard_to_slave_drbd0" from="lizard"
to="masterslave-drbd0" to_role="slave" score="-INFINITY"/>
       <rsc_location id="turtle_startup_location" rsc="turtle">
         <rule id="turtle-preferred-location" score="INFINITY"
boolean_op="and">
           <expression id="turtle-preferred-location-01"
attribute="#uname" operation="eq" value="apollo"/>
         </rule>
       </rsc_location>
       <rsc_order id="turtle_to_config" from="turtle"
to="configpoolclonetset" type="after"/>
       <rsc_order id="config_to_evms" from="configpoolcloneset"
to="evmscloneset" type="after"/>
       <rsc_order id="evms_to_evmsd" from="evmscloneset"
to="evmsdcloneset" type="after"/>
     </constraints>
   </configuration>

Log:

tengine[27044]: 2007/10/17_17:15:39 info: process_graph_event: Action
cups_monitor_0 initiated by a different transitioner
pengine[27045]: 2007/10/17_17:15:40 info: native_print: cups
(custom::ocf:Xen):      Started apollo
pengine[27045]: 2007/10/17_17:15:40 notice: NoRoleChange: Move  resource
cups   (apollo -> artemis)
pengine[27045]: 2007/10/17_17:15:40 notice: StopRsc:   apollo   Stop cups
pengine[27045]: 2007/10/17_17:15:40 WARN: custom_action: Action
cups_stop_0 on apollo is unrunnable (offline)
pengine[27045]: 2007/10/17_17:15:40 notice: StartRsc:  artemis  Start cups
pengine[27045]: 2007/10/17_17:15:40 notice: RecurringOp: artemis
   cups_monitor_10000
pengine[27045]: 2007/10/17_17:15:40 WARN: native_stop_constraints: Stop
of failed resource cups_stop_0 is implicit after apollo is fenced
crmd[5763]: 2007/10/17_17:15:48 info: do_lrm_rsc_op: Performing
op=cups_start_0 key=65:0:f0b3804b-8384-4a6f-af75-724def926045)
tengine[27044]: 2007/10/17_17:15:48 info: send_rsc_command: Initiating
action 65: cups_start_0 on artemis
lrmd[5760]: 2007/10/17_17:15:48 info: RA output: (cups:start:stderr) Error:
lrmd[5760]: 2007/10/17_17:15:48 info: RA output: (cups:start:stderr)  No
domain named 'cups'.
lrmd[5760]: 2007/10/17_17:15:57 info: RA output: (cups:start:stdout)
Using config file "/etc/xen/vm/cups.xm".
Started domain cups
crmd[5763]: 2007/10/17_17:15:57 info: process_lrm_event: LRM operation
cups_start_0 (call=31, rc=0) complete
tengine[27044]: 2007/10/17_17:15:57 info: match_graph_event: Action
cups_start_0 (65) confirmed on 607b7fd4-298b-4640-b2dc-d06e0a35bcce
pengine[27045]: 2007/10/17_17:15:57 info: native_print: cups
(custom::ocf:Xen):      Started artemis
pengine[27045]: 2007/10/17_17:15:57 notice: NoRoleChange: Leave resource
cups   (artemis)
pengine[27045]: 2007/10/17_17:15:57 notice: RecurringOp: artemis
   cups_monitor_10000
crmd[5763]: 2007/10/17_17:15:57 info: do_lrm_rsc_op: Performing
op=cups_monitor_10000 key=61:1:f0b3804b-8384-4a6f-af75-724def926045)
tengine[27044]: 2007/10/17_17:15:57 info: send_rsc_command: Initiating
action 61: cups_monitor_10000 on artemis
crmd[5763]: 2007/10/17_17:15:58 info: process_lrm_event: LRM operation
cups_monitor_10000 (call=35, rc=0) complete
tengine[27044]: 2007/10/17_17:15:58 info: match_graph_event: Action
cups_monitor_10000 (61) confirmed on 607b7fd4-298b-4640-b2dc-d06e0a35bcce
pengine[27045]: 2007/10/17_17:15:58 info: native_print: cups
(custom::ocf:Xen):      Started artemis
pengine[27045]: 2007/10/17_17:15:58 notice: NoRoleChange: Leave resource
cups   (artemis)
pengine[27045]: 2007/10/17_17:18:10 info: native_print: cups
(custom::ocf:Xen):      Started artemis
pengine[27045]: 2007/10/17_17:18:10 notice: NoRoleChange: Move  resource
cups   (artemis -> apollo)
pengine[27045]: 2007/10/17_17:18:10 notice: StopRsc:   artemis  Stop cups
pengine[27045]: 2007/10/17_17:18:10 notice: StartRsc:  apollo   Start cups
pengine[27045]: 2007/10/17_17:18:10 notice: RecurringOp: apollo
cups_monitor_10000
pengine[27045]: 2007/10/17_17:18:10 info: native_migrate_reload:
Migrating cups from artemis to apollo
tengine[27044]: 2007/10/17_17:18:10 info: send_rsc_command: Initiating
action 12: cups_monitor_0 on apollo
tengine[27044]: 2007/10/17_17:18:13 info: match_graph_event: Action
cups_monitor_0 (12) confirmed on ec04d3fb-dbad-4176-8a6a-8912aedb3129
pengine[27045]: 2007/10/17_17:18:13 info: native_print: cups
(custom::ocf:Xen):      Started artemis
pengine[27045]: 2007/10/17_17:18:13 notice: NoRoleChange: Move  resource
cups   (artemis -> apollo)
pengine[27045]: 2007/10/17_17:18:13 notice: StopRsc:   artemis  Stop cups
pengine[27045]: 2007/10/17_17:18:13 notice: StartRsc:  apollo   Start cups
pengine[27045]: 2007/10/17_17:18:13 notice: RecurringOp: apollo
cups_monitor_10000
pengine[27045]: 2007/10/17_17:18:13 info: native_migrate_reload:
Migrating cups from artemis to apollo
pengine[27045]: 2007/10/17_17:18:22 info: native_print: cups
(custom::ocf:Xen):      Started artemis
pengine[27045]: 2007/10/17_17:18:22 notice: NoRoleChange: Move  resource
cups   (artemis -> apollo)
pengine[27045]: 2007/10/17_17:18:22 notice: StopRsc:   artemis  Stop cups
pengine[27045]: 2007/10/17_17:18:22 notice: StartRsc:  apollo   Start cups
pengine[27045]: 2007/10/17_17:18:22 notice: RecurringOp: apollo
cups_monitor_10000
pengine[27045]: 2007/10/17_17:18:22 info: native_migrate_reload:
Migrating cups from artemis to apollo
pengine[27045]: 2007/10/17_17:18:35 info: native_print: cups
(custom::ocf:Xen):      Started artemis
pengine[27045]: 2007/10/17_17:18:35 notice: NoRoleChange: Move  resource
cups   (artemis -> apollo)
pengine[27045]: 2007/10/17_17:18:35 notice: StopRsc:   artemis  Stop cups
pengine[27045]: 2007/10/17_17:18:35 notice: StartRsc:  apollo   Start cups
pengine[27045]: 2007/10/17_17:18:35 notice: RecurringOp: apollo
cups_monitor_10000
pengine[27045]: 2007/10/17_17:18:35 info: native_migrate_reload: cups:
start depends on configpoolcloneset_stop_0
pengine[27045]: 2007/10/17_17:18:35 info: native_migrate_reload: cups:
start depends on configpoolcloneset_confirmed-post_notify_start_0
tengine[27044]: 2007/10/17_17:18:35 info: send_rsc_command: Initiating
action 72: cups_stop_0 on artemis
crmd[5763]: 2007/10/17_17:18:35 info: do_lrm_rsc_op: Performing
op=cups_stop_0 key=72:6:f0b3804b-8384-4a6f-af75-724def926045)
crmd[5763]: 2007/10/17_17:18:35 WARN: process_lrm_event: LRM operation
cups_monitor_10000 (call=35, rc=-2) Cancelled
lrmd[5760]: 2007/10/17_17:19:15 info: RA output: (cups:stop:stdout)
Domain cups terminated
crmd[5763]: 2007/10/17_17:19:15 info: process_lrm_event: LRM operation
cups_stop_0 (call=57, rc=0) complete
tengine[27044]: 2007/10/17_17:19:15 info: match_graph_event: Action
cups_stop_0 (72) confirmed on 607b7fd4-298b-4640-b2dc-d06e0a35bcce
tengine[27044]: 2007/10/17_17:19:24 info: send_rsc_command: Initiating
action 73: cups_start_0 on apollo
tengine[27044]: 2007/10/17_17:19:30 info: match_graph_event: Action
cups_start_0 (73) confirmed on ec04d3fb-dbad-4176-8a6a-8912aedb3129
tengine[27044]: 2007/10/17_17:19:30 info: send_rsc_command: Initiating
action 74: cups_monitor_10000 on apollo
tengine[27044]: 2007/10/17_17:19:32 info: match_graph_event: Action
cups_monitor_10000 (74) confirmed on ec04d3fb-dbad-4176-8a6a-8912aedb3129

Thanks,
Ivan
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to