On 8/22/07, Dejan Muhamedagic <[EMAIL PROTECTED]> wrote:
> On Wed, Aug 22, 2007 at 05:37:01PM +0200, Hubeli Daniel wrote:
> > Hi all,
> >
> > I'm writing you because I'm having some problems with pingd.
> >
> > My structure is:
> >
> > -     2 nodes cluster
> > -     OS: SLES 10 SP1
> > -     Ha version: 2.0.8
> >
> > I successfully configured an NFS group which works quite good. Now I'd
> > like to monitor 2 IPs; if both nodes see both IPs I'd like to run the
> > resource on my normal preferred node (assigned with a score attribute)
> > but if a node see just 1 IP and the other node see both IP I'd like to
> > switch the resource group on the node with better connectivity.
> >
> > My actual configuration seems not working but I don't understand why (to
> > me seems correctly configured).
>
> You refer to the pingd attribute in constraints by name
> "group_pingd". It should be "pingd". Or you can change the
> attribute name in the pingd primitive definition in the CIB (not
> sure about the attribute's name, you can get by checking the
> pingd RA's meta-data).

oh ... I've overlooked the wrong attribute name ... thanks Dejan for
pointing that out!

Regards,
Andreas

>
> In the logs I see references to some resources which are not in
> the CIB: nfsp*, nfst*, oe10*, perhaps more. How comes?
>
> > If someone has any hints it would be great ...
>
> Try to use ptest on peinputs which occur when the connectivity
> changes. That should tell you what the cluster thinks it's doing
> (actually, the pengine). Check the status section to see if the
> pingd attribute really gets updated correctly (cibadmin -Q).
>
> > My configuration is:
>
> Next time, please attach CIB/logs instead of pasting them in the
> message.
>
> Dejan
>
> > #
> > #
> >  ha.cf
> > #
> > autojoin any
> > crm true
> > bcast eth3
> > bcast eth1
> > node ulxxcapb
> > node ulxxcapa
> > use_logd on
> > respawn root /sbin/evmsd
> > apiauth evms uid=hacluster,root
> > apiauth ping gid=root uid=root
> > ping 9.0.2.90 9.0.1.91
> >
> > #
> > # General config:
> > #
> > <crm_config>
> >   <cluster_property_set id="cib-bootstrap-options">
> >     <attributes>
> >       <nvpair name="last-lrm-refresh"
> > id="cib-bootstrap-options-last-lrm-refresh" value="1184073021"/>
> >       <nvpair name="transition-idle-timeout"
> > id="transition-idle-timeout"                value="60"/>
> >       <nvpair name="symmetric-cluster"
> > id="symmetric-cluster"                      value="true"/>
> >       <nvpair name="no-quorum-policy"
> > id="no-quorum-policy"                       value="stop"/>
> >       <nvpair name="stonith-enabled"
> > id="stonith-enabled"                        value="false"/>
> >       <nvpair name="stonith-action"
> > id="stonith-action"                         value="reboot"/>
> >       <nvpair name="startup-fencing"
> > id="startup-fencing"                        value="true"/>
> >       <nvpair name="is-managed-default"
> > id="is-managed-default"                     value="true"/>
> >       <nvpair name="default-resource-stickiness"
> > id="default-resource-stickiness"            value="100"/>
> >       <nvpair name="default-resource-failure-stickiness"
> > id="default-resource-failure-stickiness"    value="-200"/>
> >       <nvpair name="stop-orphan-resources"
> > id="stop-orphan-resources"                  value="true"/>
> >       <nvpair name="stop-orphan-actions"
> > id="stop-orphan-actions"                    value="true"/>
> >     </attributes>
> >   </cluster_property_set>
> > </crm_config>
> >
> > #
> > # Node config:
> > #
> > <nodes>
> >   <node id="4153c055-d562-46bb-8f33-41023f000ef9" uname="ulxxcapa"
> > type="normal"/>
> >   <node id="acc23203-d2c8-419e-9dfb-ce621b332225" uname="ulxxcapb"
> > type="normal"/>
> > </nodes>
> >
> > #
> > # Resource config (just a NFS Share):
> > #
> >         <group id="group_nfs2">
> >           <primitive class="heartbeat" type="evms_failover"
> > provider="heartbeat" id="nfs2_evms_failover">
> >             <instance_attributes id="nfs2_evms_failover_attrs">
> >               <attributes>
> >                 <nvpair name="target_role"
> > id="target_role_nfs2_evms_failover" value="started"/>
> >                 <nvpair name="1"           id="1_nfs2_evms_failover"
> > value="nfs2"/>
> >               </attributes>
> >             </instance_attributes>
> >             <operations>
> >               <op name="monitor" id="monitor_nfs2_evms_failover"
> > timeout="60s"    interval="30s"/>
> >               <op name="start"   id="start_nfs2_evms_failover"
> > timeout="300s"/>
> >               <op name="stop"    id="stop_nfs2_evms_failover"
> > timeout="300s"/>
> >             </operations>
> >           </primitive>
> >           <primitive class="ocf" type="Filesystem" provider="heartbeat"
> > id="nfs2_filesystem">
> >             <instance_attributes id="nfs2_filesystem_attrs">
> >               <attributes>
> >                 <nvpair name="target_role"
> > id="target_role_nfs2_filesystem"  value="started"/>
> >                 <nvpair name="device"      id="device_nfs2_filesystem"
> > value="/dev/evms/nfs2/nfs2_lv"/>
> >                 <nvpair name="directory"
> > id="directory_nfs2_filesystem"    value="/mnt/nfs2"/>
> >                 <nvpair name="fstype"      id="fstype_nfs2_filesystem"
> > value="ext3"/>
> >               </attributes>
> >             </instance_attributes>
> >             <operations>
> >               <op name="monitor" id="monitor_nfs2_filesystem"
> > timeout="60s" interval="30s"/>
> >               <op name="start"   id="start_nfs2_filesystem"
> > timeout="300s"/>
> >               <op name="stop"    id="stop_nfs2_filesystem"
> > timeout="300s"/>
> >             </operations>
> >           </primitive>
> >           <primitive class="ocf" type="IPaddr2" provider="heartbeat"
> > id="nfs2_ip_1">
> >             <instance_attributes id="nfs2_ip_1_attrs">
> >               <attributes>
> >                 <nvpair name="target_role" id="target_role_nfs2_ip_1"
> > value="started"/>
> >                 <nvpair name="ip"          id="ip_nfs2_ip_1"
> > value="9.0.1.92"/>
> >                 <nvpair name="nic"         id="nic_nfs2_ip_1"
> > value="eth0"/>
> >               </attributes>
> >             </instance_attributes>
> >             <operations>
> >               <op name="monitor" id="monitor_nfs2_ip_1" timeout="10s"
> > interval="5s" />
> >             </operations>
> >           </primitive>
> >           <primitive class="ocf" type="IPaddr2" provider="heartbeat"
> > id="nfs2_ip_2">
> >             <instance_attributes id="nfs2_ip_2_attrs">
> >               <attributes>
> >                 <nvpair name="target_role" id="target_role_nfs2_ip_2"
> > value="started"/>
> >                 <nvpair name="ip"          id="ip_nfs2_ip_2"
> > value="9.0.2.92"/>
> >                 <nvpair name="nic"         id="nic_nfs2_ip_2"
> > value="eth2"/>
> >               </attributes>
> >             </instance_attributes>
> >             <operations>
> >               <op name="monitor" id="monitor_nfs2_ip_2" timeout="10s"
> > interval="5s"/>
> >             </operations>
> >           </primitive>
> >           <primitive id="nfs2_export" class="lsb" type="export_nfs2">
> >             <operations>
> >               <op name="monitor" id="monitor_export_nfs2" timeout="60s"
> > interval="30s" on_fail="restart"/>
> >               <op name="start"   id="start_export_nfs2"
> > timeout="300s"/>
> >               <op name="stop"    id="stop_export_nfs2"
> > timeout="300s"/>
> >             </operations>
> >           </primitive>
> >           <primitive id="nfs2_tsmc" class="lsb" type="dsmc_nfs2">
> >             <operations>
> >               <op name="monitor" id="monitor_tsmc_nfs2" timeout="60s"
> > interval="30s" on_fail="restart"/>
> >               <op name="start"   id="start_tsmc_nfs2"   timeout="300s"/>
> >               <op name="stop"    id="stop_tsmc_nfs2"    timeout="300s"/>
> >             </operations>
> >           </primitive>
> >         </group>
> >
> > #
> > # Pingd definition
> > #
> > <clone id="group_pingd">
> >
> >   <instance_attributes id="group_pingd">
> >     <attributes>
> >       <nvpair id="clone_node_max" name="clone_node_max" value="1"/>
> >     </attributes>
> >   </instance_attributes>
> >
> >   <primitive id="pingd-child" provider="heartbeat" class="ocf"
> > type="pingd">
> >     <instance_attributes id="pingd_inst_attr">
> >       <attributes>
> >          <nvpair id="pingd-dampen"     name="dampen"     value="2s"/>
> >          <nvpair id="pingd-multiplier" name="multiplier" value="100"/>
> >          <nvpair id="pingd-pidfile"    name="pidfile"
> > value="/var/run/pingd.pid"/>
> >          <nvpair id="pingd-user"       name="user"       value="root"/>
> >       </attributes>
> >     </instance_attributes>
> >     <operations>
> >       <op id="pingd-child-start"   name="start"   prereq="nothing"/>
> >       <op id="pingd-child-monitor" name="monitor" interval="4s"
> > timeout="8s" prereq="nothing"/>
> >     </operations>
> >   </primitive>
> >
> > </clone>
> >
> >
> > #
> > # Resource restriction
> > #
> > <constraints>
> >   <rsc_location id="pingd_nfs2_location" rsc="group_nfs2">
> >     <rule id="prefered_place_group_nfs2" score="50">
> >       <expression attribute="#uname" operation="eq" value="ulxxcapb"/>
> >     </rule>
> >     <rule id="pingd_nfs2_rule" score_attribute="group_pingd">
> >       <expression id="pingd_nfs2_conn_defined" attribute="group_pingd"
> > operation="defined"/>
> >     </rule>
> >   </rsc_location>
> > </constraints>
> >
> >
> > If the resource is in the default location (node B) and the node loose
> > one ping node in the log I see the following messages (but the resource
> > remain where it is):
> >
> > heartbeat[3192]: 2007/08/22_17:33:40 WARN: node 9.0.2.90: is dead
> > crmd[3240]: 2007/08/22_17:33:40 notice: crmd_ha_status_callback: Status
> > update: Node 9.0.2.90 now has status [dead]
> > heartbeat[3192]: 2007/08/22_17:33:40 info: Link 9.0.2.90:9.0.2.90 dead.
> > crmd[3240]: 2007/08/22_17:33:40 WARN: get_uuid: Could not calculate UUID
> > for 9.0.2.90
> > crmd[3240]: 2007/08/22_17:33:40 info: crmd_ha_status_callback: Ping node
> > 9.0.2.90 is dead
> > attrd[3239]: 2007/08/22_17:33:42 info: attrd_timer_callback: Sending
> > flush op to all hosts for: pingd
> > attrd[3239]: 2007/08/22_17:33:42 info: attrd_ha_callback: flush message
> > from ulxxcapb
> > attrd[3239]: 2007/08/22_17:33:42 info: attrd_ha_callback: Sent update
> > 13: pingd=100
> > cib[3236]: 2007/08/22_17:33:42 info: cib_diff_notify: Update (client:
> > 3239, call:13): 0.151.14643 -> 0.151.14644 (ok)
> > tengine[3283]: 2007/08/22_17:33:42 info: te_update_diff: Processing diff
> > (cib_modify): 0.151.14643 -> 0.151.14644
> > tengine[3283]: 2007/08/22_17:33:42 info: extract_event: Aborting on
> > transient_attributes changes for acc23203-d2c8-419e-9dfb-ce621b332225
> > tengine[3283]: 2007/08/22_17:33:42 info: update_abort_priority: Abort
> > priority upgraded to 1000000
> > tengine[3283]: 2007/08/22_17:33:42 info: te_update_diff: Aborting on
> > transient_attributes deletions
> > crmd[3240]: 2007/08/22_17:33:42 info: do_state_transition: ulxxcapb:
> > State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC
> > cause=C_IPC_MESSAGE origin=route_message ]
> > crmd[3240]: 2007/08/22_17:33:42 info: do_state_transition: All 2 cluster
> > nodes are eligible to run resources.
> > cib[11297]: 2007/08/22_17:33:42 info: write_cib_contents: Wrote version
> > 0.151.14644 of the CIB to disk (digest:
> > 2f4ba9ed708980450a3a2a4cc9af67a8)
> > pengine[3284]: 2007/08/22_17:33:42 info: log_data_element:
> > process_pe_message: [generation] <cib admin_epoch="0" have_quorum="true"
> > ignore_dtd="false" num_peers="2" cib_feature_revision="1.3"
> > generated="true" epoch="151" num_updates="14644" cib-last-written="Wed
> > Aug 22 17:06:03 2007" ccm_transition="2"
> > dc_uuid="acc23203-d2c8-419e-9dfb-ce621b332225"/>
> > pengine[3284]: 2007/08/22_17:33:42 notice: cluster_option: Using default
> > value '60s' for cluster option 'cluster-delay'
> > pengine[3284]: 2007/08/22_17:33:42 notice: cluster_option: Using default
> > value '20s' for cluster option 'default-action-timeout'
> > pengine[3284]: 2007/08/22_17:33:42 notice: cluster_option: Using default
> > value 'false' for cluster option 'remove-after-stop'
> > pengine[3284]: 2007/08/22_17:33:42 notice: cluster_option: Using default
> > value '-1' for cluster option 'pe-error-series-max'
> > pengine[3284]: 2007/08/22_17:33:42 notice: cluster_option: Using default
> > value '-1' for cluster option 'pe-warn-series-max'
> > pengine[3284]: 2007/08/22_17:33:42 notice: cluster_option: Using default
> > value '-1' for cluster option 'pe-input-series-max'
> > pengine[3284]: 2007/08/22_17:33:42 info: determine_online_status: Node
> > ulxxcapb is online
> > pengine[3284]: 2007/08/22_17:33:42 info: determine_online_status: Node
> > ulxxcapa is online
> > pengine[3284]: 2007/08/22_17:33:42 info: group_print: Resource Group:
> > group_nfsp
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:
> > nfsp_evms_failover   (heartbeat:evms_failover):      Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:
> > nfsp_filesystem      (heartbeat::ocf:Filesystem):    Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     nfsp_ip_1
> > (heartbeat::ocf:IPaddr2):       Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     nfsp_ip_2
> > (heartbeat::ocf:IPaddr2):       Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     nfsp_export
> > (lsb:export_nfsp):      Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: group_print: Resource Group:
> > group_nfst
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:
> > nfst_evms_failover   (heartbeat:evms_failover):      Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:
> > nfst_filesystem      (heartbeat::ocf:Filesystem):    Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     nfst_ip_1
> > (heartbeat::ocf:IPaddr2):       Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     nfst_ip_2
> > (heartbeat::ocf:IPaddr2):       Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     nfst_export
> > (lsb:export_nfst):      Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: group_print: Resource Group:
> > group_oe10
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:
> > oe10_evms_failover   (heartbeat:evms_failover):      Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:
> > oe10_filesystem      (heartbeat::ocf:Filesystem):    Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     oe10_ip_1
> > (heartbeat::ocf:IPaddr2):       Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     oe10_ip_2
> > (heartbeat::ocf:IPaddr2):       Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     oe10_db
> > (heartbeat::ocf:oracle):        Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     oe10_lsnr
> > (heartbeat::ocf:oralsnr):       Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     oe10_tsmc
> > (lsb:dsmc_oe10):        Stopped
> > pengine[3284]: 2007/08/22_17:33:42 info: group_print: Resource Group:
> > group_nfs2
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:
> > nfs2_evms_failover   (heartbeat:evms_failover):      Started ulxxcapb
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:
> > nfs2_filesystem      (heartbeat::ocf:Filesystem):    Started ulxxcapb
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     nfs2_ip_1
> > (heartbeat::ocf:IPaddr2):       Started ulxxcapb
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     nfs2_ip_2
> > (heartbeat::ocf:IPaddr2):       Started ulxxcapb
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     nfs2_export
> > (lsb:export_nfs2):      Started ulxxcapb
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     nfs2_tsmc
> > (lsb:dsmc_nfs2):        Started ulxxcapb
> > pengine[3284]: 2007/08/22_17:33:42 info: clone_print: Clone Set:
> > group_pingd
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     pingd-child:0
> > (heartbeat::ocf:pingd): Started ulxxcapa
> > pengine[3284]: 2007/08/22_17:33:42 info: native_print:     pingd-child:1
> > (heartbeat::ocf:pingd): Started ulxxcapb
> > pengine[3284]: 2007/08/22_17:33:42 info: log_data_element:
> > check_action_definition: params:all <parameters target_role="started"
> > 1="nfs2"/>
> > pengine[3284]: 2007/08/22_17:33:42 WARN: check_action_definition:
> > Parameters to nfs2_evms_failover_monitor_0 on ulxxcapb changed: recorded
> > d32fe35b8f6ec7db0825df29a3063746 vs. calculated (all)
> > 358c4b605b51a03afd9530f9ee88896b
> > pengine[3284]: 2007/08/22_17:33:42 info: log_data_element:
> > check_action_definition: params:all <parameters multiplier="100"
> > dampen="2s" user="root" pidfile="/var/run/pingd.pid"/>
> > pengine[3284]: 2007/08/22_17:33:42 WARN: check_action_definition:
> > Parameters to pingd-child:1_monitor_0 on ulxxcapb changed: recorded
> > 2500f62f8cbe28359717874cda643d0f vs. calculated (all)
> > 6ce8a133bf8292a7521af0d4610e60f2
> > pengine[3284]: 2007/08/22_17:33:42 info: log_data_element:
> > check_action_definition: params:all <parameters multiplier="100"
> > dampen="2s" user="root" pidfile="/var/run/pingd.pid"/>
> > pengine[3284]: 2007/08/22_17:33:42 WARN: check_action_definition:
> > Parameters to pingd-child:1_start_0 on ulxxcapb changed: recorded
> > e58195cc0db685dc80aa15525c617b8a vs. calculated (all)
> > 6ce8a133bf8292a7521af0d4610e60f2
> > pengine[3284]: 2007/08/22_17:33:42 info: log_data_element:
> > check_action_definition: params:all <parameters multiplier="100"
> > dampen="2s" user="root" pidfile="/var/run/pingd.pid"/>
> > pengine[3284]: 2007/08/22_17:33:42 WARN: check_action_definition:
> > Parameters to pingd-child:1_monitor_4000 on ulxxcapb changed: recorded
> > e58195cc0db685dc80aa15525c617b8a vs. calculated (all)
> > 6ce8a133bf8292a7521af0d4610e60f2
> > pengine[3284]: 2007/08/22_17:33:42 info: log_data_element:
> > check_action_definition: params:all <parameters multiplier="100"
> > dampen="2s" user="root" pidfile="/var/run/pingd.pid"/>
> > pengine[3284]: 2007/08/22_17:33:42 WARN: check_action_definition:
> > Parameters to pingd-child:0_monitor_0 on ulxxcapa changed: recorded
> > 2500f62f8cbe28359717874cda643d0f vs. calculated (all)
> > 6ce8a133bf8292a7521af0d4610e60f2
> > pengine[3284]: 2007/08/22_17:33:42 info: log_data_element:
> > check_action_definition: params:all <parameters multiplier="100"
> > dampen="2s" user="root" pidfile="/var/run/pingd.pid"/>
> > pengine[3284]: 2007/08/22_17:33:42 WARN: check_action_definition:
> > Parameters to pingd-child:0_start_0 on ulxxcapa changed: recorded
> > e58195cc0db685dc80aa15525c617b8a vs. calculated (all)
> > 6ce8a133bf8292a7521af0d4610e60f2
> > pengine[3284]: 2007/08/22_17:33:42 info: log_data_element:
> > check_action_definition: params:all <parameters multiplier="100"
> > dampen="2s" user="root" pidfile="/var/run/pingd.pid"/>
> > pengine[3284]: 2007/08/22_17:33:42 WARN: check_action_definition:
> > Parameters to pingd-child:0_monitor_4000 on ulxxcapa changed: recorded
> > e58195cc0db685dc80aa15525c617b8a vs. calculated (all)
> > 6ce8a133bf8292a7521af0d4610e60f2
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfsp_export and nfsp_ip_2
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfsp_ip_2 and nfsp_ip_1
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfsp_ip_1 and nfsp_filesystem
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfsp_filesystem and nfsp_evms_failover
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > nfsp_evms_failover cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > nfsp_filesystem cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > nfsp_ip_1 cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > nfsp_ip_2 cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > nfsp_export cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfst_export and nfst_ip_2
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfst_ip_2 and nfst_ip_1
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfst_ip_1 and nfst_filesystem
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfst_filesystem and nfst_evms_failover
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > nfst_evms_failover cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > nfst_filesystem cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > nfst_ip_1 cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > nfst_ip_2 cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > nfst_export cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from oe10_tsmc and oe10_lsnr
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from oe10_lsnr and oe10_db
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from oe10_db and oe10_ip_2
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from oe10_ip_2 and oe10_ip_1
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from oe10_ip_1 and oe10_filesystem
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from oe10_filesystem and oe10_evms_failover
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > oe10_evms_failover cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > oe10_filesystem cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > oe10_ip_1 cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > oe10_ip_2 cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource oe10_db
> > cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > oe10_lsnr cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 WARN: native_color: Resource
> > oe10_tsmc cannot run anywhere
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfs2_tsmc and nfs2_export
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfs2_export and nfs2_ip_2
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfs2_ip_2 and nfs2_ip_1
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfs2_ip_1 and nfs2_filesystem
> > pengine[3284]: 2007/08/22_17:33:42 info: native_color: Combine scores
> > from nfs2_filesystem and nfs2_evms_failover
> > pengine[3284]: 2007/08/22_17:33:42 notice: NoRoleChange: Leave resource
> > nfs2_evms_failover      (ulxxcapb)
> > pengine[3284]: 2007/08/22_17:33:42 notice: NoRoleChange: Leave resource
> > nfs2_filesystem (ulxxcapb)
> > pengine[3284]: 2007/08/22_17:33:42 notice: NoRoleChange: Leave resource
> > nfs2_ip_1       (ulxxcapb)
> > pengine[3284]: 2007/08/22_17:33:42 notice: NoRoleChange: Leave resource
> > nfs2_ip_2       (ulxxcapb)
> > pengine[3284]: 2007/08/22_17:33:42 notice: NoRoleChange: Leave resource
> > nfs2_export     (ulxxcapb)
> > pengine[3284]: 2007/08/22_17:33:42 notice: NoRoleChange: Leave resource
> > nfs2_tsmc       (ulxxcapb)
> > pengine[3284]: 2007/08/22_17:33:42 notice: NoRoleChange: Restart
> > resource pingd-child:0 (ulxxcapa)
> > pengine[3284]: 2007/08/22_17:33:42 notice: RecurringOp: ulxxcapa
> > pingd-child:0_monitor_4000
> > pengine[3284]: 2007/08/22_17:33:42 notice: NoRoleChange: Restart
> > resource pingd-child:1 (ulxxcapb)
> > pengine[3284]: 2007/08/22_17:33:42 notice: RecurringOp: ulxxcapb
> > pingd-child:1_monitor_4000
> > crmd[3240]: 2007/08/22_17:33:42 info: do_state_transition: ulxxcapb:
> > State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [
> > input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=route_message ]
> > tengine[3283]: 2007/08/22_17:33:42 info: unpack_graph: Unpacked
> > transition 17: 10 actions in 10 synapses
> > tengine[3283]: 2007/08/22_17:33:42 info: te_pseudo_action: Pseudo action
> > 49 fired and confirmed
> > tengine[3283]: 2007/08/22_17:33:42 info: send_rsc_command: Initiating
> > action 45: pingd-child:0_stop_0 on ulxxcapa
> > tengine[3283]: 2007/08/22_17:33:42 info: send_rsc_command: Initiating
> > action 46: pingd-child:1_stop_0 on ulxxcapb
> > crmd[3240]: 2007/08/22_17:33:42 info: do_lrm_rsc_op: Performing
> > op=pingd-child:1_stop_0 key=46:17:8db77d5a-c345-4e9a-b711-97943abfd78e)
> > crmd[3240]: 2007/08/22_17:33:42 WARN: process_lrm_event: LRM operation
> > pingd-child:1_monitor_4000 (call=98, rc=-2) Cancelled
> > pengine[3284]: 2007/08/22_17:33:42 WARN: process_pe_message: Transition
> > 17: WARNINGs found during PE processing. PEngine Input stored in:
> > /var/lib/heartbeat/pengine/pe-warn-1355.bz2
> > pengine[3284]: 2007/08/22_17:33:42 info: process_pe_message:
> > Configuration WARNINGs found during PE processing.  Please run
> > "crm_verify -L" to identify issues.
> > crmd[3240]: 2007/08/22_17:33:42 info: process_lrm_event: LRM operation
> > pingd-child:1_stop_0 (call=111, rc=0) complete
> > cib[3236]: 2007/08/22_17:33:43 info: cib_diff_notify: Update (client:
> > 3240, call:141): 0.151.14644 -> 0.151.14645 (ok)
> > tengine[3283]: 2007/08/22_17:33:43 info: te_update_diff: Processing diff
> > (cib_update): 0.151.14644 -> 0.151.14645
> > tengine[3283]: 2007/08/22_17:33:43 info: match_graph_event: Action
> > pingd-child:1_stop_0 (46) confirmed on
> > acc23203-d2c8-419e-9dfb-ce621b332225
> > cib[11303]: 2007/08/22_17:33:43 info: write_cib_contents: Wrote version
> > 0.151.14645 of the CIB to disk (digest:
> > 03ecd35c34dbb001aa6b5734c025de04)
> > cib[3236]: 2007/08/22_17:33:43 info: cib_diff_notify: Update (client:
> > 16681, call:82): 0.151.14645 -> 0.151.14646 (ok)
> > tengine[3283]: 2007/08/22_17:33:43 info: te_update_diff: Processing diff
> > (cib_update): 0.151.14645 -> 0.151.14646
> > tengine[3283]: 2007/08/22_17:33:43 info: match_graph_event: Action
> > pingd-child:0_stop_0 (45) confirmed on
> > 4153c055-d562-46bb-8f33-41023f000ef9
> > tengine[3283]: 2007/08/22_17:33:43 info: te_pseudo_action: Pseudo action
> > 50 fired and confirmed
> > tengine[3283]: 2007/08/22_17:33:43 info: te_pseudo_action: Pseudo action
> > 47 fired and confirmed
> > tengine[3283]: 2007/08/22_17:33:43 info: send_rsc_command: Initiating
> > action 16: pingd-child:0_start_0 on ulxxcapa
> > tengine[3283]: 2007/08/22_17:33:43 info: send_rsc_command: Initiating
> > action 14: pingd-child:1_start_0 on ulxxcapb
> > crmd[3240]: 2007/08/22_17:33:43 info: do_lrm_rsc_op: Performing
> > op=pingd-child:1_start_0 key=14:17:8db77d5a-c345-4e9a-b711-97943abfd78e)
> > cib[11309]: 2007/08/22_17:33:43 info: write_cib_contents: Wrote version
> > 0.151.14646 of the CIB to disk (digest:
> > 4577ba9efc08b59b7d15c4d0287f37c4)
> > crmd[3240]: 2007/08/22_17:33:43 info: process_lrm_event: LRM operation
> > pingd-child:1_start_0 (call=112, rc=0) complete
> > cib[3236]: 2007/08/22_17:33:43 info: cib_diff_notify: Update (client:
> > 3240, call:142): 0.151.14646 -> 0.151.14647 (ok)
> > tengine[3283]: 2007/08/22_17:33:43 info: te_update_diff: Processing diff
> > (cib_update): 0.151.14646 -> 0.151.14647
> > tengine[3283]: 2007/08/22_17:33:43 info: match_graph_event: Action
> > pingd-child:1_start_0 (14) confirmed on
> > acc23203-d2c8-419e-9dfb-ce621b332225
> > tengine[3283]: 2007/08/22_17:33:43 info: send_rsc_command: Initiating
> > action 7: pingd-child:1_monitor_4000 on ulxxcapb
> > crmd[3240]: 2007/08/22_17:33:43 info: do_lrm_rsc_op: Performing
> > op=pingd-child:1_monitor_4000
> > key=7:17:8db77d5a-c345-4e9a-b711-97943abfd78e)
> > crmd[3240]: 2007/08/22_17:33:43 info: process_lrm_event: LRM operation
> > pingd-child:1_monitor_4000 (call=113, rc=0) complete
> > cib[3236]: 2007/08/22_17:33:44 info: cib_diff_notify: Update (client:
> > 3240, call:143): 0.151.14647 -> 0.151.14648 (ok)
> > tengine[3283]: 2007/08/22_17:33:44 info: te_update_diff: Processing diff
> > (cib_update): 0.151.14647 -> 0.151.14648
> > tengine[3283]: 2007/08/22_17:33:44 info: match_graph_event: Action
> > pingd-child:1_monitor_4000 (7) confirmed on
> > acc23203-d2c8-419e-9dfb-ce621b332225
> > cib[11332]: 2007/08/22_17:33:44 info: write_cib_contents: Wrote version
> > 0.151.14647 of the CIB to disk (digest:
> > 4ef8ac941bf8f010b1af432b52f5c800)
> > cib[11337]: 2007/08/22_17:33:44 info: write_cib_contents: Wrote version
> > 0.151.14648 of the CIB to disk (digest:
> > 0211132390c452995fffb4865a233d42)
> > cib[3236]: 2007/08/22_17:33:44 info: cib_diff_notify: Update (client:
> > 16681, call:83): 0.151.14648 -> 0.151.14649 (ok)
> > tengine[3283]: 2007/08/22_17:33:44 info: te_update_diff: Processing diff
> > (cib_update): 0.151.14648 -> 0.151.14649
> > tengine[3283]: 2007/08/22_17:33:44 info: match_graph_event: Action
> > pingd-child:0_start_0 (16) confirmed on
> > 4153c055-d562-46bb-8f33-41023f000ef9
> > tengine[3283]: 2007/08/22_17:33:44 info: send_rsc_command: Initiating
> > action 8: pingd-child:0_monitor_4000 on ulxxcapa
> > tengine[3283]: 2007/08/22_17:33:44 info: te_pseudo_action: Pseudo action
> > 48 fired and confirmed
> > cib[11338]: 2007/08/22_17:33:44 info: write_cib_contents: Wrote version
> > 0.151.14649 of the CIB to disk (digest:
> > fa532c5822e4a7847ef22b869f638faa)
> > cib[3236]: 2007/08/22_17:33:45 info: cib_diff_notify: Update (client:
> > 16681, call:84): 0.151.14649 -> 0.151.14650 (ok)
> > crmd[3240]: 2007/08/22_17:33:45 info: do_state_transition: ulxxcapb:
> > State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
> > cause=C_IPC_MESSAGE origin=route_message ]
> > tengine[3283]: 2007/08/22_17:33:45 info: te_update_diff: Processing diff
> > (cib_update): 0.151.14649 -> 0.151.14650
> > tengine[3283]: 2007/08/22_17:33:45 info: match_graph_event: Action
> > pingd-child:0_monitor_4000 (8) confirmed on
> > 4153c055-d562-46bb-8f33-41023f000ef9
> > tengine[3283]: 2007/08/22_17:33:45 info: run_graph: Transition 17:
> > (Complete=10, Pending=0, Fired=0, Skipped=0, Incomplete=0)
> > tengine[3283]: 2007/08/22_17:33:45 info: notify_crmd: Transition 17
> > status: te_complete - <null>
> > cib[11339]: 2007/08/22_17:33:45 info: write_cib_contents: Wrote version
> > 0.151.14650 of the CIB to disk (digest:
> > 6b1ee544423b64236748810e98e83ab6)
> > heartbeat[3192]: 2007/08/22_17:33:49 WARN: 1 lost packet(s) for
> > [ulxxcapa] [1841:1843]
> > heartbeat[3192]: 2007/08/22_17:33:49 info: No pkts missing from
> > ulxxcapa!
> > ccm[3235]: 2007/08/22_17:33:56 info: client (pid=11467) removed from ccm
> >
> >
> > Kind regards,
> > Daniel
> >
> >
> > Diese Nachricht ist ausschliesslich f?r den Adressaten bestimmt und 
> > beinhaltet unter Umst?nden vertrauliche Mitteilungen. Da die 
> > Vertraulichkeit von e-Mail-Nachrichten nicht gew?hrleistet werden kann, 
> > ?bernehmen wir keine Haftung f?r die Gew?hrung der Vertraulichkeit und 
> > Unversehrtheit dieser Mitteilung. Bei irrt?mlicher Zustellung bitten wir 
> > Sie um Benachrichtigung per e-Mail und um L?schung dieser Nachricht sowie 
> > eventueller Anh?nge. Jegliche unberechtigte Verwendung oder Verbreitung 
> > dieser Informationen ist streng verboten.
> >
> > This message is intended only for the named recipient and may contain 
> > confidential or privileged information. As the confidentiality of email 
> > communication cannot be guaranteed, we do not accept any responsibility for 
> > the confidentiality and the intactness of this message. If you have 
> > received it in error, please advise the sender by return e-mail and delete 
> > this message and any attachments. Any unauthorised use or dissemination of 
> > this information is strictly prohibited.
> > _______________________________________________
> > Linux-HA mailing list
> > [email protected]
> > http://lists.linux-ha.org/mailman/listinfo/linux-ha
> > See also: http://linux-ha.org/ReportingProblems
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to