Thanks once again Marian. I appreciate your help and patience.
I followed your links and used the instructions to create a new
cib.xml. Unfortunately, the same problem exists. When I issue service
heartbeat stop, the ip address does not get released from the node. Here
is an output from crm_mon after I issued service heartbeat stop on the
"main" node.
============
Last updated: Mon Mar 22 09:06:25 2010
Stack: Heartbeat
Current DC: dbsuat1b.intranet.mydomain.com
(e99889ee-da15-4b09-bfc7-641e3ac0687f) - partition with quorum
Version: 1.0.6-17fe0022afda074a937d934b3eb625eccd1f01ef
2 Nodes configured, unknown expected votes
2 Resources configured.
============
Online: [ dbsuat1b.intranet.mydomanin.com ]
OFFLINE: [ dbsuat1a.intranet.mydomain.com ]
IPaddr_172_28_185_49 (ocf::heartbeat:IPaddr): Started
dbsuat1b.intranet.mydomain.com (unmanaged) FAILED
failover-ip (ocf::heartbeat:IPaddr): Started
dbsuat1b.intranet.mydomain.com
Failed actions:
IPaddr_172_28_185_49_start_0 (node=dbsuat1b.intranet.mydomain.com,
call=4, rc=1, status=complete): unknown error
IPaddr_172_28_185_49_stop_0 (node=dbsuat1b.intranet.mydomain.com,
call=6, rc=-2, status=Timed Out): unknown exec error
And here is the new cib.xml - am I missing anything?:
<cib crm_feature_set="3.0.1" admin_epoch="0" epoch="13" have-quorum="1"
remote-tls-port="0" validate-with="pacemaker-1.0"
dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f" num_updates="0"
cib-last-written="Mon Mar 22 08:51:05 2010">
<configuration>
<crm_config>
<cluster_property_set id="cib-bootstrap-options">
<nvpair id="nvpair.id17898566" name="symmetric-cluster"
value="true"/>
<nvpair id="nvpair.id17898575" name="no-quorum-policy"
value="stop"/>
<nvpair id="nvpair.id17896945"
name="default-resource-stickiness" value="0"/>
<nvpair id="nvpair.id17896954"
name="default-resource-failure-stickiness" value="0"/>
<nvpair id="nvpair.id17896963" name="stonith-enabled"
value="false"/>
<nvpair id="nvpair.id17898428" name="stonith-action"
value="reboot"/>
<nvpair id="nvpair.id17898436" name="startup-fencing" value="true"/>
<nvpair id="nvpair.id17898445" name="stop-orphan-resources"
value="true"/>
<nvpair id="nvpair.id17898455" name="stop-orphan-actions"
value="true"/>
<nvpair id="nvpair.id17898464" name="remove-after-stop"
value="false"/>
<nvpair id="nvpair.id17898473" name="short-resource-names"
value="true"/>
<nvpair id="nvpair.id17897927" name="transition-idle-timeout"
value="5min"/>
<nvpair id="nvpair.id17897936" name="default-action-timeout"
value="20s"/>
<nvpair id="nvpair.id17897946" name="is-managed-default"
value="true"/>
<nvpair id="nvpair.id17897955" name="cluster-delay" value="60s"/>
<nvpair id="nvpair.id17897964" name="pe-error-series-max"
value="-1"/>
<nvpair id="nvpair.id17897973" name="pe-warn-series-max"
value="-1"/>
<nvpair id="nvpair.id17897982" name="pe-input-series-max"
value="-1"/>
<nvpair id="nvpair.id17897383" name="dc-version"
value="1.0.6-17fe0022afda074a937d934b3eb625eccd1f01ef"/>
<nvpair id="nvpair.id17897392" name="cluster-infrastructure"
value="Heartbeat"/>
</cluster_property_set>
</crm_config>
<nodes>
<node id="db80324b-c9de-4995-a66a-eedf93abb42c"
uname="dbsuat1a.intranet.mydomain.com" type="normal"/>
<node id="e99889ee-da15-4b09-bfc7-641e3ac0687f"
uname="dbsuat1b.intranet.mydomain.com" type="normal"/>
</nodes>
<resources>
<primitive id="IPaddr_172_28_185_49" class="ocf"
provider="heartbeat" type="IPaddr">
<meta_attributes id="primitive-IPaddr_172_28_185_49.meta"/>
<operations>
<op name="monitor" interval="5s" id="IPaddr_172_28_185_49_mon"
timeout="5s"/>
</operations>
<instance_attributes id="instance_attributes.id17897446">
<nvpair id="nvpair.id17897452" name="ip" value="172.28.185.49"/>
</instance_attributes>
</primitive>
<primitive class="ocf" id="failover-ip" provider="heartbeat"
type="IPaddr">
<instance_attributes id="failover-ip-instance_attributes">
<nvpair id="failover-ip-instance_attributes-ip" name="ip"
value="172.28.185.49"/>
</instance_attributes>
<operations>
<op id="failover-ip-monitor-10s" interval="10s" name="monitor"/>
</operations>
</primitive>
</resources>
<constraints>
<rsc_location id="rsc_location_IPaddr_172_28_185_49"
rsc="IPaddr_172_28_185_49">
<rule id="prefered_location_IPaddr_172_28_185_49" score="100">
<expression attribute="#uname" id="expression.id17897605"
operation="eq" value="DBSUAT1A.intranet.mydomain.com"/>
</rule>
</rsc_location>
</constraints>
<rsc_defaults/>
<op_defaults/>
</configuration>
</cib>
Marian Marinov wrote:
> Hello mike,
>
> Your problem is pretty simple. You simply don't have configured IPaddr
> resource
> within pacemaker.
>
> Please look here for IPaddr:
> http://www.linux-ha.org/doc/re-ra-IPaddr.html
> And here for IPaddr2:
> http://www.linux-ha.org/doc/re-ra-IPaddr2.html
>
> You should remove the haresources file and configure pacemaker so it can
> handle
> your IP address.
>
> You can find basic configurations here:
> http://clusterlabs.org/wiki/Example_configurations
> http://clusterlabs.org/wiki/Example_XML_configurations
>
> Best regards,
> Marian
>
> On Monday 22 March 2010 01:25:58 mike wrote:
>
>> Hi Marian - I have included my cib.xml file below.
>> What I have found tonight is that by commenting out the crm entry in the
>> ha.cf and re-enabling the haresources file, I am able to fail the ip
>> back and forth at will. Here is what my haresources file looks like:
>> *DBSUAT1A.intranet.mydomain.com IPaddr::172.28.185.49*
>>
>> My cib.xml file which was generated from the above haresources file
>> using /usr/lib64/heartbeat/haresources2cib.py
>> [r...@dbsuat1b support]# cat cib.xml
>> <cib admin_epoch="0" epoch="7" validate-with="transitional-0.6"
>> crm_feature_set="3.0.1" have-quorum="1"
>> dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f" num_updates="0"
>> cib-last-written="Sun Mar 21 19:10:03 2010">
>> <configuration>
>> <crm_config>
>> <cluster_property_set id="cib-bootstrap-options">
>> <attributes>
>> <nvpair id="cib-bootstrap-options-symmetric-cluster"
>> name="symmetric-cluster" value="true"/>
>> <nvpair id="cib-bootstrap-options-no-quorum-policy"
>> name="no-quorum-policy" value="stop"/>
>> <nvpair id="cib-bootstrap-options-default-resource-stickiness"
>> name="default-resource-stickiness" value="0"/>
>> <nvpair
>> id="cib-bootstrap-options-default-resource-failure-stickiness"
>> name="default-resource-failure-stickiness" value="0"/>
>> <nvpair id="cib-bootstrap-options-stonith-enabled"
>> name="stonith-enabled" value="false"/>
>> <nvpair id="cib-bootstrap-options-stonith-action"
>> name="stonith-action" value="reboot"/>
>> <nvpair id="cib-bootstrap-options-startup-fencing"
>> name="startup-fencing" value="true"/>
>> <nvpair id="cib-bootstrap-options-stop-orphan-resources"
>> name="stop-orphan-resources" value="true"/>
>> <nvpair id="cib-bootstrap-options-stop-orphan-actions"
>> name="stop-orphan-actions" value="true"/>
>> <nvpair id="cib-bootstrap-options-remove-after-stop"
>> name="remove-after-stop" value="false"/>
>> <nvpair id="cib-bootstrap-options-short-resource-names"
>> name="short-resource-names" value="true"/>
>> <nvpair id="cib-bootstrap-options-transition-idle-timeout"
>> name="transition-idle-timeout" value="5min"/>
>> <nvpair id="cib-bootstrap-options-default-action-timeout"
>> name="default-action-timeout" value="20s"/>
>> <nvpair id="cib-bootstrap-options-is-managed-default"
>> name="is-managed-default" value="true"/>
>> <nvpair id="cib-bootstrap-options-cluster-delay"
>> name="cluster-delay" value="60s"/>
>> <nvpair id="cib-bootstrap-options-pe-error-series-max"
>> name="pe-error-series-max" value="-1"/>
>> <nvpair id="cib-bootstrap-options-pe-warn-series-max"
>> name="pe-warn-series-max" value="-1"/>
>> <nvpair id="cib-bootstrap-options-pe-input-series-max"
>> name="pe-input-series-max" value="-1"/>
>> <nvpair id="cib-bootstrap-options-dc-version"
>> name="dc-version" value="1.0.6-17fe0022afda074a937d934b3eb625eccd1f01ef"/>
>> <nvpair id="cib-bootstrap-options-cluster-infrastructure"
>> name="cluster-infrastructure" value="Heartbeat"/>
>> </attributes>
>> </cluster_property_set>
>> </crm_config>
>> <nodes>
>> <node id="db80324b-c9de-4995-a66a-eedf93abb42c"
>> uname="dbsuat1a.intranet.mydomain.com" type="normal"/>
>> <node id="e99889ee-da15-4b09-bfc7-641e3ac0687f"
>> uname="dbsuat1b.intranet.mydomain.com" type="normal"/>
>> </nodes>
>> <resources>
>> <primitive class="ocf" id="IPaddr_172_28_185_49"
>> provider="heartbeat" type="IPaddr">
>> <operations>
>> <op id="IPaddr_172_28_185_49_mon" interval="5s" name="monitor"
>> timeout="5s"/>
>> </operations>
>> <instance_attributes id="IPaddr_172_28_185_49_inst_attr">
>> <attributes>
>> <nvpair id="IPaddr_172_28_185_49_attr_0" name="ip"
>> value="172.28.185.49"/>
>> </attributes>
>> </instance_attributes>
>> </primitive>
>> </resources>
>> <constraints>
>> <rsc_location id="rsc_location_IPaddr_172_28_185_49"
>> rsc="IPaddr_172_28_185_49">
>> <rule id="prefered_location_IPaddr_172_28_185_49" score="100">
>> <expression attribute="#uname"
>> id="prefered_location_IPaddr_172_28_185_49_expr" operation="eq"
>> value="DBSUAT1A.intranet.mydomain.com"/>
>> </rule>
>> </rsc_location>
>> </constraints>
>> </configuration>
>> </cib>
>>
>> Marian Marinov wrote:
>>
>>> Can you please give us your crm configuration ?
>>>
>>> Marian
>>>
>>> On Sunday 21 March 2010 23:30:46 mike wrote:
>>>
>>>> Thank you Marian. I removed th efile as you suggested but unfortunately
>>>> it has made no difference. The ip address is simply not being released
>>>> when I stop the heartbeat process.
>>>>
>>>> Anyone have an ideas where I could start to look at this? The only way I
>>>> can get the ip address released is to reboot the node.
>>>>
>>>> thanks
>>>>
>>>> Marian Marinov wrote:
>>>>
>>>>> On Saturday 20 March 2010 03:56:27 mike wrote:
>>>>>
>>>>>> Hi guys,
>>>>>>
>>>>>> I have a simple 2 node cluster with a VIP running on RHEL 5.3 on s390.
>>>>>> Nothing else configured yet.
>>>>>>
>>>>>> When I start up the cluster, all is well. The VIP starts up on the
>>>>>> home node and crm_mon shows the resource and nodes as on line. No
>>>>>> errors in the logs.
>>>>>>
>>>>>> If I issue service heartbeat stop on the main node, the ip fails over
>>>>>> to the back up node and crm_mon shows as I would expect it should,
>>>>>> i.e. the ip address is on the back up node and that the other node is
>>>>>> offline. However, if I do a ifconfig on the main node I see that the
>>>>>> eth0:0 entry is still there so in effect the vip address is now on
>>>>>> both servers.
>>>>>>
>>>>>> If both nodes were up and running and I rebooted the main node then
>>>>>> the failover works perfectly.
>>>>>>
>>>>>> Would anyone know why the nodes seem unable to release the vip unless
>>>>>> rebooted?
>>>>>>
>>>>>> ha-log:
>>>>>> Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> stage6: Scheduling Node dbsuat1a.intranet.mydomain.com for shutdown
>>>>>> Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com pengine: [4714]:
>>>>>> notice: LogActions: Move resource IPaddr_172_28_185_49 (Started
>>>>>> dbsuat1a.intranet.mydomain.com -> dbsuat1b.intranet.mydomain.com)
>>>>>> Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_state_transition: State transition S_POLICY_ENGINE ->
>>>>>> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE
>>>>>> origin=handle_response ]
>>>>>> Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> process_pe_message: Transition 5: PEngine Input stored in:
>>>>>> /usr/var/lib/pengine/pe-input-337.bz2
>>>>>> Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> process_pe_message: Configuration WARNINGs found during PE processing.
>>>>>> Please run "crm_verify -L" to identify issues.
>>>>>> Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> unpack_graph: Unpacked transition 5: 5 actions in 5 synapses
>>>>>> Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_te_invoke: Processing graph 5 (ref=pe_calc-dc-1269021312-26)
>>>>>> derived from /usr/var/lib/pengine/pe-input-337.bz2
>>>>>> Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> te_rsc_command: Initiating action 6: stop IPaddr_172_28_185_49_stop_0
>>>>>> on dbsuat1a.intranet.mydomain.com (local)
>>>>>> Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_lrm_rsc_op: Performing
>>>>>> key=6:5:0:888fa84e-3267-409e-966b-2ab01e579c0f
>>>>>> op=IPaddr_172_28_185_49_stop_0 )
>>>>>> Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com lrmd: [4528]: info:
>>>>>> rsc:IPaddr_172_28_185_49:5: stop
>>>>>> Mar 19 13:55:12 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> process_lrm_event: LRM operation IPaddr_172_28_185_49_monitor_5000
>>>>>> (call=4, status=1, cib-update=0, confirmed=true) Cancelled
>>>>>> Mar 19 13:55:32 DBSUAT1A.intranet.mydomain.com lrmd: [4528]: WARN:
>>>>>> IPaddr_172_28_185_49:stop process (PID 5474) timed out (try 1).
>>>>>> Killing with signal SIGTERM (15).
>>>>>> Mar 19 13:55:32 DBSUAT1A.intranet.mydomain.com lrmd: [4528]: WARN:
>>>>>> Managed IPaddr_172_28_185_49:stop process 5474 killed by signal 15
>>>>>> [SIGTERM - Termination (ANSI)].
>>>>>> Mar 19 13:55:32 DBSUAT1A.intranet.mydomain.com lrmd: [4528]: WARN:
>>>>>> operation stop[5] on ocf::IPaddr::IPaddr_172_28_185_49 for client
>>>>>> 4531, its parameters: ip=[172.28.185.49] CRM_meta_timeout=[20000]
>>>>>> crm_feature_set=[3.0.1] : pid [5474] timed out
>>>>>> Mar 19 13:55:32 DBSUAT1A.intranet.mydomain.com crmd: [4531]: ERROR:
>>>>>> process_lrm_event: LRM operation IPaddr_172_28_185_49_stop_0 (5) Timed
>>>>>> Out (timeout=20000ms)
>>>>>> Mar 19 13:55:32 DBSUAT1A.intranet.mydomain.com crmd: [4531]: WARN:
>>>>>> status_from_rc: Action 6 (IPaddr_172_28_185_49_stop_0) on
>>>>>> dbsuat1a.intranet.mydomain.com failed (target: 0 vs. rc: -2): Error
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: WARN:
>>>>>> update_failcount: Updating failcount for IPaddr_172_28_185_49 on
>>>>>> dbsuat1a.intranet.mydomain.com after failed stop: rc=-2
>>>>>> (update=INFINITY, time=1269021333)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> abort_transition_graph: match_graph_event:272 - Triggered transition
>>>>>> abort (complete=0, tag=lrm_rsc_op, id=IPaddr_172_28_185_49_stop_0,
>>>>>> magic=2:-2;6:5:0:888fa84e-3267-409e-966b-2ab01e579c0f, cib=0.23.16) :
>>>>>> Event failed
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> update_abort_priority: Abort priority upgraded from 0 to 1
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> update_abort_priority: Abort action done superceeded by restart
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> match_graph_event: Action IPaddr_172_28_185_49_stop_0 (6) confirmed on
>>>>>> dbsuat1a.intranet.mydomain.com (rc=4)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> run_graph: ====================================================
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: notice:
>>>>>> run_graph: Transition 5 (Complete=1, Pending=0, Fired=0, Skipped=4,
>>>>>> Incomplete=0, Source=/usr/var/lib/pengine/pe-input-337.bz2): Stopped
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> te_graph_trigger: Transition 5 is now complete
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info:
>>>>>> find_hash_entry: Creating hash entry for
>>>>>> fail-count-IPaddr_172_28_185_49 Mar 19 13:55:33
>>>>>> DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_state_transition: State transition S_TRANSITION_ENGINE ->
>>>>>> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL
>>>>>> origin=notify_crmd ] Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com
>>>>>> crmd: [4531]: info: do_state_transition: All 2 cluster nodes are
>>>>>> eligible to run resources. Mar 19 13:55:33
>>>>>> DBSUAT1A.intranet.mydomain.com attrd: [4530]: info:
>>>>>> attrd_trigger_update: Sending flush op to all hosts for:
>>>>>> fail-count-IPaddr_172_28_185_49 (INFINITY)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info:
>>>>>> attrd_perform_update: Sent update 24:
>>>>>> fail-count-IPaddr_172_28_185_49=INFINITY
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_pe_invoke: Query 53: Requesting the current CIB: S_POLICY_ENGINE
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info:
>>>>>> find_hash_entry: Creating hash entry for
>>>>>> last-failure-IPaddr_172_28_185_49 Mar 19 13:55:33
>>>>>> DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> abort_transition_graph: te_update_diff:146 - Triggered transition
>>>>>> abort (complete=1, tag=transient_attributes,
>>>>>> id=db80324b-c9de-4995-a66a-eedf93abb42c, magic=NA, cib=0.23.17) :
>>>>>> Transient attribute: update
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info:
>>>>>> attrd_trigger_update: Sending flush op to all hosts for:
>>>>>> last-failure-IPaddr_172_28_185_49 (1269021333)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info:
>>>>>> attrd_perform_update: Sent update 27:
>>>>>> last-failure-IPaddr_172_28_185_49=1269021333
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_pe_invoke_callback: Invoking the PE: ref=pe_calc-dc-1269021333-28,
>>>>>> seq=2, quorate=1
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> abort_transition_graph: te_update_diff:146 - Triggered transition
>>>>>> abort (complete=1, tag=transient_attributes,
>>>>>> id=db80324b-c9de-4995-a66a-eedf93abb42c, magic=NA, cib=0.23.18) :
>>>>>> Transient attribute: update
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]:
>>>>>> notice: update_validation: Upgrading transitional-0.6-style
>>>>>> configuration to pacemaker-1.0 with /usr/share/pacemaker/upgrade06.xsl
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> update_validation: Transformation /usr/share/pacemaker/upgrade06.xsl
>>>>>> successful
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]:
>>>>>> notice: update_validation: Upgraded from transitional-0.6 to
>>>>>> pacemaker-1.0 validation
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN:
>>>>>> cli_config_update: Your configuration was internally updated to the
>>>>>> latest version (pacemaker-1.0)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_pe_invoke: Query 54: Requesting the current CIB: S_POLICY_ENGINE
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_pe_invoke: Query 55: Requesting the current CIB: S_POLICY_ENGINE
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_pe_invoke_callback: Invoking the PE: ref=pe_calc-dc-1269021333-29,
>>>>>> seq=2, quorate=1
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' =
>>>>>> 0 Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]:
>>>>>> info: determine_online_status: Node dbsuat1a.intranet.mydomain.com is
>>>>>> shutting down
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN:
>>>>>> unpack_rsc_op: Processing failed op IPaddr_172_28_185_49_stop_0 on
>>>>>> dbsuat1a.intranet.mydomain.com: unknown exec error (-2)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> native_add_running: resource IPaddr_172_28_185_49 isnt managed
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> determine_online_status: Node dbsuat1b.intranet.mydomain.com is online
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]:
>>>>>> notice: native_print: IPaddr_172_28_185_49 (ocf::heartbeat:IPaddr):
>>>>>> Started dbsuat1a.intranet.mydomain.com (unmanaged) FAILED
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> get_failcount: IPaddr_172_28_185_49 has failed 1000000 times on
>>>>>> dbsuat1a.intranet.mydomain.com
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN:
>>>>>> common_apply_stickiness: Forcing IPaddr_172_28_185_49 away from
>>>>>> dbsuat1a.intranet.mydomain.com after 1000000 failures (max=1000000)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> native_color: Unmanaged resource IPaddr_172_28_185_49 allocated to
>>>>>> 'nowhere': failed
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> stage6: Scheduling Node dbsuat1a.intranet.mydomain.com for shutdown
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]:
>>>>>> notice: LogActions: Leave resource IPaddr_172_28_185_49 (Started
>>>>>> unmanaged) Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd:
>>>>>> [4531]: info: handle_response: pe_calc calculation
>>>>>> pe_calc-dc-1269021333-28 is obsolete Mar 19 13:55:33
>>>>>> DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> process_pe_message: Transition 6: PEngine Input stored in:
>>>>>> /usr/var/lib/pengine/pe-input-338.bz2
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> process_pe_message: Configuration WARNINGs found during PE processing.
>>>>>> Please run "crm_verify -L" to identify issues.
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]:
>>>>>> notice: update_validation: Upgrading transitional-0.6-style
>>>>>> configuration to pacemaker-1.0 with /usr/share/pacemaker/upgrade06.xsl
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> update_validation: Transformation /usr/share/pacemaker/upgrade06.xsl
>>>>>> successful
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]:
>>>>>> notice: update_validation: Upgraded from transitional-0.6 to
>>>>>> pacemaker-1.0 validation
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN:
>>>>>> cli_config_update: Your configuration was internally updated to the
>>>>>> latest version (pacemaker-1.0)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' =
>>>>>> 0 Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]:
>>>>>> info: determine_online_status: Node dbsuat1a.intranet.mydomain.com is
>>>>>> shutting down
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN:
>>>>>> unpack_rsc_op: Processing failed op IPaddr_172_28_185_49_stop_0 on
>>>>>> dbsuat1a.intranet.mydomain.com: unknown exec error (-2)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> native_add_running: resource IPaddr_172_28_185_49 isnt managed
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> determine_online_status: Node dbsuat1b.intranet.mydomain.com is online
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]:
>>>>>> notice: native_print: IPaddr_172_28_185_49 (ocf::heartbeat:IPaddr):
>>>>>> Started dbsuat1a.intranet.mydomain.com (unmanaged) FAILED
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> get_failcount: IPaddr_172_28_185_49 has failed 1000000 times on
>>>>>> dbsuat1a.intranet.mydomain.com
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: WARN:
>>>>>> common_apply_stickiness: Forcing IPaddr_172_28_185_49 away from
>>>>>> dbsuat1a.intranet.mydomain.com after 1000000 failures (max=1000000)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> native_color: Unmanaged resource IPaddr_172_28_185_49 allocated to
>>>>>> 'nowhere': failed
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> stage6: Scheduling Node dbsuat1a.intranet.mydomain.com for shutdown
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]:
>>>>>> notice: LogActions: Leave resource IPaddr_172_28_185_49 (Started
>>>>>> unmanaged) Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine:
>>>>>> [4714]: info: process_pe_message: Transition 7: PEngine Input stored
>>>>>> in:
>>>>>> /usr/var/lib/pengine/pe-input-339.bz2
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_state_transition: State transition S_POLICY_ENGINE ->
>>>>>> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE
>>>>>> origin=handle_response ]
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> process_pe_message: Configuration WARNINGs found during PE processing.
>>>>>> Please run "crm_verify -L" to identify issues.
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> unpack_graph: Unpacked transition 7: 1 actions in 1 synapses
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_te_invoke: Processing graph 7 (ref=pe_calc-dc-1269021333-29)
>>>>>> derived from /usr/var/lib/pengine/pe-input-339.bz2
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> te_crm_command: Executing crm-event (10): do_shutdown on
>>>>>> dbsuat1a.intranet.mydomain.com
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> te_crm_command: crm-event (10) is a local shutdown
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> run_graph: ====================================================
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: notice:
>>>>>> run_graph: Transition 7 (Complete=1, Pending=0, Fired=0, Skipped=0,
>>>>>> Incomplete=0, Source=/usr/var/lib/pengine/pe-input-339.bz2): Complete
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> te_graph_trigger: Transition 7 is now complete
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_state_transition: State transition S_TRANSITION_ENGINE ->
>>>>>> S_STOPPING [ input=I_STOP cause=C_FSA_INTERNAL origin=notify_crmd ]
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_dc_release: DC role released
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> stop_subsystem: Sent -TERM to pengine: [4714]
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_te_control: Transitioner is now inactive
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com pengine: [4714]: info:
>>>>>> crm_signal_dispatch: Invoking handler for signal 15: Terminated
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_te_control: Disconnecting STONITH...
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> tengine_stonith_connection_destroy: Fencing daemon disconnected
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: notice:
>>>>>> Not currently connected.
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_shutdown: Terminating the pengine
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> stop_subsystem: Sent -TERM to pengine: [4714]
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_shutdown: Waiting for subsystems to exit
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: WARN:
>>>>>> register_fsa_input_adv: do_shutdown stalled the FSA with pending
>>>>>> inputs Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]:
>>>>>> info: do_shutdown: All subsystems stopped, continuing
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: WARN:
>>>>>> do_log: FSA: Input I_RELEASE_SUCCESS from do_dc_release() received in
>>>>>> state S_STOPPING
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_shutdown: Terminating the pengine
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> stop_subsystem: Sent -TERM to pengine: [4714]
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_shutdown: Waiting for subsystems to exit
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_shutdown: All subsystems stopped, continuing
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> crmdManagedChildDied: Process pengine:[4714] exited (signal=0,
>>>>>> exitcode=0) Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd:
>>>>>> [4531]: info: pe_msg_dispatch: Received HUP from pengine:[4714]
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> pe_connection_destroy: Connection to the Policy Engine released
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_shutdown: All subsystems stopped, continuing
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: ERROR:
>>>>>> verify_stopped: Resource IPaddr_172_28_185_49 was active at shutdown.
>>>>>> You may ignore this error if it is unmanaged.
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_lrm_control: Disconnected from the LRM
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com ccm: [4526]: info:
>>>>>> client (pid=4531) removed from ccm
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_ha_control: Disconnected from Heartbeat
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_cib_control: Disconnecting CIB
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com cib: [4527]: info:
>>>>>> cib_process_readwrite: We are now in R/O mode
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> crmd_cib_connection_destroy: Connection to the CIB terminated...
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_exit: Performing A_EXIT_0 - gracefully exiting the CRMd
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> free_mem: Dropping I_TERMINATE: [ state=S_STOPPING
>>>>>> cause=C_FSA_INTERNAL origin=do_stop ]
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com crmd: [4531]: info:
>>>>>> do_exit: [crmd] stopped (0)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: killing /usr/lib64/heartbeat/attrd process group 4530 with
>>>>>> signal 15 Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd:
>>>>>> [4530]: info: crm_signal_dispatch: Invoking handler for signal 15:
>>>>>> Terminated Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd:
>>>>>> [4530]: info: attrd_shutdown: Exiting
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info:
>>>>>> main: Exiting...
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com attrd: [4530]: info:
>>>>>> attrd_cib_connection_destroy: Connection to the CIB terminated...
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: killing /usr/lib64/heartbeat/stonithd process group 4529 with
>>>>>> signal 15 Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com stonithd:
>>>>>> [4529]: notice: /usr/lib64/heartbeat/stonithd normally quit.
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: killing /usr/lib64/heartbeat/lrmd -r process group 4528 with
>>>>>> signal 15 Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com lrmd: [4528]:
>>>>>> info: lrmd is shutting down
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com lrmd: [4528]: WARN:
>>>>>> resource IPaddr_172_28_185_49 is left in UNKNOWN status.(last op stop
>>>>>> finished without LRM_OP_DONE status.)
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: killing /usr/lib64/heartbeat/cib process group 4527 with signal
>>>>>> 15 Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com cib: [4527]: info:
>>>>>> crm_signal_dispatch: Invoking handler for signal 15: Terminated Mar 19
>>>>>> 13:55:33 DBSUAT1A.intranet.mydomain.com cib: [4527]: info:
>>>>>> cib_shutdown: Disconnected 0 clients
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com cib: [4527]: info:
>>>>>> cib_process_disconnect: All clients disconnected...
>>>>>> Mar 19 13:55:33 DBSUAT1A.intranet.mydomain.com cib: [4527]: info:
>>>>>> initiate_exit: Sending disconnect notification to 2 peers...
>>>>>> Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com cib: [4527]: info:
>>>>>> cib_process_shutdown_req: Shutdown ACK from
>>>>>> dbsuat1b.intranet.mydomain.com Mar 19 13:55:34
>>>>>> DBSUAT1A.intranet.mydomain.com cib: [4527]: info: terminate_cib:
>>>>>> cib_process_shutdown_req: Disconnecting heartbeat Mar 19 13:55:34
>>>>>> DBSUAT1A.intranet.mydomain.com cib: [4527]: info: terminate_cib:
>>>>>> Exiting...
>>>>>> Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com cib: [4527]: info:
>>>>>> cib_process_request: Operation complete: op cib_shutdown_req for
>>>>>> section 'all'
>>>>>> (origin=dbsuat1b.intranet.mydomain.com/dbsuat1b.intranet.mydomain.com/
>>>>>> (n ull ), version=0.0.0): ok (rc=0)
>>>>>> Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com cib: [4527]: info:
>>>>>> ha_msg_dispatch: Lost connection to heartbeat service.
>>>>>> Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com cib: [4527]: info:
>>>>>> main: Done
>>>>>> Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com ccm: [4526]: info:
>>>>>> client (pid=4527) removed from ccm
>>>>>> Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: killing /usr/lib64/heartbeat/ccm process group 4526 with signal
>>>>>> 15 Mar 19 13:55:34 DBSUAT1A.intranet.mydomain.com ccm: [4526]: info:
>>>>>> received SIGTERM, going to shut down
>>>>>> Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: killing HBFIFO process 4522 with signal 15
>>>>>> Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: killing HBWRITE process 4523 with signal 15
>>>>>> Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: killing HBREAD process 4524 with signal 15
>>>>>> Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: Core process 4524 exited. 3 remaining
>>>>>> Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: Core process 4523 exited. 2 remaining
>>>>>> Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: Core process 4522 exited. 1 remaining
>>>>>> Mar 19 13:55:35 DBSUAT1A.intranet.mydomain.com heartbeat: [4519]:
>>>>>> info: dbsuat1a.intranet.mydomain.com Heartbeat shutdown complete.
>>>>>>
>>>>>>
>>>>>> my ha.cf
>>>>>> # Logging
>>>>>> debug 1
>>>>>> debugfile /var/log/ha-debug
>>>>>> logfile /var/log/ha-log
>>>>>> logfacility local0
>>>>>> #use_logd true
>>>>>> #logfacility daemon
>>>>>>
>>>>>> # Misc Options
>>>>>> traditional_compression off
>>>>>> compression bz2
>>>>>> coredumps true
>>>>>>
>>>>>> # Communications
>>>>>> udpport 691
>>>>>> bcast eth0
>>>>>> ##autojoin any
>>>>>> autojoin none
>>>>>>
>>>>>> # Thresholds (in seconds)
>>>>>> keepalive 1
>>>>>> warntime 6
>>>>>> deadtime 10
>>>>>> initdead 15
>>>>>>
>>>>>> node dbsuat1a.intranet.mydomain.com
>>>>>> node dbsuat1b.intranet.mydomain.com
>>>>>> #enable pacemaker
>>>>>> crm yes
>>>>>> #enable STONITH
>>>>>> #crm respawn
>>>>>>
>>>>>> my haresources:
>>>>>> DBSUAT1A.intranet.mydomain.com 172.28.185.49
>>>>>>
>>>>> I don't have a very good advice, but you shouldn't use haresources
>>>>> anymore. You should use pacemaker for configuring the cluster.
>>>>>
>>>>> You have said that you wish to use pacemaker(crm) with this line of
>>>>> your config: crm yes
>>>>>
>>>>> Remove the haresources file, restart the heartbeat on both nodes and
>>>>> redo the tests.
>>>>>
>>>>>
>>>>> -----------------------------------------------------------------------
>>>>> -
>>>>>
>>>>> _______________________________________________
>>>>> Linux-HA mailing list
>>>>> [email protected]
>>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>>>> See also: http://linux-ha.org/ReportingProblems
>>>>>
>>>> _______________________________________________
>>>> Linux-HA mailing list
>>>> [email protected]
>>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>>> See also: http://linux-ha.org/ReportingProblems
>>>>
>>> ------------------------------------------------------------------------
>>>
>>> _______________________________________________
>>> Linux-HA mailing list
>>> [email protected]
>>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>>> See also: http://linux-ha.org/ReportingProblems
>>>
>> _______________________________________________
>> Linux-HA mailing list
>> [email protected]
>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> See also: http://linux-ha.org/ReportingProblems
>>
>>
>
>
> ------------------------------------------------------------------------
>
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems