Jerome Yanga wrote: > Andrew, > > I apologize for my sending my previous email abruptly. > > I have followed your recommendation and installed Pacemaker. > > Here is my config. > > Packages Installed: > heartbeat-2.99.2-6.1 > heartbeat-common-2.99.2-6.1 > heartbeat-debug-2.99.2-6.1 > heartbeat-ldirectord-2.99.2-6.1 > heartbeat-resources-2.99.2-6.1 > libheartbeat2-2.99.2-6.1 > libpacemaker3-1.0.1-3.1 > pacemaker-1.0.1-3.1 > pacemaker-debug-1.0.1-3.1 > pacemaker-pygui-1.4-11.9 > pacemaker-pygui-debug-1.4-11.9 > > > > ha.cf: > # Logging > debug 1 > use_logd false > logfacility daemon > > # Misc Options > traditional_compression off > compression bz2 > coredumps true > > # Communications > udpport 691 > bcast eth1 eth0 > autojoin any > > # Thresholds (in seconds) > keepalive 1 > warntime 6 > deadtime 10 > initdead 15 > > ping 10.50.254.254 > crm respawn > apiauth mgmtd uid=root > respawn root /usr/lib/heartbeat/mgmtd -v > > > cib.xml: > <cib admin_epoch="0" validate-with="pacemaker-1.0" crm_feature_set="3.0" > have-quorum="1" epoch="57" dc-uuid="5e3e3c2d-55e7-4c51-90be-5c4a1912bf3e" > num_updates="0" cib-last-written="Mon Jan 26 13:57:32 2009"> > <configuration> > <crm_config> > <cluster_property_set id="cib-bootstrap-options"> > <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" > value="1.0.1-node: 6fc5ce8302abf145a02891ec41e5a492efbe8efe"/> > </cluster_property_set> > </crm_config> > <nodes> > <node id="5e3e3c2d-55e7-4c51-90be-5c4a1912bf3e" uname="nomen.esri.com" > type="normal"> > <instance_attributes id="nodes-5e3e3c2d-55e7-4c51-90be-5c4a1912bf3e"> > <nvpair id="standby-5e3e3c2d-55e7-4c51-90be-5c4a1912bf3e" > name="standby" value="off"/> > </instance_attributes> > </node> > <node id="27f54ec3-b626-4b4f-b8a6-4ed0b768513c" uname="rubric.esri.com" > type="normal"> > <instance_attributes id="nodes-27f54ec3-b626-4b4f-b8a6-4ed0b768513c"> > <nvpair id="standby-27f54ec3-b626-4b4f-b8a6-4ed0b768513c" > name="standby" value="off"/> > </instance_attributes> > </node> > </nodes> > <resources> > <group id="Directory_Server"> > <meta_attributes id="Directory_Server-meta_attributes"> > <nvpair id="Directory_Server-meta_attributes-collocated" > name="collocated" value="true"/> > <nvpair id="Directory_Server-meta_attributes-ordered" > name="ordered" value="true"/> > <nvpair id="Directory_Server-meta_attributes-resource_stickiness" > name="resource_stickiness" value="100"/> > </meta_attributes> > <primitive class="ocf" id="VIP" provider="heartbeat" type="IPaddr"> > <instance_attributes id="VIP-instance_attributes"> > <nvpair id="VIP-instance_attributes-ip" name="ip" > value="10.50.26.250"/> > </instance_attributes> > <operations id="VIP-ops"> > <op id="VIP-monitor-5s" interval="5s" name="monitor" > timeout="5s"/> > </operations> > </primitive> > <primitive class="ocf" id="ECAS" provider="esri" type="ecas"> > <operations id="ECAS-ops"> > <op id="ECAS-monitor-3s" interval="3s" name="monitor" > timeout="3s"/> > </operations> > <meta_attributes id="ECAS-meta_attributes"> > <nvpair id="ECAS-meta_attributes-target-role" name="target-role" > value="Started"/> > </meta_attributes> > </primitive> > <primitive class="ocf" id="FDS_Admin" provider="esri" type="fdsadm"> > <operations id="FDS_Admin-ops"> > <op id="FDS_Admin-monitor-3s" interval="3s" name="monitor" > timeout="3s"/> > </operations> > </primitive> > </group> > </resources> > <constraints> > <rsc_location id="cli-prefer-Directory_Server" rsc="Directory_Server"> > <rule id="cli-prefer-rule-Directory_Server" score="INFINITY" > boolean-op="and"> > <expression id="cli-prefer-expr-Directory_Server" > attribute="#uname" operation="eq" value="rubric.esri.com" type="string"/> > </rule> > </rsc_location> > <rsc_location id="cli-prefer-FDS_Admin" rsc="FDS_Admin"> > <rule id="cli-prefer-rule-FDS_Admin" score="INFINITY" > boolean-op="and"> > <expression id="cli-prefer-expr-FDS_Admin" attribute="#uname" > operation="eq" value="nomen.esri.com" type="string"/> > </rule> > </rsc_location> > </constraints> > </configuration> > </cib> > > > > I still have the following issues when I only had heartbeat 2.1.3-1. My > concerns are still as follows: > > 01) When a node comes back up after a restart of heartbeat, resources gets > bounced when it rejoins the cluster.
Well, you have defined rsc_location constraints with a score of INFINITY, so that is expected. > 02) Stopping one resource in a group does not failover the group to the > other node. Lookup migration-threshold. Regards Dominik _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
