On 27 May 2014, at 3:45 pm, K Mehta <kiranmehta1...@gmail.com> wrote:
> No reason. Need to know which combination is expected to work. The bets approach is to update both. That combination is sure to work. > using 0.9.26 with new 1.1.10 pacemaker > and 0.9.90 with old pacemaker has issues in resource deletion. > > Below is the commands executed on 0.9.26 with 1.1.10 pacemaker > > [root@vsanqa11 tmp]# rpm -qa | grep pcs; rpm -qa | grep pacemaker; rpm -qa | > grep corosync; rpm -qa | grep libqb > pcs-0.9.26-10.el6.noarch > pacemaker-cli-1.1.10-14.el6_5.3.x86_64 > pacemaker-libs-1.1.10-14.el6_5.3.x86_64 > pacemaker-1.1.10-14.el6_5.3.x86_64 > pacemaker-cluster-libs-1.1.10-14.el6_5.3.x86_64 > corosynclib-1.4.1-17.el6_5.1.x86_64 > corosync-1.4.1-17.el6_5.1.x86_64 > libqb-devel-0.16.0-2.el6.x86_64 > libqb-0.16.0-2.el6.x86_64 > > > [root@vsanqa11 tmp]# pcs status > Last updated: Mon May 26 22:39:42 2014 > Last change: Mon May 26 22:39:01 2014 via cibadmin on vsanqa11 > Stack: cman > Current DC: vsanqa11 - partition with quorum > Version: 1.1.10-14.el6_5.3-368c726 > 2 Nodes configured > 2 Resources configured > > > Online: [ vsanqa11 vsanqa12 ] > > Full list of resources: > > Master/Slave Set: ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a > [vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a] > Masters: [ vsanqa12 ] > Slaves: [ vsanqa11 ] > > > > [root@vsanqa11 tmp]# pcs config > Corosync Nodes: > > Pacemaker Nodes: > vsanqa11 vsanqa12 > > Resources: > > Location Constraints: > Resource: ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a > Enabled on: vsanqa11 > Enabled on: vsanqa12 > Resource: vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a > Enabled on: vsanqa11 > Enabled on: vsanqa12 > Ordering Constraints: > Colocation Constraints: > > Cluster Properties: > dc-version: 1.1.10-14.el6_5.3-368c726 > cluster-infrastructure: cman > last-lrm-refresh: 1401098102 > expected-quorum-votes: 2 > stonith-enabled: false > no-quorum-policy: ignore > > > > [root@vsanqa11 tmp]# pcs resource show ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a > Resource: ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a > cluster_uuid: 9fc36888-cf2a-417a-907c-db3f5e9b7a8a > clone-max: 2 > globally-unique: false > target-role: started > op monitor interval=31s role=Slave timeout=100s > [root@vsanqa11 tmp]# pcs resource show > vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a > Resource: vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a > cluster_uuid: 9fc36888-cf2a-417a-907c-db3f5e9b7a8a > op monitor interval=31s role=Slave timeout=100s > > > [root@vsanqa11 tmp]# pcs resource delete > ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a > ERROR: Unable to update cib > Call cib_replace failed (-203): Update does not conform to the configured > schema > <cib admin_epoch="0" cib-last-written="Mon May 26 22:40:45 2014" > crm_feature_set="3.0.7" dc-uuid="vsanqa11" epoch="10550" have-quorum="1" > num_updates="1" update-client="cibadmin" update-origin="vsanqa11" > validate-with="pacemaker-1.2"> > <configuration> > <crm_config> > <cluster_property_set id="cib-bootstrap-options"> > <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" > value="1.1.10-14.el6_5.3-368c726"/> > <nvpair id="cib-bootstrap-options-cluster-infrastructure" > name="cluster-infrastructure" value="cman"/> > <nvpair id="cib-bootstrap-options-last-lrm-refresh" > name="last-lrm-refresh" value="1401098102"/> > <nvpair id="cib-bootstrap-options-expected-quorum-votes" > name="expected-quorum-votes" value="2"/> > <nvpair id="cib-bootstrap-options-stonith-enabled" > name="stonith-enabled" value="false"/> > <nvpair id="cib-bootstrap-options-no-quorum-policy" > name="no-quorum-policy" value="ignore"/> > </cluster_property_set> > </crm_config> > <nodes> > <node id="vsanqa11" uname="vsanqa11"/> > <node id="vsanqa12" uname="vsanqa12"/> > </nodes> > <resources/> > <constraints> > <rsc_location > id="location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY" > node="vsanqa11" rsc="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" > score="INFINITY"/> > <rsc_location > id="location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY" > node="vsanqa12" rsc="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" > score="INFINITY"/> > <rsc_location > id="location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY" > node="vsanqa11" rsc="ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" > score="INFINITY"/> > <rsc_location > id="location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY" > node="vsanqa12" rsc="ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" > score="INFINITY"/> > </constraints> > <rsc_defaults> > <meta_attributes id="rsc_defaults-options"> > <nvpair id="rsc_defaults-options-resource-stickiness" > name="resource-stickiness" value="100"/> > <nvpair id="rsc_defaults-options-timeout" name="timeout" > value="100s"/> > </meta_attributes> > </rsc_defaults> > </configuration> > <status> > <node_state crm-debug-origin="do_update_resource" crmd="online" > expected="member" id="vsanqa11" in_ccm="true" join="member" uname="vsanqa11"> > <transient_attributes id="vsanqa11"> > <instance_attributes id="status-vsanqa11"> > <nvpair id="status-vsanqa11-probe_complete" name="probe_complete" > value="true"/> > <nvpair > id="status-vsanqa11-master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" > name="master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" value="4"/> > </instance_attributes> > </transient_attributes> > <lrm id="vsanqa11"> > <lrm_resources> > <lrm_resource id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" > type="vgc-cm-agent.ocf" class="ocf" provider="heartbeat"> > <lrm_rsc_op id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_last_0" > operation_key="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_start_0" > operation="start" crm-debug-origin="do_update_resource" > crm_feature_set="3.0.7" > transition-key="7:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" > transition-magic="0:0;7:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" > call-id="290" rc-code="0" op-status="0" interval="0" last-run="1401169141" > last-rc-change="1401169141" exec-time="1103" queue-time="0" > op-digest="494c7d757cceae4f35487404ebc12a10"/> > <lrm_rsc_op > id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_monitor_31000" > operation_key="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_monitor_31000" > operation="monitor" crm-debug-origin="do_update_resource" > crm_feature_set="3.0.7" > transition-key="8:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" > transition-magic="0:0;8:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" > call-id="293" rc-code="0" op-status="0" interval="31000" > last-rc-change="1401169142" exec-time="76" queue-time="0" > op-digest="f81bd1d7870f6bb69f88312740132d65"/> > </lrm_resource> > </lrm_resources> > </lrm> > </node_state> > <node_state crm-debug-origin="do_update_resource" crmd="online" > expected="member" id="vsanqa12" in_ccm="true" join="member" uname="vsanqa12"> > <transient_attributes id="vsanqa12"> > <instance_attributes id="status-vsanqa12"> > <nvpair id="status-vsanqa12-probe_complete" name="probe_complete" > value="true"/> > <nvpair > id="status-vsanqa12-master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" > name="master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" value="5"/> > </instance_attributes> > </transient_attributes> > <lrm id="vsanqa12"> > <lrm_resources> > <lrm_resource id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" > type="vgc-cm-agent.ocf" class="ocf" provider="heartbeat"> > <lrm_rsc_op id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_last_0" > operation_key="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_promote_0" > operation="promote" crm-debug-origin="do_update_resource" > crm_feature_set="3.0.7" > transition-key="11:196:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" > transition-magic="0:0;11:196:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" > call-id="275" rc-code="0" op-status="0" interval="0" last-run="1401169174" > last-rc-change="1401169174" exec-time="196" queue-time="0" > op-digest="494c7d757cceae4f35487404ebc12a10"/> > </lrm_resource> > </lrm_resources> > </lrm> > </node_state> > </status> > </cib> > > > ****deleted on base (non clone) resource works however it stays in orphaned > state for few seconds > > [root@vsanqa11 tmp]# pcs resource delete > vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a; pcs status > Removing Constraint - > location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY > Removing Constraint - > location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY > Removing Constraint - > location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY > Removing Constraint - > location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY > Deleting Resource - vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a > Last updated: Mon May 26 22:41:08 2014 > Last change: Mon May 26 22:41:07 2014 via cibadmin on vsanqa11 > Stack: cman > Current DC: vsanqa11 - partition with quorum > Version: 1.1.10-14.el6_5.3-368c726 > 2 Nodes configured > 0 Resources configured > > > Online: [ vsanqa11 vsanqa12 ] > > Full list of resources: > > vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a > (ocf::heartbeat:vgc-cm-agent.ocf): ORPHANED Master [ vsanqa11 vsanqa12 ] > > [root@vsanqa11 tmp]# pcs status > Last updated: Mon May 26 22:41:18 2014 > Last change: Mon May 26 22:41:07 2014 via cibadmin on vsanqa11 > Stack: cman > Current DC: vsanqa11 - partition with quorum > Version: 1.1.10-14.el6_5.3-368c726 > 2 Nodes configured > 0 Resources configured > > > Online: [ vsanqa11 vsanqa12 ] > > Full list of resources: > > > > On Tue, May 27, 2014 at 11:01 AM, Andrew Beekhof <and...@beekhof.net> wrote: > > On 27 May 2014, at 2:34 pm, K Mehta <kiranmehta1...@gmail.com> wrote: > > > I have seen that 0.9.26 works with 1.1.8 pacemaker and 0.9.90 works with > > 1.1.10 pacemaker. > > However, with 0.9.90 pcs and 1.1.8 pacemaker, pcs delete resource > > <multistate resource name> fails with error CIB update failed because of > > schema error > > Any specific reason to stay on 1.1.8? > > > > > > > On Tue, May 27, 2014 at 5:28 AM, Andrew Beekhof <and...@beekhof.net> wrote: > > > > On 26 May 2014, at 5:15 pm, K Mehta <kiranmehta1...@gmail.com> wrote: > > > > > pcs versions 0.9.26 and 0.9.90 > > > pacemaker versions 1.1.8 and 1.1.10 > > > > > > Which pcs versions are expected to work with which pacemaker versions ? > > > > I think for the most part, all versions will work together. > > There may be the odd command that requires a flag exposed by newer versions > > of pacemaker, but that should be minimal. > > > > > > > > > > > Regards, > > > Kiran > > > _______________________________________________ > > > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > > > > > Project Home: http://www.clusterlabs.org > > > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > > > Bugs: http://bugs.clusterlabs.org > > > > > > _______________________________________________ > > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > > > Project Home: http://www.clusterlabs.org > > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > > Bugs: http://bugs.clusterlabs.org > > > > > > _______________________________________________ > > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > > > Project Home: http://www.clusterlabs.org > > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > > Bugs: http://bugs.clusterlabs.org > > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org
signature.asc
Description: Message signed with OpenPGP using GPGMail
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org