On 27 May 2014, at 3:45 pm, K Mehta <kiranmehta1...@gmail.com> wrote:

> No reason. Need to know which combination is expected to work.

The bets approach is to update both.  That combination is sure to work.

> using 0.9.26 with new 1.1.10 pacemaker
> and 0.9.90 with old pacemaker has issues in resource deletion.
> 
> Below is the commands executed on 0.9.26 with 1.1.10 pacemaker 
> 
> [root@vsanqa11 tmp]# rpm -qa | grep pcs; rpm -qa | grep pacemaker; rpm -qa | 
> grep corosync; rpm -qa | grep libqb
> pcs-0.9.26-10.el6.noarch
> pacemaker-cli-1.1.10-14.el6_5.3.x86_64
> pacemaker-libs-1.1.10-14.el6_5.3.x86_64
> pacemaker-1.1.10-14.el6_5.3.x86_64
> pacemaker-cluster-libs-1.1.10-14.el6_5.3.x86_64
> corosynclib-1.4.1-17.el6_5.1.x86_64
> corosync-1.4.1-17.el6_5.1.x86_64
> libqb-devel-0.16.0-2.el6.x86_64
> libqb-0.16.0-2.el6.x86_64
> 
> 
> [root@vsanqa11 tmp]# pcs status
> Last updated: Mon May 26 22:39:42 2014
> Last change: Mon May 26 22:39:01 2014 via cibadmin on vsanqa11
> Stack: cman
> Current DC: vsanqa11 - partition with quorum
> Version: 1.1.10-14.el6_5.3-368c726
> 2 Nodes configured
> 2 Resources configured
> 
> 
> Online: [ vsanqa11 vsanqa12 ]
> 
> Full list of resources:
> 
>  Master/Slave Set: ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a 
> [vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a]
>      Masters: [ vsanqa12 ]
>      Slaves: [ vsanqa11 ]
> 
> 
> 
> [root@vsanqa11 tmp]# pcs config
> Corosync Nodes:
> 
> Pacemaker Nodes:
>  vsanqa11 vsanqa12
> 
> Resources:
> 
> Location Constraints:
>   Resource: ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
>     Enabled on: vsanqa11
>     Enabled on: vsanqa12
>   Resource: vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
>     Enabled on: vsanqa11
>     Enabled on: vsanqa12
> Ordering Constraints:
> Colocation Constraints:
> 
> Cluster Properties:
>  dc-version: 1.1.10-14.el6_5.3-368c726
>  cluster-infrastructure: cman
>  last-lrm-refresh: 1401098102
>  expected-quorum-votes: 2
>  stonith-enabled: false
>  no-quorum-policy: ignore
> 
> 
> 
> [root@vsanqa11 tmp]# pcs resource show ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
> Resource: ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
>   cluster_uuid: 9fc36888-cf2a-417a-907c-db3f5e9b7a8a
>   clone-max: 2
>   globally-unique: false
>   target-role: started
>   op monitor interval=31s role=Slave timeout=100s
> [root@vsanqa11 tmp]# pcs resource show 
> vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
> Resource: vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
>   cluster_uuid: 9fc36888-cf2a-417a-907c-db3f5e9b7a8a
>   op monitor interval=31s role=Slave timeout=100s
> 
> 
> [root@vsanqa11 tmp]# pcs resource delete 
> ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
> ERROR: Unable to update cib
> Call cib_replace failed (-203): Update does not conform to the configured 
> schema
> <cib admin_epoch="0" cib-last-written="Mon May 26 22:40:45 2014" 
> crm_feature_set="3.0.7" dc-uuid="vsanqa11" epoch="10550" have-quorum="1" 
> num_updates="1" update-client="cibadmin" update-origin="vsanqa11" 
> validate-with="pacemaker-1.2">
>   <configuration>
>     <crm_config>
>       <cluster_property_set id="cib-bootstrap-options">
>         <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" 
> value="1.1.10-14.el6_5.3-368c726"/>
>         <nvpair id="cib-bootstrap-options-cluster-infrastructure" 
> name="cluster-infrastructure" value="cman"/>
>         <nvpair id="cib-bootstrap-options-last-lrm-refresh" 
> name="last-lrm-refresh" value="1401098102"/>
>         <nvpair id="cib-bootstrap-options-expected-quorum-votes" 
> name="expected-quorum-votes" value="2"/>
>         <nvpair id="cib-bootstrap-options-stonith-enabled" 
> name="stonith-enabled" value="false"/>
>         <nvpair id="cib-bootstrap-options-no-quorum-policy" 
> name="no-quorum-policy" value="ignore"/>
>       </cluster_property_set>
>     </crm_config>
>     <nodes>
>       <node id="vsanqa11" uname="vsanqa11"/>
>       <node id="vsanqa12" uname="vsanqa12"/>
>     </nodes>
>     <resources/>
>     <constraints>
>       <rsc_location 
> id="location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY" 
> node="vsanqa11" rsc="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" 
> score="INFINITY"/>
>       <rsc_location 
> id="location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY" 
> node="vsanqa12" rsc="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" 
> score="INFINITY"/>
>       <rsc_location 
> id="location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY" 
> node="vsanqa11" rsc="ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" 
> score="INFINITY"/>
>       <rsc_location 
> id="location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY" 
> node="vsanqa12" rsc="ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" 
> score="INFINITY"/>
>     </constraints>
>     <rsc_defaults>
>       <meta_attributes id="rsc_defaults-options">
>         <nvpair id="rsc_defaults-options-resource-stickiness" 
> name="resource-stickiness" value="100"/>
>         <nvpair id="rsc_defaults-options-timeout" name="timeout" 
> value="100s"/>
>       </meta_attributes>
>     </rsc_defaults>
>   </configuration>
>   <status>
>     <node_state crm-debug-origin="do_update_resource" crmd="online" 
> expected="member" id="vsanqa11" in_ccm="true" join="member" uname="vsanqa11">
>       <transient_attributes id="vsanqa11">
>         <instance_attributes id="status-vsanqa11">
>           <nvpair id="status-vsanqa11-probe_complete" name="probe_complete" 
> value="true"/>
>           <nvpair 
> id="status-vsanqa11-master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" 
> name="master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" value="4"/>
>         </instance_attributes>
>       </transient_attributes>
>       <lrm id="vsanqa11">
>         <lrm_resources>
>           <lrm_resource id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" 
> type="vgc-cm-agent.ocf" class="ocf" provider="heartbeat">
>             <lrm_rsc_op id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_last_0" 
> operation_key="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_start_0" 
> operation="start" crm-debug-origin="do_update_resource" 
> crm_feature_set="3.0.7" 
> transition-key="7:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" 
> transition-magic="0:0;7:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" 
> call-id="290" rc-code="0" op-status="0" interval="0" last-run="1401169141" 
> last-rc-change="1401169141" exec-time="1103" queue-time="0" 
> op-digest="494c7d757cceae4f35487404ebc12a10"/>
>             <lrm_rsc_op 
> id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_monitor_31000" 
> operation_key="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_monitor_31000" 
> operation="monitor" crm-debug-origin="do_update_resource" 
> crm_feature_set="3.0.7" 
> transition-key="8:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" 
> transition-magic="0:0;8:195:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" 
> call-id="293" rc-code="0" op-status="0" interval="31000" 
> last-rc-change="1401169142" exec-time="76" queue-time="0" 
> op-digest="f81bd1d7870f6bb69f88312740132d65"/>
>           </lrm_resource>
>         </lrm_resources>
>       </lrm>
>     </node_state>
>     <node_state crm-debug-origin="do_update_resource" crmd="online" 
> expected="member" id="vsanqa12" in_ccm="true" join="member" uname="vsanqa12">
>       <transient_attributes id="vsanqa12">
>         <instance_attributes id="status-vsanqa12">
>           <nvpair id="status-vsanqa12-probe_complete" name="probe_complete" 
> value="true"/>
>           <nvpair 
> id="status-vsanqa12-master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" 
> name="master-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" value="5"/>
>         </instance_attributes>
>       </transient_attributes>
>       <lrm id="vsanqa12">
>         <lrm_resources>
>           <lrm_resource id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a" 
> type="vgc-cm-agent.ocf" class="ocf" provider="heartbeat">
>             <lrm_rsc_op id="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_last_0" 
> operation_key="vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a_promote_0" 
> operation="promote" crm-debug-origin="do_update_resource" 
> crm_feature_set="3.0.7" 
> transition-key="11:196:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" 
> transition-magic="0:0;11:196:0:79ecdaeb-e637-4fdf-b8e8-ebfc7e2eca39" 
> call-id="275" rc-code="0" op-status="0" interval="0" last-run="1401169174" 
> last-rc-change="1401169174" exec-time="196" queue-time="0" 
> op-digest="494c7d757cceae4f35487404ebc12a10"/>
>           </lrm_resource>
>         </lrm_resources>
>       </lrm>
>     </node_state>
>   </status>
> </cib>
> 
> 
> ****deleted on base (non clone) resource works however it stays in orphaned 
> state for few seconds
> 
> [root@vsanqa11 tmp]# pcs resource delete 
> vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a; pcs status
> Removing Constraint - 
> location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY
> Removing Constraint - 
> location-ms-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY
> Removing Constraint - 
> location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa11-INFINITY
> Removing Constraint - 
> location-vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a-vsanqa12-INFINITY
> Deleting Resource - vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a
> Last updated: Mon May 26 22:41:08 2014
> Last change: Mon May 26 22:41:07 2014 via cibadmin on vsanqa11
> Stack: cman
> Current DC: vsanqa11 - partition with quorum
> Version: 1.1.10-14.el6_5.3-368c726
> 2 Nodes configured
> 0 Resources configured
> 
> 
> Online: [ vsanqa11 vsanqa12 ]
> 
> Full list of resources:
> 
>  vha-9fc36888-cf2a-417a-907c-db3f5e9b7a8a       
> (ocf::heartbeat:vgc-cm-agent.ocf):       ORPHANED Master [ vsanqa11 vsanqa12 ]
> 
> [root@vsanqa11 tmp]# pcs status
> Last updated: Mon May 26 22:41:18 2014
> Last change: Mon May 26 22:41:07 2014 via cibadmin on vsanqa11
> Stack: cman
> Current DC: vsanqa11 - partition with quorum
> Version: 1.1.10-14.el6_5.3-368c726
> 2 Nodes configured
> 0 Resources configured
> 
> 
> Online: [ vsanqa11 vsanqa12 ]
> 
> Full list of resources:
> 
> 
> 
> On Tue, May 27, 2014 at 11:01 AM, Andrew Beekhof <and...@beekhof.net> wrote:
> 
> On 27 May 2014, at 2:34 pm, K Mehta <kiranmehta1...@gmail.com> wrote:
> 
> > I have seen that 0.9.26 works with 1.1.8 pacemaker and 0.9.90 works with 
> > 1.1.10 pacemaker.
> > However, with 0.9.90 pcs and 1.1.8 pacemaker, pcs delete resource 
> > <multistate resource name> fails with error CIB update failed because of 
> > schema error
> 
> Any specific reason to stay on 1.1.8?
> 
> >
> >
> > On Tue, May 27, 2014 at 5:28 AM, Andrew Beekhof <and...@beekhof.net> wrote:
> >
> > On 26 May 2014, at 5:15 pm, K Mehta <kiranmehta1...@gmail.com> wrote:
> >
> > > pcs versions 0.9.26 and 0.9.90
> > > pacemaker versions 1.1.8 and 1.1.10
> > >
> > > Which pcs versions are expected to work with which pacemaker versions ?
> >
> > I think for the most part, all versions will work together.
> > There may be the odd command that requires a flag exposed by newer versions 
> > of pacemaker, but that should be minimal.
> >
> > >
> > >
> > > Regards,
> > >  Kiran
> > > _______________________________________________
> > > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> > >
> > > Project Home: http://www.clusterlabs.org
> > > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > > Bugs: http://bugs.clusterlabs.org
> >
> >
> > _______________________________________________
> > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> >
> >
> > _______________________________________________
> > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> 
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
> 
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org

Attachment: signature.asc
Description: Message signed with OpenPGP using GPGMail

_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to