On 11/01/2016 02:31 AM, Siwakoti, Ganesh wrote: > Hi, > > > i'm using CentOS release 6.8 (Final) as a KVM and i configured 3 > nodes(PM1.local,PM2.local and PM3.local), and using > CMAN clustering. Resources running at two nodes as Active node then > another one node is for Fail-over resource as a Standby node. Configured > resources are groupA and groupB, 5 resources are configured in each > resource group. groupA is Basically run at PM1.local and groupB at > PM2.local. If fail any resource group it'll fail-over at PM3.local. On > PM3.local can run only one Resource group at a time. groupA should not > run at PM2.local and groupB should not run at PM1.local. groupA or > groupB resources not move back on own nodes from PM3.local > automatically, if need to move back on own nodes, can move by > manually.someone pease help me.
Your configuration looks good to me. I don't think you need the -inf rules for fail-counts; migration-threshold=1 handles that. Are you seeing any problems? > > > version information: > > CentOS release 6.8 (Final) (KVM) > > Pacemaker 1.1.14-8.el6_8.1 > > Cman Version: 6.2.0 > > > I want to be sure that my configuration will work properly or not. > > > my Cluster configuration and crm_configuration are on below: > > > cluster.conf: > > <?xml version="1.0"?> > > <cluster config_version="23" name="clusterpm666"> > > <fence_daemon/> > > <clusternodes> > > <clusternode name="pm1.local" nodeid="1"> > > <fence> > > <method name="pcmk-redirect"> > > <device name="pcmk" > port="pm1.local"/> > > </method> > > </fence> > > <altname name="192.168.14.210"/> > > </clusternode> > > <clusternode name="pm2.local" nodeid="2"> > > <fence> > > <method name="pcmk-redirect"> > > <device name="pcmk" > port="pm2.local"/> > > </method> > > </fence> > > <altname name="192.168.14.211"/> > > </clusternode> > > <clusternode name="pm3.local" nodeid="3"> > > <fence> > > <method name="pcmk-redirect"> > > <device name="pcmk" > port="pm3.local"/> > > </method> > > </fence> > > <altname name="192.168.14.212"/> > > </clusternode> > > </clusternodes> > > <fencedevices> > > <fencedevice agent="fence_pcmk" name="pcmk"/> > > </fencedevices> > > <rm> > > <failoverdomains/> > > <resources/> > > </rm> > > </cluster> > > > crm_configuration: > > > node pm1.local \ > > attributes \ > > utilization capacity=1 > > node pm2.local \ > > attributes \ > > utilization capacity=1 > > node pm3.local \ > > attributes \ > > utilization capacity=1 > > primitive asteriskA asterisk \ > > params binary="/usr/sbin/asterisk" canary_binary=astcanary > additional_parameters="-vvvg -I" config="/etc/asteri > > sk_pm1/asterisk.conf" user=root group=root additional_parameters="-vvvg > -I" realtime=true maxfiles=32768 \ > > meta migration-threshold=2 \ > > op monitor interval=20s start-delay=5s timeout=30s \ > > op stop interval=0s on-fail=ignore > > primitive asteriskB asterisk \ > > params binary="/usr/sbin/asterisk" canary_binary=astcanary > additional_parameters="-vvvg -I" config="/etc/asteri > > sk_pm2/asterisk.conf" user=root group=root additional_parameters="-vvvg > -I" realtime=true maxfiles=32768 \ > > meta migration-threshold=2 \ > > op monitor interval=20s start-delay=5s timeout=30s \ > > op stop interval=0s on-fail=ignore > > primitive changeSrcIpA ocf:pacemaker:changeSrcIp \ > > params vip=192.168.12.215 mask=23 device=eth0 \ > > op start interval=0s timeout=0 \ > > op monitor interval=10s \ > > op stop interval=0s on-fail=ignore > > primitive changeSrcIpB ocf:pacemaker:changeSrcIp \ > > params vip=192.168.12.216 mask=23 device=eth0 \ > > op start interval=0s timeout=0 \ > > op monitor interval=10s \ > > op stop interval=0s on-fail=ignore > > primitive cronA lsb:crond \ > > meta migration-threshold=2 \ > > op monitor interval=20s start-delay=5s timeout=15s \ > > op stop interval=0s on-fail=ignore > > primitive cronB lsb:crond \ > > meta migration-threshold=2 \ > > op monitor interval=20s start-delay=5s timeout=15s \ > > op stop interval=0s on-fail=ignore > > primitive vip-local-checkA VIPcheck \ > > params target_ip=192.168.12.215 count=1 wait=5 \ > > op start interval=0s on-fail=restart timeout=60s \ > > op monitor interval=10s timeout=60s \ > > op stop interval=0s on-fail=ignore timeout=60s \ > > utilization capacity=1 > > primitive vip-local-checkB VIPcheck \ > > params target_ip=192.168.12.216 count=1 wait=5 \ > > op start interval=0s on-fail=restart timeout=60s \ > > op monitor interval=10s timeout=60s \ > > op stop interval=0s on-fail=ignore timeout=60s \ > > utilization capacity=1 > > primitive vip-localA IPaddr2 \ > > params ip=192.168.12.215 cidr_netmask=23 nic=eth0 iflabel=0 > broadcast=192.168.13.255 \ > > op start interval=0s timeout=20s \ > > op monitor interval=5s timeout=20s \ > > op stop interval=0s on-fail=ignore > > primitive vip-localB IPaddr2 \ > > params ip=192.168.12.216 cidr_netmask=23 nic=eth0 iflabel=0 > broadcast=192.168.13.255 \ > > op start interval=0s timeout=20s \ > > op monitor interval=5s timeout=20s \ > > op stop interval=0s on-fail=ignore > > group groupA vip-local-checkA vip-localA changeSrcIpA cronA asteriskA \ > > meta target-role=Started > > group groupB vip-local-checkB vip-localB changeSrcIpB cronB asteriskB \ > > meta > > location location-groupA-avoid-failed-node groupA \ > > rule -inf: defined fail-count-vip-local-checkA \ > > rule -inf: defined fail-count-vip-localA \ > > rule -inf: defined fail-count-changeSrcIpA \ > > rule -inf: defined fail-count-cronA \ > > rule -inf: defined fail-count-asteriskA \ > > rule -inf: defined fail-count-vip-local-checkB \ > > rule -inf: defined fail-count-vip-localB \ > > rule -inf: defined fail-count-changeSrcIpB \ > > rule -inf: defined fail-count-cronB \ > > rule -inf: defined fail-count-asteriskB > > location location-groupA-pm1.local-10000 groupA 10000: pm1.local > > location location-groupA-pm2.local--inf groupA resource-discovery=never > -inf: pm2.local > > location location-groupA-pm3.local-1000 groupA 1000: pm3.local > > location location-groupB-avoid-failed-node groupB \ > > rule -inf: defined fail-count-vip-local-checkA \ > > rule -inf: defined fail-count-vip-localA \ > > rule -inf: defined fail-count-changeSrcIpA \ > > rule -inf: defined fail-count-cronA \ > > rule -inf: defined fail-count-asteriskA \ > > rule -inf: defined fail-count-vip-local-checkB \ > > rule -inf: defined fail-count-vip-localB \ > > rule -inf: defined fail-count-changeSrcIpB \ > > rule -inf: defined fail-count-cronB \ > > rule -inf: defined fail-count-asteriskB > > location location-groupB-pm1.local--inf groupB resource-discovery=never > -inf: pm1.local > > location location-groupB-pm2.local-10000 groupB 10000: pm2.local > > location location-groupB-pm3.local-1000 groupB 1000: pm3.local > > property cib-bootstrap-options: \ > > have-watchdog=false \ > > dc-version=1.1.14-8.el6_8.1-70404b0 \ > > cluster-infrastructure=cman \ > > stonith-enabled=false \ > > no-quorum-policy=ignore \ > > enable-startup-probes=true \ > > maintenance-mode=false \ > > startup-fencing=false \ > > dc-deadtime=20s \ > > last-lrm-refresh=1475445380 \ > > placement-strategy=balanced > > rsc_defaults rsc_defaults-options: \ > > migration-threshold=1 \ > > resource-stickiness=INFINITY _______________________________________________ Users mailing list: Users@clusterlabs.org http://clusterlabs.org/mailman/listinfo/users Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org