group grp_application res_mount-application res_Service-IP 
res_mount-CIFSshareData res_mount-CIFSshareData2  res_mount-CIFSshareData3 
res_MyApplication

is just a shortcut for 

colocation res_Service-IP with res_mount-application
colocation res_mount-CIFSshareData with res_Service-IP
...

and

colocation res_mount-application then res_Service-IP
colocation res_Service-IP then res_mount-CIFSshareData
...

just re-order and remove any bit you don't need anymore



> On 21 Jan 2015, at 3:43 am, Sven Moeller <smoel...@nichthelfer.de> wrote:
> 
> Hi,
> 
> i have a Cluster running on SLES 11 SP3. A resource group is defined. So 
> every entry in that resource group is a hard dependency for the folliowing 
> resource in that group. If just only one resource fails on both nodes, the 
> resource res_MyApplication won't be started. I want to change that behavior. 
> I want the cluster to start the resources on just one host. The local 
> resources should be started on just the same host, regardless of the status 
> of the CIFS Shares. If the CIFS Shares are available they have to be mounted 
> on the same node running the application. I would try to accomplish this by 
> deleting the resource group grp_application, creating colocation constrains 
> for local fs, service ip and application. Additionally I would create 
> co-location constrains for each CIFS Share to mount it on the node running 
> the application. Any hints/thoughts on that? Is that the "right way" to 
> achieve what I want?
> 
> Here's the currently running config:
> 
> node mynodea\
>        attributes standby="off"
> node mynodeb \
>        attributes standby="off"
> primitive res_Service-IP ocf:heartbeat:IPaddr2 \
>        params ip="192.168.10.120" cidr_netmask="24" \
>        op monitor interval="10s" timeout="20s" depth="0" \
>        meta target-role="started"
> primitive res_MyApplication lsb:myapp \
>        operations $id="res_MyApplication-operations" \
>        op start interval="0" timeout="180" \
>        op stop interval="0" timeout="180" \
>        op status interval="0" timeout="600" \
>        op monitor interval="15" timeout="15" start-delay="120" \
>        op meta-data interval="0" timeout="600" \
>        meta target-role="Started"
> primitive res_mount-CIFSshareData ocf:heartbeat:Filesystem \
>        params device="//CIFS/Share1" directory="/Datafstype="cifs" 
> options="uid=30666,gid=30666" \
>        op monitor interval="20" timeout="40" depth="0" \
>        meta target-role="started"
> primitive res_mount-CIFSshareData2 ocf:heartbeat:Filesystem \
>        params device="//CIFS/Share2" directory="/Data2" fstype="cifs" 
> options="uid=30666,gid=30666" \
>        op monitor interval="20" timeout="40" depth="0" \
>        meta target-role="started"
> primitive res_mount-CIFSshareData3 ocf:heartbeat:Filesystem \
>        params device="//CIFS/Share3" directory="/Data3" fstype="cifs" 
> options="uid=30666,gid=30666" \
>        op monitor interval="20" timeout="40" depth="0" \
>        meta target-role="started"
> primitive res_mount-application ocf:heartbeat:Filesystem \
>        params 
> device="/dev/disk/by-id/scsi-36000d56e0000324561d9dcae19034e90-part1" 
> directory="/MyApp" fstype="ext3" \
>        op monitor interval="20" timeout="40" depth="0" \
>        meta target-role="started"
> primitive stonith-sbd stonith:external/sbd \
>        params 
> sbd_device="/dev/disk/by-id/scsi-36000d56e0000a31a36d8dcaebaf5a439" \
>        meta target-role="Started"
> group grp_application res_mount-application res_Service-IP 
> res_mount-CIFSshareData res_mount-CIFSshareData2  res_mount-CIFSshareData3 
> res_MyApplication
> location prefer-application grp_application 50: nodea
> property $id="cib-bootstrap-options" \
>        stonith-enabled="true" \
>        no-quorum-policy="ignore" \
>        placement-strategy="balanced" \
>        dc-version="1.1.9-2db99f1" \
>        cluster-infrastructure="classic openais (with plugin)" \
>        expected-quorum-votes="2" \
>        last-lrm-refresh="1412758622" \
>        stonith-action="poweroff" \
>        stonith-timeout="216s" \
>        maintenance-mode="true"
> rsc_defaults $id="rsc-options" \
>        resource-stickiness="100" \
>        migration-threshold="3"
> op_defaults $id="op-options" \
>        timeout="600" \
>        record-pending="true"
> 
> Thanks and kind regards
> 
> Sven
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org


_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to