2012/2/19 Viacheslav Biriukov <[email protected]>: > Hi. > You don't need cman in your configuration.The best way is use gfs-pcmk and > dlm-pcmk.
Debatable. > > > 16 февраля 2012 г. 12:30 пользователь emmanuel segura <[email protected]> > написал: > >> I don't think so >> >> I remember gfs2 && ocfs2 on pacemaker & redhat cluster use clvm >> >> In redhat cluster dlm it's started with cman >> >> On pacemaker you should do a clone resource >> >> >> 2012/2/16 diego fanesi <[email protected]> >>> >>> Try to set expected-quorum-votes to 1. I'm not sure but in my opinion >>> that is the problem. >>> >>> Il giorno 14/feb/2012 06:43, "Богомолов Дмитрий Викторович" >>> <[email protected]> ha scritto: >>> >>>> Hi, >>>> I have a trouble with my test configuration. >>>> I build an Actice/Active cluster >>>> Ubuntu(11.10)+DRBD+Cman+Pacemaker+gfs2+Xen for test purpose. >>>> Now i am doing some tests with availability. I am try to start cluster >>>> on one node. >>>> >>>> Trouble is - the Filesystem primitive ClusterFS (fs type=gfs2) does not >>>> start when one of two nodes is switched off. >>>> >>>> Here my configuration: >>>> >>>> node blaster \ >>>> attributes standby="off" >>>> node turrel \ >>>> attributes standby="off" >>>> primitive ClusterData ocf:linbit:drbd \ >>>> params drbd_resource="clusterdata" \ >>>> op monitor interval="60s" >>>> primitive ClusterFS ocf:heartbeat:Filesystem \ >>>> params device="/dev/drbd/by-res/clusterdata" >>>> directory="/mnt/cluster" fstype="gfs2" \ >>>> op start interval="0" timeout="60s" \ >>>> op stop interval="0" timeout="60s" \ >>>> op monitor interval="60s" timeout="60s" >>>> primitive ClusterIP ocf:heartbeat:IPaddr2 \ >>>> params ip="192.168.122.252" cidr_netmask="32" >>>> clusterip_hash="sourceip" \ >>>> op monitor interval="30s" >>>> primitive SSH-stonith stonith:ssh \ >>>> params hostlist="turrel blaster" \ >>>> op monitor interval="60s" >>>> primitive XenDom ocf:heartbeat:Xen \ >>>> params xmfile="/etc/xen/xen1.example.com.cfg" \ >>>> meta allow-migrate="true" is-managed="true" target-role="Stopped" >>>> \ >>>> utilization cores="1" mem="512" \ >>>> op monitor interval="30s" timeout="30s" \ >>>> op start interval="0" timeout="90s" \ >>>> op stop interval="0" timeout="300s" >>>> ms ClusterDataClone ClusterData \ >>>> meta master-max="2" master-node-max="1" clone-max="2" >>>> clone-node-max="1" notify="true" >>>> clone ClusterFSClone ClusterFS \ >>>> meta target-role="Started" is-managed="true" >>>> clone IP ClusterIP \ >>>> meta globally-unique="true" clone-max="2" clone-node-max="2" >>>> clone SSH-stonithClone SSH-stonith >>>> location prefere-blaster XenDom 50: blaster >>>> colocation XenDom-with-ClusterFS inf: XenDom ClusterFSClone >>>> colocation fs_on_drbd inf: ClusterFSClone ClusterDataClone:Master >>>> order ClusterFS-after-ClusterData inf: ClusterDataClone:promote >>>> ClusterFSClone:start >>>> order XenDom-after-ClusterFS inf: ClusterFSClone XenDom >>>> property $id="cib-bootstrap-options" \ >>>> dc-version="1.1.5-01e86afaaa6d4a8c4836f68df80ababd6ca3902f" \ >>>> cluster-infrastructure="cman" \ >>>> expected-quorum-votes="2" \ >>>> stonith-enabled="true" \ >>>> no-quorum-policy="ignore" \ >>>> last-lrm-refresh="1329194925" >>>> rsc_defaults $id="rsc-options" \ >>>> resource-stickiness="100" >>>> >>>> Here is an $crm resource show: >>>> >>>> Master/Slave Set: ClusterDataClone [ClusterData] >>>> Masters: [ turrel ] >>>> Stopped: [ ClusterData:1 ] >>>> Clone Set: IP [ClusterIP] (unique) >>>> ClusterIP:0 (ocf::heartbeat:IPaddr2) Started >>>> ClusterIP:1 (ocf::heartbeat:IPaddr2) Started >>>> Clone Set: ClusterFSClone [ClusterFS] >>>> Stopped: [ ClusterFS:0 ClusterFS:1 ] >>>> Clone Set: SSH-stonithClone [SSH-stonith] >>>> Started: [ turrel ] >>>> Stopped: [ SSH-stonith:1 ] >>>> XenDom (ocf::heartbeat:Xen) Stopped >>>> >>>> I tryed: >>>> crm(live)resource# cleanup ClusterFSClone >>>> Cleaning up ClusterFS:0 on turrel >>>> Cleaning up ClusterFS:1 on turrel >>>> Waiting for 3 replies from the CRMd... OK >>>> >>>> I can see only warn message in /var/log/cluster/corosync.log >>>> Feb 14 16:25:56 turrel pengine: [1640]: WARN: unpack_rsc_op: Processing >>>> failed op ClusterFS:0_start_0 on turrel: unknown exec error (-2) >>>> and >>>> Feb 14 16:25:56 turrel pengine: [1640]: WARN: common_apply_stickiness: >>>> Forcing ClusterFSClone away from turrel after 1000000 failures >>>> (max=1000000) >>>> Feb 14 16:25:56 turrel pengine: [1640]: WARN: common_apply_stickiness: >>>> Forcing ClusterFSClone away from turrel after 1000000 failures >>>> (max=1000000) >>>> >>>> Direct me, please, what i need to check or else? >>>> >>>> Best regards, >>>> Dmitriy Bogomolov >>>> >>>> _______________________________________________ >>>> Pacemaker mailing list: [email protected] >>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>> >>>> Project Home: http://www.clusterlabs.org >>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>> Bugs: http://bugs.clusterlabs.org >>> >>> >>> _______________________________________________ >>> Pacemaker mailing list: [email protected] >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >>> >> >> >> >> -- >> esta es mi vida e me la vivo hasta que dios quiera >> >> _______________________________________________ >> Pacemaker mailing list: [email protected] >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> > > > > -- > Viacheslav Biriukov > BR > http://biriukov.com > > > _______________________________________________ > Pacemaker mailing list: [email protected] > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > _______________________________________________ Pacemaker mailing list: [email protected] http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org
