I tried using ping instead of pingd and I added "number" to the evaluation, I get the same results :/
primitive p_ping ocf:pacemaker:ping params host_list=172.20.0.254 clone c_ping p_ping meta globally-unique=false location loc_ping g_cluster_services rule -inf: not_defined p_ping or p_ping number:lte 0 Regards, Craig. On 6 October 2010 20:43, Jayakrishnan <jayakrishnan...@gmail.com> wrote: > > Hello, > > Guess the change:-- > location loc_pingd g_cluster_services rule -inf: not_defined pingd or pingd > number:lte 0 > > should work > > > -- > Regards, > > Jayakrishnan. L > > Visit: > www.foralllinux.blogspot.com > www.jayakrishnan.bravehost.com > > > On Wed, Oct 6, 2010 at 11:56 AM, Claus Denk <d...@us.es> wrote: >> >> I am having a similar problem, so let's wait for the experts, But in the >> meanwhile, try changing >> >> >> location loc_pingd g_cluster_services rule -inf: not_defined p_pingd >> or p_pingd lte 0 >> >> to >> >> location loc_pingd g_cluster_services rule -inf: not_defined pingd >> or pingd number:lte 0 >> >> and see what happens. As far as I have read, it is also more recommended >> to use the "ping" >> resource instead of "pingd"... >> >> kind regards, Claus >> >> >> >> >> >> >> On 10/06/2010 05:45 AM, Craig Hurley wrote: >>> >>> Hello, >>> >>> I have a 2 node cluster, running DRBD, heartbeat and pacemaker in >>> active/passive mode. On both nodes, eth0 is connected to the main >>> network, eth1 is used to connect the nodes directly to each other. >>> The nodes share a virtual IP address on eth0. Pacemaker is also >>> controlling a custom service with an LSB compliant script in >>> /etc/init.d/. All of this is working fine and I'm happy with it. >>> >>> I'd like to configure the nodes so that they fail over if eth0 goes >>> down (or if they cannot access a particular gateway), so I tried >>> adding the following (as per >>> http://www.clusterlabs.org/wiki/Example_configurations#Set_up_pingd) >>> >>> primitive p_pingd ocf:pacemaker:pingd params host_list=172.20.0.254 op >>> monitor interval=15s timeout=5s >>> clone c_pingd p_pingd meta globally-unique=false >>> location loc_pingd g_cluster_services rule -inf: not_defined p_pingd >>> or p_pingd lte 0 >>> >>> ... but when I do add that, all resource are stopped and they don't >>> come back up on either node. Am I making a basic mistake or do you >>> need more info from me? >>> >>> All help is appreciated, >>> Craig. >>> >>> >>> pacemaker >>> Version: 1.0.8+hg15494-2ubuntu2 >>> >>> heartbeat >>> Version: 1:3.0.3-1ubuntu1 >>> >>> drbd8-utils >>> Version: 2:8.3.7-1ubuntu2.1 >>> >>> >>> r...@rpalpha:~$ sudo crm configure show >>> node $id="32482293-7b0f-466e-b405-c64bcfa2747d" rpalpha >>> node $id="3f2aac12-05aa-4ac7-b91f-c47fa28efb44" rpbravo >>> primitive p_drbd_data ocf:linbit:drbd \ >>> params drbd_resource="data" \ >>> op monitor interval="30s" >>> primitive p_fs_data ocf:heartbeat:Filesystem \ >>> params device="/dev/drbd/by-res/data" directory="/mnt/data" >>> fstype="ext4" >>> primitive p_ip ocf:heartbeat:IPaddr2 \ >>> params ip="172.20.50.3" cidr_netmask="255.255.0.0" nic="eth0" \ >>> op monitor interval="30s" >>> primitive p_rp lsb:rp \ >>> op monitor interval="30s" \ >>> meta target-role="Started" >>> group g_cluster_services p_ip p_fs_data p_rp >>> ms ms_drbd p_drbd_data \ >>> meta master-max="1" master-node-max="1" clone-max="2" >>> clone-node-max="1" notify="true" >>> location loc_preferred_master g_cluster_services inf: rpalpha >>> colocation colo_mnt_on_master inf: g_cluster_services ms_drbd:Master >>> order ord_mount_after_drbd inf: ms_drbd:promote g_cluster_services:start >>> property $id="cib-bootstrap-options" \ >>> dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \ >>> cluster-infrastructure="Heartbeat" \ >>> no-quorum-policy="ignore" \ >>> stonith-enabled="false" \ >>> expected-quorum-votes="2" \ >>> >>> >>> r...@rpalpha:~$ sudo cat /etc/ha.d/ha.cf >>> node rpalpha >>> node rpbravo >>> >>> keepalive 2 >>> warntime 5 >>> deadtime 15 >>> initdead 60 >>> >>> mcast eth0 239.0.0.43 694 1 0 >>> bcast eth1 >>> >>> use_logd yes >>> autojoin none >>> crm respawn >>> >>> >>> r...@rpalpha:~$ sudo cat /etc/drbd.conf >>> global { >>> usage-count no; >>> } >>> common { >>> protocol C; >>> >>> handlers {} >>> >>> startup {} >>> >>> disk {} >>> >>> net { >>> cram-hmac-alg sha1; >>> shared-secret "foobar"; >>> } >>> >>> syncer { >>> verify-alg sha1; >>> rate 100M; >>> } >>> } >>> resource data { >>> device /dev/drbd0; >>> meta-disk internal; >>> on rpalpha { >>> disk /dev/mapper/rpalpha-data; >>> address 192.168.1.1:7789; >>> } >>> on rpbravo { >>> disk /dev/mapper/rpbravo-data; >>> address 192.168.1.2:7789; >>> } >>> } >>> _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker