Well I posted my config and I've tried various things and tested this setup... and it still behaves incorrectly: going primary in the event of a complete loss of network connectivity.

I mean... its an NFS server... *network* filesystem. If it can't connect to the network *at* *all* it makes no sense to become the primary NFS server...

I'd really appreciate some comment on what may be wrong in the config files that I've posted. If theres any further info that I need to post please mention it.

Thanks!


Steve Wray wrote:
Dejan Muhamedagic wrote:
Hi,

On Wed, Jan 16, 2008 at 03:29:59PM +1300, Steve Wray wrote:
Hi there,

I have two servers set up with drbd, heartbeat and NFS.

They are only talking over the network, there is no serial heartbeat, and can't be at the moment.

What I'd like to happen is that if either one loses network connectivity they "sit down and shut up", ie the one which cannot reach, say, the router, does not try to become the active node; it sends drbd into standalone, does not mount the drbd filesystem, does not bring up NFS and does not bring up the floating interface. It has lost network, it should go into a 'passive' mode and wait for the network to come back.

I tried using 'ping' with "respawn hacluster /usr/lib/heartbeat/ipfail" and to list a few IP addresses which should be reachable, but this doesn't seem to do what I want. Rather, when it loses network and cannot ping them it puts drbd into standalone, mounts the drbd filesystem, starts NFS and brings up the interface on the floating address... seems a bit wierd to me.

Is there an FAQ or something for this?

If you're using v1 style configuration, then ipfail is the right
thing. Otherwise, you should go with the pingd. See:

I'm using v1.


http://www.linux-ha.org/ipfail

I had read this, which is where I got the idea from.

However, it doesn't seem to be working.

Here is what I have in ha.cf:

logfacility     local0
keepalive 100ms
deadping 5
deadtime 30
warntime 10
ucast eth0 10.10.2.26
ucast eth0 10.10.2.27
node drbd-test-1
node drbd-test-2
auto_failback on
ping 10.10.10.1
respawn hacluster /usr/lib/heartbeat/ipfail



Here is what I have in haresources:

drbd-test-1  \
        drbddisk::drbdtest \
        Filesystem::/dev/drbd0::/data::ext3 \
        killnfsd \
        nfs-common \
        nfs-kernel-server \
        Delay::20::0 \
        IPaddr::10.10.2.28/16/eth0


If I go onto the node which is currently secondary and take down eth0 it takes over as primary, mounts drbd0, starts nfs and tries to bring up the floating IP.

This is the opposite of the behavior I expected to achieve with this config.


Note that this is a /16 network ie 255.255.0.0 so all listed IP addresses are in the same network space. 10.10.10.1 is the default route for both nodes.

Both nodes are running on the same Xen dom0 host ie there is no physical network between them (I can't pull a network cable out to test loss of connectivity so I use ifdown) and are running Debian Etch.


http://linux-ha.org/pingd

Thanks,

Dejan


Thanks
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to