Re: [Linux-HA] Problem around stonith resources IP targets with fence_ipmilan

2010-10-13 Thread Alain.Moulle
Hi, another clue perhaps, I have the same problem with a 3 nodes cluster. And another one : if I start Pacemaker so that DC is the second node node 2 , with the same stonith primitives configuration , and if I power off the first node node1, node1 is correctly fenced by the restofencenode1

Re: [Linux-HA] Problem around stonith resources IP targets with fence_ipmilan

2010-10-13 Thread Alain.Moulle
Hi, I've recompiled /usr/sbin/stonith in rhel6 and configured the same restofence primitives but with ocf agent 'ipmi' instead of fence_ipmilan, and same location constraints for the restofence primitives,and it works perfectly fine : when I power off a node, this is this node which is fenced

[Linux-HA] Documentation of heartbeat protocol

2010-10-13 Thread Steve Davies
Hi, I've done some googling, and read a lot of the source - I can reverse-engineer most of the heartbeat protocol, but would like a definition of the various exchanges, and how each node is meant to respond to various cases. In particular I am interested to understand the meaning of the various

Re: [Linux-HA] Documentation of heartbeat protocol

2010-10-13 Thread Lars Marowsky-Bree
On 2010-10-13T20:36:59, Steve Davies davies...@gmail.com wrote: In particular I am interested to understand the meaning of the various sequence numbers and so forth, and their implications when hosts fail-over, die, return to active status etc. Basically the sort of thing you would find in a

[Linux-HA] Stonith log entries

2010-10-13 Thread mike
Hi all, I've started building a simple 2 node http cluster. I've built several clusters so this should be a joke. I got the first node fired up and noticed these entries over and over again in my logs. Oct 13 21:18:46 Firethorn crmd: [2403]: ERROR: te_connect_stonith: Sign-in failed:

Re: [Linux-HA] Stonith log entries

2010-10-13 Thread mike
Fedora 13 on i686 btw. On 10-10-13 09:26 PM, mike wrote: Hi all, I've started building a simple 2 node http cluster. I've built several clusters so this should be a joke. I got the first node fired up and noticed these entries over and over again in my logs. Oct 13 21:18:46 Firethorn crmd:

[Linux-HA] ip takeover and drbd mount is not working

2010-10-13 Thread Linux Cook
hi! I'm trying to configure drbd with heartbeat. My configuration are as follows: *drbd.conf** global { usage-count no; } resource postgres { protocol C; startup { wfc-timeout 0; degr-wfc-timeout 120; } disk { on-io-error detach; } # or panic, ... net { cram-hmac-alg sha1;