Re: [Linux-ha-dev] Dovecot OCF Resource Agent

2011-07-22 Thread jeroen

On Fri, 15 Apr 2011 14:45:59 +0200, Raoul Bhatia [IPAX]
r.bha...@ipax.at
wrote:
 On 04/15/2011 01:19 PM, Andrew Beekhof wrote:
 On Fri, Apr 15, 2011 at 12:53 PM, Raoul Bhatia [IPAX] r.bha...@ipax.at
 wrote:
 On 04/15/2011 11:10 AM, jer...@intuxicated.org wrote:

 Yes, it does the same thing but contains some additional features,
like
 logging into a mailbox.

 first of all, i do not know how the others think about a ocf ra
 implemented in c. i'll suggest waiting for comments from dejan or
 fghass.
 
 the ipv6addr agent was written in C too
 the OCF standard does not dictate the language to be used - its really
 a matter of whether C is the best tool for this job
 
 thank you andrew!
 
 jeroen, can you please create a github fork off
 https://github.com/ClusterLabs/ (it's really easy!)
 
 and add your resource agent in the same fashion as IPv6addr.c [1] ?
 
 thanks,
 raoul
 
 [1]

https://github.com/ClusterLabs/resource-agents/blob/master/heartbeat/IPv6addr.c

Hi,

I finally found some time to get the code on GitHub.

https://github.com/perrit/dovecot-ocf-resource-agent

As you can see it's kind of hard to merge the code in the same way as
IPv6addr.c as it currently spans multiple files. Would you like me to just
put it in a directory? Maybe it's a good idea to split the dovecot part and
the mailbox login part, so that there's a mailbox login resource agent
becomes more like the ping resource agent?

Regards,
Jeroen
___
Linux-HA-Dev: Linux-HA-Dev@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha-dev
Home Page: http://linux-ha.org/


Re: [Linux-ha-dev] [PATCH] fix config parameter type for pgsql

2011-07-22 Thread Serge Dubrouski
Thanks for catching this! Could somebody apply it?

On Thu, Jul 21, 2011 at 7:43 PM, Takatoshi MATSUO matsuo@gmail.comwrote:

 Hi

 I found a slight bug for pgsql.
 You know,  type of config parameter is not integer but string.


 Best Regard,
 Takatoshi MATSUO

 ___
 Linux-HA-Dev: Linux-HA-Dev@lists.linux-ha.org
 http://lists.linux-ha.org/mailman/listinfo/linux-ha-dev
 Home Page: http://linux-ha.org/




-- 
Serge Dubrouski.
___
Linux-HA-Dev: Linux-HA-Dev@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha-dev
Home Page: http://linux-ha.org/


[Linux-HA] Need help for ocf-ra agent, searching an imap agent

2011-07-22 Thread Jörg Lang
hallo to all,

first sorry for my bad english.

then i want to ask if anyone can help me.
i search an ocf-ra agent to use an Cyrus2.2 IMAP server on an linux-ha
system.
unfortunately i can not find a ocf-ra agent.

please can anyone help me?

thanks a lot and greetings
joerg
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] help is needed as the stonith_host directive is not happending!

2011-07-22 Thread Avestan

Hello,

Sorry for the lack of information. You guys are so good that sometimes I
think you have a crystal-ball. ;o)

As the following shows, I am running Heartbeat and STONITH version 2.0.8
release 1 on Fedora 7.

[root@shemshak~]# rpm -qa | grep -i heartbeat
heartbeat-2.0.8-1.fc7
[root@shemshak~]# rpm -qa | grep -i stonith
stonith-2.0.8-1.fc7

This is an old system which I built over 2 years ago and still runs like a
clock. I have recently added two STONITH Devices (APC9225 MasterSwithch Plus
with APC9617 Network Management card)

and here is the heartbeat configuration file /etc/ha.d/ha.cf:

# Heartbeat logging configuration
debugfile   /var/log/ha-debug
logfile /var/log/ha-log
logfacility local0

# Heartbeat cluster members
node shemshak
node dizin

# Heartbeat communication timing
keepalive 2
deadtime 32
initdead 64

# Heartbeat communication paths
udpport 694
bcast  eth1
#ucast eth1 192.168.1.21
#ucast eth1 192.168.1.22
#ucast eth0 192.168.1.81
#ucast eth0 192.168.1.82
baud 19200
serial /dev/ttyS0

# Don't fail back automatically - on/off
auto_failback on

# Monitoring of network connection to default gateway
ping 192.168.1.1

#respawn hacluster /usr/lib64/heartbeat/ipfail

#STONITH
stonith_host Testing apcmaster 192.168.1.56 apc apc

Here is also my log file /var/log/ha-log after stopping the heartbeat on
the primary host by issuing service heartbeat stop command at
2011/07/22_08:30:48

[root@shemshak ~]# tail -f /var/log/ha-log
heartbeat[4741]: 2011/07/21_18:36:04 info: Current arena value: 0
heartbeat[4741]: 2011/07/21_18:36:04 info: MSG stats: 0/190108 ms age 10
[pid4749/HBWRITE]
heartbeat[4741]: 2011/07/21_18:36:04 info: ha_malloc stats: 379/5069800 
38076/18447 [pid4749/HBWRITE]
heartbeat[4741]: 2011/07/21_18:36:04 info: RealMalloc stats: 50112 total
malloc bytes. pid [4749/HBWRITE]
heartbeat[4741]: 2011/07/21_18:36:04 info: Current arena value: 0
heartbeat[4741]: 2011/07/21_18:36:04 info: MSG stats: 0/86408 ms age 20
[pid4750/HBREAD]
heartbeat[4741]: 2011/07/21_18:36:04 info: ha_malloc stats: 380/1815007 
38160/18491 [pid4750/HBREAD]
heartbeat[4741]: 2011/07/21_18:36:04 info: RealMalloc stats: 39660 total
malloc bytes. pid [4750/HBREAD]
heartbeat[4741]: 2011/07/21_18:36:04 info: Current arena value: 0
heartbeat[4741]: 2011/07/21_18:36:04 info: These are nothing to worry about.

heartbeat[4741]: 2011/07/22_08:30:48 info: Heartbeat shutdown in progress.
(4741)
heartbeat[17136]: 2011/07/22_08:30:48 info: Giving up all HA resources.
ResourceManager[17146]: 2011/07/22_08:30:48 info: Releasing resource group:
shemshak 192.168.1.8/24/eth0
ResourceManager[17146]: 2011/07/22_08:30:48 info: Running
/etc/ha.d/resource.d/IPaddr 192.168.1.8/24/eth0 stop
IPaddr[17204]:  2011/07/22_08:30:48 INFO: /sbin/ifconfig eth0:0 192.168.1.8
down
IPaddr[17183]:  2011/07/22_08:30:48 INFO:  Success
heartbeat[17136]: 2011/07/22_08:30:48 info: All HA resources relinquished.
heartbeat[4741]: 2011/07/22_08:30:49 WARN: 1 lost packet(s) for [dizin]
[134127:134129]
heartbeat[4741]: 2011/07/22_08:30:49 info: No pkts missing from dizin!
heartbeat[4741]: 2011/07/22_08:30:50 info: killing HBFIFO process 4744 with
signal 15
heartbeat[4741]: 2011/07/22_08:30:50 info: killing HBWRITE process 4745 with
signal 15
heartbeat[4741]: 2011/07/22_08:30:50 info: killing HBREAD process 4746 with
signal 15
heartbeat[4741]: 2011/07/22_08:30:50 info: killing HBWRITE process 4747 with
signal 15
heartbeat[4741]: 2011/07/22_08:30:50 info: killing HBREAD process 4748 with
signal 15
heartbeat[4741]: 2011/07/22_08:30:50 info: killing HBWRITE process 4749 with
signal 15
heartbeat[4741]: 2011/07/22_08:30:50 info: killing HBREAD process 4750 with
signal 15
heartbeat[4741]: 2011/07/22_08:30:50 info: Core process 4749 exited. 7
remaining
heartbeat[4741]: 2011/07/22_08:30:50 info: Core process 4747 exited. 6
remaining
heartbeat[4741]: 2011/07/22_08:30:50 info: Core process 4746 exited. 5
remaining
heartbeat[4741]: 2011/07/22_08:30:50 info: Core process 4745 exited. 4
remaining
heartbeat[4741]: 2011/07/22_08:30:50 info: Core process 4744 exited. 3
remaining
heartbeat[4741]: 2011/07/22_08:30:50 info: Core process 4750 exited. 2
remaining
heartbeat[4741]: 2011/07/22_08:30:50 info: Core process 4748 exited. 1
remaining
heartbeat[4741]: 2011/07/22_08:30:51 info: shemshak Heartbeat shutdown
complete.

when I check the log file I don't see the directive stonith_host taking
place!

I know the STONITH demean and the device working as I am able to control the
device by directly issuing the STONITH commands such as:

stonith -t apcmaster -p 192.168.1.56 apc apc -T off Testing
stonith -t apcmaster -p 192.168.1.56 apc apc -T on Testing

Thank you for your help.

Avestan




Nikita Michalko wrote:
 
 Hi Avestan,
 
 do you use really V1/haresource? What version of HA? config?
 We have no crystall ball anymore ;-)
 
 
 Nikita Michalko
 
 
 Am Mittwoch 20 Juli 2011 18:08:56 schrieb Avestan:
 Hello everyone,
 
 I am trying to add a STONITH device into my 

[Linux-HA] logged messages

2011-07-22 Thread Léon Keijser
Hi,

I've set up a cluster and see these messages in syslog:

2011-07-22T19:15:15+02:00 nfs01 exportfs[7765]: INFO: Directory /srv/nfs
is exported to 192.168.100.22/32 (started).
2011-07-22T19:15:15+02:00 nfs01 exportfs[7773]: INFO: Directory /srv/nfs
is exported to 192.168.100.23/32 (started).
2011-07-22T19:15:21+02:00 nfs01 attrd: [19717]: info:
attrd_trigger_update: Sending flush op to all hosts for: ping (0)
2011-07-22T19:15:21+02:00 nfs01 corosync[19703]:   [pcmk  ] ERROR:
send_cluster_msg_raw: Child 7795 spawned to record non-fatal assertion
failure line 1591: rc == 0
2011-07-22T19:15:21+02:00 nfs01 corosync[19703]:   [pcmk  ] ERROR:
send_cluster_msg_raw: Message not sent (-1): attrd_trigger_update
t=attrd src=nfs01 task=flush attr_name=ping
attr_section=status attr_dampening=5s attr
2011-07-22T19:15:21+02:00 nfs01 corosync[19703]:   [pcmk  ] WARN:
route_ais_message: Sending message to all.attrd failed: cluster
delivery failed (rc=-1)
2011-07-22T19:15:25+02:00 nfs01 exportfs[7800]: INFO: Directory /srv/nfs
is exported to 192.168.100.20/32 (started).
2011-07-22T19:15:25+02:00 nfs01 exportfs[7808]: INFO: Directory /srv/nfs
is exported to 192.168.100.22/32 (started).
2011-07-22T19:15:25+02:00 nfs01 exportfs[7816]: INFO: Directory /srv/nfs
is exported to 192.168.100.23/32 (started).


Is this normal? If not, how can i fix it?


My config:



node nfs01 \
attributes standby=off
node nfs02
primitive drbd ocf:linbit:drbd \
params drbd_resource=drbd0 \
op start interval=0 timeout=240 \
op stop interval=0 timeout=100 \
op monitor interval=59s role=Master timeout=30s \
op monitor interval=60s role=Slave timeout=30s
primitive exportfs-shell01 ocf:heartbeat:exportfs \
params directory=/srv/nfs clientspec=192.168.100.20/32
fsid=shell01 \
op start interval=0 timeout=40s
options=rw,async,no_subtree_check,no_root_squash fsid=1 \
op monitor interval=10s timeout=30s \
op start interval=0 timeout=40s \
op stop interval=0 timeout=40s
primitive exportfs-web01 ocf:heartbeat:exportfs \
params directory=/srv/nfs clientspec=192.168.100.22/32 fsid=web01
\
op start interval=0 timeout=40s
options=rw,async,no_subtree_check,no_root_squash fsid=2 \
op monitor interval=10s timeout=30s \
op start interval=0 timeout=40s \
op stop interval=0 timeout=40s
primitive exportfs-web02 ocf:heartbeat:exportfs \
params directory=/srv/nfs clientspec=192.168.100.23/32 fsid=web02
\
op start interval=0 timeout=40s
options=rw,async,no_subtree_check,no_root_squash fsid=2 \
op monitor interval=10s timeout=30s \
op start interval=0 timeout=40s \
op stop interval=0 timeout=40s
primitive failover-ip ocf:heartbeat:IPaddr \
params ip=192.168.100.30 \
op monitor interval=30s
primitive fs-nfs ocf:heartbeat:Filesystem \
params device=/dev/drbd0 directory=/srv/nfs fstype=ext4 \
op start interval=0 timeout=60 \
op stop interval=0 timeout=120
primitive nfs-common lsb:nfs-common \
op monitor interval=10s timeout=30s \
meta target-role=Started
primitive nfs-kernel-server lsb:nfs-kernel-server \
op monitor interval=10s timeout=30s \
meta target-role=Started
primitive ping ocf:pacemaker:ping \
params host_list=192.168.100.1 multiplier=100 name=ping \
op monitor interval=20s timeout=60s \
op start interval=0 timeout=60s
primitive portmap lsb:portmap \
op monitor interval=10s timeout=30s
group nfs portmap nfs-common nfs-kernel-server
group share-nfs fs-nfs exportfs-shell01 exportfs-web01 exportfs-web02
failover-ip \
meta target-role=Started
ms ms_drbd drbd \
meta master-max=1 master-node-max=1 clone-max=2
clone-node-max=1 notify=true
clone nfs_clone nfs \
meta globally-unique=false target-role=Started
location share-nfs_on_connected_node share-nfs \
rule $id=share-nfs_on_connected_node-rule -inf: not_defined ping or
ping lte 0
colocation share-nfs_on_ms_drbd inf: share-nfs ms_drbd:Master
order share-nfs_after_ms_drbd inf: ms_drbd:promote share-nfs:start
property $id=cib-bootstrap-options \
dc-version=1.0.9-74392a28b7f31d7ddc86689598bd23114f58978b \
cluster-infrastructure=openais \
expected-quorum-votes=2 \
stonith-enabled=false \
no-quorum-policy=ignore \
last-lrm-refresh=1311232624



kind regards,

Léon

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

[Linux-HA] Does HA monitor its resources?

2011-07-22 Thread Hai Tao

Does HA monitor its resources? If I manually disable the floating IP, (for 
example, ifdown eth0:0), will HA be able to detect that and re-enable the IP?


Thanks.
 
Hai Tao
 


  
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Does HA monitor its resources?

2011-07-22 Thread mike
It should re-enable it yes. Incidentally, you can set the number of 
times it will do this in your cib.xml file in /var/lib/heartbeat/crm.

For instance, on one of my clusters I only want the resource re-started 
3 times. On a fourth failure I want it to do a failover. I set it with 
this entry in the cib.xml:

nvpair id=nvpair.id17897906 
name=default-resource-failure-stickiness value=50/

Hope this helps.

-mike


On 11-07-22 02:55 PM, Hai Tao wrote:
 Does HA monitor its resources? If I manually disable the floating IP, (for 
 example, ifdown eth0:0), will HA be able to detect that and re-enable the IP?


 Thanks.

 Hai Tao



   
 ___
 Linux-HA mailing list
 Linux-HA@lists.linux-ha.org
 http://lists.linux-ha.org/mailman/listinfo/linux-ha
 See also: http://linux-ha.org/ReportingProblems


___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Does HA monitor its resources?

2011-07-22 Thread Hai Tao

How can I disable HA without stopping the resources then?

I like to disable HA by stopping the heartbeat, but once I do that a failover 
will happen, but I do not want that.  

Thanks.
 
Hai Tao
 


 Date: Fri, 22 Jul 2011 17:23:58 -0300
 From: mgbut...@nbnet.nb.ca
 To: linux-ha@lists.linux-ha.org
 Subject: Re: [Linux-HA] Does HA monitor its resources?
 
 It should re-enable it yes. Incidentally, you can set the number of 
 times it will do this in your cib.xml file in /var/lib/heartbeat/crm.
 
 For instance, on one of my clusters I only want the resource re-started 
 3 times. On a fourth failure I want it to do a failover. I set it with 
 this entry in the cib.xml:
 
 nvpair id=nvpair.id17897906 
 name=default-resource-failure-stickiness value=50/
 
 Hope this helps.
 
 -mike
 
 
 On 11-07-22 02:55 PM, Hai Tao wrote:
  Does HA monitor its resources? If I manually disable the floating IP, (for 
  example, ifdown eth0:0), will HA be able to detect that and re-enable the 
  IP?
 
 
  Thanks.
 
  Hai Tao
 
 
 
  
  ___
  Linux-HA mailing list
  Linux-HA@lists.linux-ha.org
  http://lists.linux-ha.org/mailman/listinfo/linux-ha
  See also: http://linux-ha.org/ReportingProblems
 
 
 ___
 Linux-HA mailing list
 Linux-HA@lists.linux-ha.org
 http://lists.linux-ha.org/mailman/listinfo/linux-ha
 See also: http://linux-ha.org/ReportingProblems
  
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems