Re: [Linux-ha-dev] [Patch 3]Change avoiding the stop error of the mysql resource agent.

2011-09-14 Thread Raoul Bhatia [IPAX]
Hi Hideo-san!


On 09/12/2011 03:19 AM, renayama19661...@ybb.ne.jp wrote:
 Hi Raoul,

 How about the modified patch of this place?

unfortunatly, i'm currently a little swamped.
i'll look into that asap. please bear with me a little longer!

thanks,
raoul
___
Linux-HA-Dev: Linux-HA-Dev@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha-dev
Home Page: http://linux-ha.org/


[Linux-HA] Resource fail and node fence

2011-09-14 Thread RaSca
Hi all,
I've got a two node pacemaker/corosync cluster with some virtual domain 
resources on some DRBD devices.
Every DRBD device is configured in dual primary setup and I have enabled 
the live migration. Cluster has also stonith enabled.

My problem is that if a live migration for a single virtualdomain 
resource fails, then this node gets fenced, making unavailable also all 
the other virtual machines (that gets restarted on the other node after 
a poweroff).

As I saw the way to make a single resource fail not fencing the node 
where it fails is to declare an on-fail=restart option for the virtual 
domain. Is it the correct approach or is there a more elegant way to 
obtain what I want?

Thanks to all,

-- 
RaSca
Mia Mamma Usa Linux: Niente è impossibile da capire, se lo spieghi bene!
ra...@miamammausalinux.org
http://www.miamammausalinux.org

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] What's wrong in my configuration for GFS2 under Pacemaker ?

2011-09-14 Thread alain . moulle
Hi,

only for information : 
It works fine now, it was just due to the fact that on chili1 , gfs2-utils 
rpm was not installed,
and it is strange that I succeeded to mount the FS gfs2 without this rpm 
installed which 
install /sbin/mount.gfs2 ...
And perhaps just a dependancy is missing in rpm 
gfs-pcmk-3.0.12-23.el6.Bull.1.x86_64 on gfs2-utils.

Regards
Alain




De :alain.mou...@bull.net
A : linux-ha@lists.linux-ha.org
Date :  13/09/2011 08:59
Objet : [Linux-HA] What's wrong in my configuration for GFS2 under 
Pacemaker ?
Envoyé par :linux-ha-boun...@lists.linux-ha.org



Hi,

( Pacemaker 1.1.2.7 on RHEL6 )

I just tried to configure a GFS2 Filesystem under Pacemaker, but
the FS is always mounted only on one side, not both sides :

Clone Set: dlm-clone [dlm]
 Started: [ chili0 chili1 ]
 Clone Set: gfs-clone [gfs-control]
 Started: [ chili0 chili1 ]
 Clone Set: fsGFS2 [fs-GFS2]
 Started: [ chili1 ]
 Stopped: [ fs-GFS2:0 ]

and after a while the node chili0 where the FS mount fails is fenced by 
chili1

Is there something missing or wrong in my configuration below ?
or is there an already known issue on this subject ?
Thanks
Alain Moullé

My configuration around gfs2 items:
node chili0
node chili1
primitive dlm ocf:pacemaker:controld \
op start interval=0 timeout=90s \
op stop interval=0 timeout=100 \
op monitor interval=120s
primitive fs-GFS2 ocf:heartbeat:Filesystem \
params device=/dev/sdl directory=/mnt/gfs2 fstype=gfs2 \
op start interval=0 timeout=60 \
op stop interval=0 timeout=60 \
op monitor interval=20 timeout=40 \
primitive gfs-control ocf:pacemaker:controld \
params daemon=gfs_controld.pcmk args=-g 0 \
op monitor interval=120s \
op start interval=0 timeout=90s \
op stop interval=0 timeout=120s
clone dlm-clone dlm \
meta interleave=true
clone fsGFS2 fs-GFS2
clone gfs-clone gfs-control \
meta interleave=true
colocation gfs-with-dlm inf: gfs-clone dlm-clone
order start-gfs-after-dlm inf: dlm-clone gfs-clone


___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] How to configure only logs warning or critical

2011-09-14 Thread alain . moulle
Hi

I wonder if someone could list where I can tune the log level info for all 
softwares involved when using the stack Pacemaker/corosync .
I set debug:off in corosync.conf but there are still lots of logs 
information
in syslog , generated by lrmd etc.

Is it possible to have the exhaustive list of software levels where we can
tune the log level ?

Thanks
Regards
Alain
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] heartbeat and squid

2011-09-14 Thread Nicolas Repentin
Hi all,

I've got a question for heartbeat. 
How can I made this :

If squid stop or be killed on node1, how make node2 be master ?

Actually, node2 become master only when node1 is down, or heartbeat
service on node1 is down, but if I kill squid, nothing happen.

I'm using Centos 6 and last heartbeat version.

Thanks a lot for your responses !


-- 
Nicolas
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] heartbeat and squid

2011-09-14 Thread Dejan Muhamedagic
Hi,

On Thu, Sep 01, 2011 at 06:30:46PM +0200, Nicolas Repentin wrote:
 Hi all,
 
 I've got a question for heartbeat. 
 How can I made this :
 
 If squid stop or be killed on node1, how make node2 be master ?
 
 Actually, node2 become master only when node1 is down, or heartbeat
 service on node1 is down, but if I kill squid, nothing happen.
 
 I'm using Centos 6 and last heartbeat version.

Using just heartbeat and no pacemaker? Only pacemaker has service
monitoring.

Thanks,

Dejan

 Thanks a lot for your responses !
 
 
 -- 
 Nicolas
 ___
 Linux-HA mailing list
 Linux-HA@lists.linux-ha.org
 http://lists.linux-ha.org/mailman/listinfo/linux-ha
 See also: http://linux-ha.org/ReportingProblems
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] only one node complains heartbeat over /dev/ttyS0

2011-09-14 Thread Hai Tao

anyone has thoughts on it?


Thanks.
 
Hai Tao
 


 From: taoh...@hotmail.com
 To: linux-ha@lists.linux-ha.org
 Date: Mon, 12 Sep 2011 15:49:33 -0700
 Subject: [Linux-HA] only one node complains heartbeat over /dev/ttyS0
 
 
 I have a two node HA, and heartbeat is configured using a network LAN and a 
 serial cable.
 
 However, I keep seeing this message in 02:
 
 WARN: glib: TTY write timeout on [/dev/ttyS0] (no connection or bad cable? 
 [see documentation])
 
 But on 01, everthing is ok:
 
 info: These are nothing to worry about.
 
 I have checked that both nodes have heartbeat running. My question is if 01 
 is receiving heartbeat from 02, and this is over a serial link, why 02 cannot 
 receive heartbeat from 01 over /dev/ttyS0?
 
 
 Thanks.
 
 Hai Tao 
 ___
 Linux-HA mailing list
 Linux-HA@lists.linux-ha.org
 http://lists.linux-ha.org/mailman/listinfo/linux-ha
 See also: http://linux-ha.org/ReportingProblems
  
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] howto use heartbeat API

2011-09-14 Thread 4070421
Hi,all
 I am program with heartbeat3.0.4, I have install it for three node,
now I want to use heartbeat local IPC and cluster IPC to communicate with other 
node .
   And my question is  how to compile, it always say that can't find 
hb_api.h ...hb_api_core.h.. how to wirte the makefile ?
And  how tu use local IPC and cluster IPC api  to communicate ?
one node want send message to other node , i have to code a  CLI and a Daemon,  
daemon get Command from CLI from local IPC, and send message (the command from 
CLI) to other node via cluster IPC,  HOW  to do it
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] howto use heartbeat API

2011-09-14 Thread Lars Ellenberg
On Thu, Sep 15, 2011 at 01:47:33AM +0800, 4070421 wrote:
 Hi,all
  I am program with heartbeat3.0.4, I have install it for three node,
 now I want to use heartbeat local IPC and cluster IPC to communicate with 
 other node .
And my question is  how to compile, it always say that can't find 
 hb_api.h ...hb_api_core.h.. how to wirte the makefile ?
 And  how tu use local IPC and cluster IPC api  to communicate ?
 one node want send message to other node , i have to code a  CLI and a 
 Daemon,  daemon get Command from CLI from local IPC, and send message (the 
 command from CLI) to other node via cluster IPC,  HOW  to do it

Since the project is no longer one monolithic source repo.
you need the cluster-glue sources/libraries/headers as well.
(http://hg.linux-ha.org/glue)

In the heartbeat source tree, there are
 lib/hbclient/api_test.c
 membership/ccm/ccm_testclient.c

which may help you get going.

I'm curious, what is your project about?

If you stay with heartbeat, I strongly recommend to use latest mercurial
(which is one changeset beyond 3.0.5).  http://hg.linux-ha.org/dev/shortlog

3.0.4 to 3.0.5 contains only a handful of changesets.
But they fix very bad behaviour in case you should ever
have to deal with packet loss, for whatever reason.

Have you considered using corosync instead?
I'm not trying to push you either way, 
I'd just like to know what alternatives have been considered,
and what you based your decision for heartbeat on.

Cheers,

-- 
: Lars Ellenberg
: LINBIT | Your Way to High Availability
: DRBD/HA support and consulting http://www.linbit.com
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] Strange behavior ldirectord direct routing

2011-09-14 Thread mike
Hope someone here can give me some pointers. I've set up a ldirectord 
LinuxHA cluster.

When I start it up its a simple set up. I have one VIP pointing to 2 
backend servers and everything seems to come up ok.

Output of ip addr: (weird set up I know - not my network tho just trying 
to get this working)
1: lo: LOOPBACK,UP,LOWER_UP mtu 16436 qdisc noqueue
 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
 inet 127.0.0.1/8 scope host lo
2: eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc pfifo_fast 
qlen 1000
 link/ether 5c:f3:fc:78:c9:98 brd ff:ff:ff:ff:ff:ff
3: eth1: BROADCAST,MULTICAST mtu 1500 qdisc noop qlen 1000
 link/ether 5c:f3:fc:78:c9:9a brd ff:ff:ff:ff:ff:ff
4: usb0: BROADCAST,MULTICAST mtu 1500 qdisc noop qlen 1000
 link/ether 5e:f3:fc:7b:c9:9b brd ff:ff:ff:ff:ff:ff
5: eth0.101@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue
 link/ether 5c:f3:fc:78:c9:98 brd ff:ff:ff:ff:ff:ff
 inet 172.28.89.46/25 brd 172.28.89.127 scope global eth0.101
 inet 172.28.191.155/25 brd 172.28.191.127 scope global eth0.101 
--- This is the VIP

The VIP above points to the backend servers 172.28.191.139 and 
172.28.191.140.

Without HA running I can telnet to the backend servers to the proper 
port (8080 in this case). As soon as I start up HA and ldirector, I am 
unable to telnet to the backend servers on port 8080 on the node that is 
primary. I can telnet to them on the backup node and on any other server 
on the network but not from the primary node.

I don't have access to the backend servers but the SA tells me he set up 
the loopback with an alias to the VIP as I asked and he also said he 
added the appropriate entries to prevent the arp issues into sysctl.conf 
on the backend servers.

So what the heck is going on? Why does HA suddenly make it impossible to 
connect to the backend servers from the primary node?

Another problem I have is that I cannot telnet to the VIP on port 8080 
no matter what node it is running on. I think if I can resolve the 
problem above, this one will go away too.

Thanks
-mike

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Strange behavior ldirectord direct routing

2011-09-14 Thread mike
Replying to my own post here but the fact that the VIP is on a different 
subnet has me curious if this could be an issue:

ip addr show eth0.101
5: eth0.101@eth0: BROADCAST,MULTICAST,UP,LOWER_UP mtu 1500 qdisc noqueue
 link/ether 5c:f3:fc:78:c9:98 brd ff:ff:ff:ff:ff:ff
 inet 172.28.89.46/25 brd 172.28.89.127 scope global eth0.101
 inet 172.28.191.155/25 brd 172.28.191.127 scope global eth0.101 -- VIP


On 11-09-14 06:37 PM, mike wrote:
 Hope someone here can give me some pointers. I've set up a ldirectord
 LinuxHA cluster.

 When I start it up its a simple set up. I have one VIP pointing to 2
 backend servers and everything seems to come up ok.

 Output of ip addr: (weird set up I know - not my network tho just trying
 to get this working)
 1: lo:LOOPBACK,UP,LOWER_UP  mtu 16436 qdisc noqueue
   link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
   inet 127.0.0.1/8 scope host lo
 2: eth0:BROADCAST,MULTICAST,UP,LOWER_UP  mtu 1500 qdisc pfifo_fast
 qlen 1000
   link/ether 5c:f3:fc:78:c9:98 brd ff:ff:ff:ff:ff:ff
 3: eth1:BROADCAST,MULTICAST  mtu 1500 qdisc noop qlen 1000
   link/ether 5c:f3:fc:78:c9:9a brd ff:ff:ff:ff:ff:ff
 4: usb0:BROADCAST,MULTICAST  mtu 1500 qdisc noop qlen 1000
   link/ether 5e:f3:fc:7b:c9:9b brd ff:ff:ff:ff:ff:ff
 5: eth0.101@eth0:BROADCAST,MULTICAST,UP,LOWER_UP  mtu 1500 qdisc noqueue
   link/ether 5c:f3:fc:78:c9:98 brd ff:ff:ff:ff:ff:ff
   inet 172.28.89.46/25 brd 172.28.89.127 scope global eth0.101
   inet 172.28.191.155/25 brd 172.28.191.127 scope global eth0.101
 --- This is the VIP

 The VIP above points to the backend servers 172.28.191.139 and
 172.28.191.140.

 Without HA running I can telnet to the backend servers to the proper
 port (8080 in this case). As soon as I start up HA and ldirector, I am
 unable to telnet to the backend servers on port 8080 on the node that is
 primary. I can telnet to them on the backup node and on any other server
 on the network but not from the primary node.

 I don't have access to the backend servers but the SA tells me he set up
 the loopback with an alias to the VIP as I asked and he also said he
 added the appropriate entries to prevent the arp issues into sysctl.conf
 on the backend servers.

 So what the heck is going on? Why does HA suddenly make it impossible to
 connect to the backend servers from the primary node?

 Another problem I have is that I cannot telnet to the VIP on port 8080
 no matter what node it is running on. I think if I can resolve the
 problem above, this one will go away too.

 Thanks
 -mike

 ___
 Linux-HA mailing list
 Linux-HA@lists.linux-ha.org
 http://lists.linux-ha.org/mailman/listinfo/linux-ha
 See also: http://linux-ha.org/ReportingProblems


___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems