Re: [Pacemaker] errors in corosync.log

2010-01-18 Thread Andrew Beekhof
On Sat, Jan 16, 2010 at 9:20 PM, Shravan Mishra shravan.mis...@gmail.com wrote: Hi Guys, I'm running the following version of pacemaker and corosync corosync=1.1.1-1-2 pacemaker=1.0.9-2-1 Every thing had been running fine for quite some time now but then I started seeing following errors

Re: [Pacemaker] Split Site 2-way clusters

2010-01-18 Thread Andrew Beekhof
On Thu, Jan 14, 2010 at 11:44 PM, Miki Shapiro miki.shap...@coles.com.au wrote: Confused. I *am* running DRBD in dual-master mode /me cringes... this sounds to me like an impossibly dangerous idea. Can someone from linbit comment on this please? Am I imagining this? (apologies, I should

Re: [Pacemaker] Pre-Announce: End of 0.6 support is near

2010-01-18 Thread Andrew Beekhof
On Tue, Jan 12, 2010 at 3:55 PM, Emmanuel Lesouef e.leso...@crbn.fr wrote: Le Tue, 12 Jan 2010 14:56:31 +0100, Michael Schwartzkopff mi...@multinet.de a écrit : Am Dienstag, 12. Januar 2010 14:48:12 schrieb Emmanuel Lesouef: Hi, We use a rather old (in fact, very old) combination :

Re: [Pacemaker] DC election with downed node in 2-way cluster

2010-01-18 Thread Andrew Beekhof
On Thu, Jan 14, 2010 at 4:40 AM, Miki Shapiro miki.shap...@coles.com.au wrote: And the node really did power down? Yes. 100% certain and positive. OFF. But the other node didn't notice?!? Its resources (drbd master and the fence clone) did notice. Its dc-election-mechanism did NOT notice

Re: [Pacemaker] Split Site 2-way clusters

2010-01-18 Thread Florian Haas
On 2010-01-18 11:41, Colin wrote: Hi All, we are currently looking at nearly the same issue, in fact I just wanted to start a similarly titled thread when I stumbled over these messages… The setup we are evaluating is actually a 2*N-node-cluster, i.e. two slightly separated sites with N

Re: [Pacemaker] Pre-Announce: End of 0.6 support is near

2010-01-18 Thread Florian Haas
On 2010-01-18 11:18, Andrew Beekhof wrote: Biggest caveat is the networking issue that makes pacemaker 1.0 wire-incompatible with pacemaker 0.6 (and heartbeat 2.1.x). So rolling upgrades are out and you'd need to look at one of the other upgrade strategies. Even though I've bugged you about

Re: [Pacemaker] [Linux-HA] Announce: Hawk (HA Web Konsole)

2010-01-18 Thread Andrew Beekhof
I look forward to taking this for a spin! Do we have a bugzilla component for it yet? On Sat, Jan 16, 2010 at 2:14 PM, Tim Serong tser...@novell.com wrote: Greetings All, This is to announce the development of the Hawk project, a web-based GUI for Pacemaker HA clusters. So, why another

[Pacemaker] Announce: Pacemaker 1.0.7 (stable) Released

2010-01-18 Thread Andrew Beekhof
The latest installment of the Pacemaker 1.0 stable series is now ready for general consumption. In this release, we’ve made a number improvements to clone handling - particularly the way ordering constraints are processed - as well as some really nice improvements to the shell. The next 1.0

Re: [Pacemaker] Split Site 2-way clusters

2010-01-18 Thread Colin
On Mon, Jan 18, 2010 at 11:52 AM, Florian Haas florian.h...@linbit.com wrote: the current approach is to utilize 2 Pacemaker clusters, each highly available in its own right, and employing manual failover. As described here: Thanks for the pointer! Perhaps site is not quite the correct term

Re: [Pacemaker] Announce: Pacemaker 1.0.7 (stable) Released

2010-01-18 Thread Andreas Mock
-Ursprüngliche Nachricht- Von: Andrew Beekhof and...@beekhof.net Gesendet: 18.01.10 12:43:30 An: The Pacemaker cluster resource manager pacemaker@oss.clusterlabs.org Betreff: [Pacemaker] Announce: Pacemaker 1.0.7 (stable) Released The latest installment of the Pacemaker 1.0 stable

Re: [Pacemaker] errors in corosync.log

2010-01-18 Thread Shravan Mishra
Hi , Since the interfaces on the two nodes are connected via cross over cable so there is no chance of that happening and since I'm using rrp: passive, which means that the other ring i.e. ring 1 will come into play only when ring 0 fails,I assume. I say this because ring 1 interface is on the

[Pacemaker] mcast vs broadcast

2010-01-18 Thread Shravan Mishra
Hi all, Following is my corosync.conf. Even though broadcast is enabled I see mcasted messages like these in corosync.log. Is it ok? even when the broadcast is on and not mcast. == Jan 18 09:50:40 corosync [TOTEM ] mcasted message added to pending queue Jan 18 09:50:40 corosync [TOTEM ]

[Pacemaker] 1.0.7 upgraded, restarting resources problem

2010-01-18 Thread Martin Gombač
Hi, i have one m/s drbd resource and one Xen instance on top. Both m/s are primary. When i restart node that's _not_ hosting the Xen instance (ibm1), pacemaker restarts running Xen instance on the other node (ibm2). There is no need to do that. I thought it got fixed

Re: [Pacemaker] errors in corosync.log

2010-01-18 Thread Steven Dake
One possibility is you have a different cluster in your network on the same multicast address and port. Regards -steve On Sat, 2010-01-16 at 15:20 -0500, Shravan Mishra wrote: Hi Guys, I'm running the following version of pacemaker and corosync corosync=1.1.1-1-2 pacemaker=1.0.9-2-1

Re: [Pacemaker] mcast vs broadcast

2010-01-18 Thread Steven Dake
On Mon, 2010-01-18 at 11:25 -0500, Shravan Mishra wrote: Hi all, Following is my corosync.conf. Even though broadcast is enabled I see mcasted messages like these in corosync.log. Is it ok? even when the broadcast is on and not mcast. Yes you are using broadcast and the debug