On Sat, Jan 16, 2010 at 9:20 PM, Shravan Mishra
shravan.mis...@gmail.com wrote:
Hi Guys,
I'm running the following version of pacemaker and corosync
corosync=1.1.1-1-2
pacemaker=1.0.9-2-1
Every thing had been running fine for quite some time now but then I
started seeing following errors
On Thu, Jan 14, 2010 at 11:44 PM, Miki Shapiro
miki.shap...@coles.com.au wrote:
Confused.
I *am* running DRBD in dual-master mode
/me cringes... this sounds to me like an impossibly dangerous idea.
Can someone from linbit comment on this please? Am I imagining this?
(apologies, I should
On Tue, Jan 12, 2010 at 3:55 PM, Emmanuel Lesouef e.leso...@crbn.fr wrote:
Le Tue, 12 Jan 2010 14:56:31 +0100,
Michael Schwartzkopff mi...@multinet.de a écrit :
Am Dienstag, 12. Januar 2010 14:48:12 schrieb Emmanuel Lesouef:
Hi,
We use a rather old (in fact, very old) combination :
On Thu, Jan 14, 2010 at 4:40 AM, Miki Shapiro miki.shap...@coles.com.au wrote:
And the node really did power down?
Yes. 100% certain and positive. OFF.
But the other node didn't notice?!?
Its resources (drbd master and the fence clone) did notice.
Its dc-election-mechanism did NOT notice
On 2010-01-18 11:41, Colin wrote:
Hi All,
we are currently looking at nearly the same issue, in fact I just
wanted to start a similarly titled thread when I stumbled over these
messages…
The setup we are evaluating is actually a 2*N-node-cluster, i.e. two
slightly separated sites with N
On 2010-01-18 11:18, Andrew Beekhof wrote:
Biggest caveat is the networking issue that makes pacemaker 1.0
wire-incompatible with pacemaker 0.6 (and heartbeat 2.1.x).
So rolling upgrades are out and you'd need to look at one of the other
upgrade strategies.
Even though I've bugged you about
I look forward to taking this for a spin!
Do we have a bugzilla component for it yet?
On Sat, Jan 16, 2010 at 2:14 PM, Tim Serong tser...@novell.com wrote:
Greetings All,
This is to announce the development of the Hawk project,
a web-based GUI for Pacemaker HA clusters.
So, why another
The latest installment of the Pacemaker 1.0 stable series is now ready for
general consumption.
In this release, we’ve made a number improvements to clone handling -
particularly the way ordering constraints are processed - as well as some
really nice improvements to the shell.
The next 1.0
On Mon, Jan 18, 2010 at 11:52 AM, Florian Haas florian.h...@linbit.com wrote:
the current approach is to utilize 2 Pacemaker clusters, each highly
available in its own right, and employing manual failover. As described
here:
Thanks for the pointer! Perhaps site is not quite the correct term
-Ursprüngliche Nachricht-
Von: Andrew Beekhof and...@beekhof.net
Gesendet: 18.01.10 12:43:30
An: The Pacemaker cluster resource manager pacemaker@oss.clusterlabs.org
Betreff: [Pacemaker] Announce: Pacemaker 1.0.7 (stable) Released
The latest installment of the Pacemaker 1.0 stable
Hi ,
Since the interfaces on the two nodes are connected via cross over
cable so there is no chance of that happening and since I'm using rrp:
passive, which means that the other ring i.e. ring 1 will come into
play only when ring 0 fails,I assume. I say this because ring 1
interface is on the
Hi all,
Following is my corosync.conf.
Even though broadcast is enabled I see mcasted messages like these
in corosync.log.
Is it ok? even when the broadcast is on and not mcast.
==
Jan 18 09:50:40 corosync [TOTEM ] mcasted message added to pending queue
Jan 18 09:50:40 corosync [TOTEM ]
Hi,
i have one m/s drbd resource and one Xen instance on top. Both m/s are
primary.
When i restart node that's _not_ hosting the Xen instance (ibm1),
pacemaker restarts running Xen instance on the other node (ibm2). There
is no need to do that. I thought it got fixed
One possibility is you have a different cluster in your network on the
same multicast address and port.
Regards
-steve
On Sat, 2010-01-16 at 15:20 -0500, Shravan Mishra wrote:
Hi Guys,
I'm running the following version of pacemaker and corosync
corosync=1.1.1-1-2
pacemaker=1.0.9-2-1
On Mon, 2010-01-18 at 11:25 -0500, Shravan Mishra wrote:
Hi all,
Following is my corosync.conf.
Even though broadcast is enabled I see mcasted messages like these
in corosync.log.
Is it ok? even when the broadcast is on and not mcast.
Yes you are using broadcast and the debug
15 matches
Mail list logo