Sorry William
But if you wanna implement dual primary i think you don't nee promote for
your drbd
Try to use clone without master/slave
2012/1/30 William Seligman selig...@nevis.columbia.edu
I'm trying to follow the directions for setting up a dual-primary DRBD
setup
with CMAN and Pacemaker.
Krille wrote:
On Tuesday 31 January 2012 00:12:52 emmanuel segura wrote:
But if you wanna implement dual primary i think you don't nee promote
for
your drbd
Try to use clone without master/slave
At least when you use the linbit-ra, using it without a master-clone
will give
you one
William can you try like this
primitive AdminDrbd ocf:linbit:drbd \
params drbd_resource=admin \
op monitor interval=60s role=Master
clone Adming AdminDrbd
2012/1/31 William Seligman selig...@nevis.columbia.edu
On 1/31/12 3:47 PM, emmanuel segura wrote:
William try to follow
Sorry But can we see your configuration?
2012/2/2 alain.mou...@bull.net
Hi
Don't remember in details, it was at the end of 2010 ... but :
Why pacemaker is fencing a node ?
because it was one of my simple HA test : for example, make the heartbeat
no more working so
that pacemaker fences
colocation altogether inf: apache mysql drbd_fs drbd_ms:Master
2012/2/10 Ryan Stepalavich rstepalav...@gmail.com
I'm using Pacemaker to handle my cluster resources (on top of heartbeat).
Everything works except the collocation parameter. I want all of my
resources to stay on the same node at
can you show me your /etc/cluster/cluster.conf?
because i think your problem it's a fencing-loop
Il giorno 01 marzo 2012 01:03, William Seligman selig...@nevis.columbia.edu
ha scritto:
On 2/28/12 7:26 PM, Lars Ellenberg wrote:
On Tue, Feb 28, 2012 at 03:51:29PM -0500, William Seligman
...@nevis.columbia.edu
ha scritto:
On 3/1/12 4:15 AM, emmanuel segura wrote:
can you show me your /etc/cluster/cluster.conf?
because i think your problem it's a fencing-loop
Here it is:
/etc/cluster/cluster.conf:
?xml version=1.0?
cluster config_version=17 name=Nevis_HA
logging debug=off
Ok william
if this it'sn the problem, when you show me your pacemaker cib xml
crm configure show OUTPUT
Il giorno 01 marzo 2012 18:10, William Seligman selig...@nevis.columbia.edu
ha scritto:
On 3/1/12 6:34 AM, emmanuel segura wrote:
try to change the fence daemon tag like
are you sure the exportfs agent can be use it with clone active/active?
Il giorno 03 marzo 2012 00:12, William Seligman selig...@nevis.columbia.edu
ha scritto:
One step forward, two steps back.
I'm working on a two-node primary-primary cluster. I'm debugging problems
I have
with the
Sorry Willian
But i think clvmd it must be used with
ocf:lvm2:clvmd
esample
crm confgiure
primitive clvmd ocf:lvm2:clvmd params daemon_timeout=30
clone cln_clvmd clvmd
and rember clvmd depend on dlm, so for the dlm you sould same
Il giorno 13 marzo 2012 17:29, William Seligman
Hello Willian
So if you using cman why you use lsb::clvmd
I think you are very confused
Il giorno 13 marzo 2012 22:42, William Seligman selig...@nevis.columbia.edu
ha scritto:
On 3/13/12 12:29 PM, William Seligman wrote:
I'm not sure if this is a Linux-HA question; please direct me to
Hello William
I think it's better you make clvmd start at boot
chkconfig cman on ; chkconfig clvmd on
Il giorno 13 marzo 2012 23:29, William Seligman selig...@nevis.columbia.edu
ha scritto:
On 3/13/12 5:50 PM, emmanuel segura wrote:
So if you using cman why you use lsb::clvmd
I
6:02 AM, emmanuel segura wrote:
I think it's better you make clvmd start at boot
chkconfig cman on ; chkconfig clvmd on
I've already tried it. It doesn't work. The problem is that my LVM
information is on the drbd. If I start up clvmd before drbd, it won't find
the logical volumes.
I
for cman+pacemaker you don't need qdisk
Il giorno 14 marzo 2012 23:17, William Seligman selig...@nevis.columbia.edu
ha scritto:
On 3/14/12 9:20 AM, emmanuel segura wrote:
Hello William
i did new you are using drbd and i dont't know what type of configuration
you using
But it's better you
yes william
Now try clvmd -d and see what happen
locking_type = 3 it's lvm cluster lock type
Il giorno 15 marzo 2012 16:15, William Seligman selig...@nevis.columbia.edu
ha scritto:
On 3/15/12 5:18 AM, emmanuel segura wrote:
The first thing i seen in your clvmd log it's
Hello William
Ho did you created your volume group
give me the output of vgs command when the cluster it's up
Il giorno 15 marzo 2012 17:06, William Seligman selig...@nevis.columbia.edu
ha scritto:
On 3/15/12 11:50 AM, emmanuel segura wrote:
yes william
Now try clvmd -d and see what
scritto:
On 3/15/12 12:15 PM, emmanuel segura wrote:
Ho did you created your volume group
pvcreate /dev/drbd0
vgcreate -c y ADMIN /dev/drbd0
lvcreate -L 200G -n usr ADMIN # ... and so on
# Nevis-HA is the cluster name I used in cluster.conf
mkfs.gfs2 -p lock_dlm -j 2 -t Nevis_HA:usr /dev
, emmanuel segura wrote:
I don't see any error and the answer for your question it's yes
can you show me your /etc/cluster/cluster.conf and your crm configure
show
like that more later i can try to look if i found some fix
Thanks for taking a look.
My cluster.conf: http://pastebin.com
===
Il giorno 15 marzo 2012 23:50, William Seligman selig...@nevis.columbia.edu
ha scritto:
On 3/15/12 6:07 PM, William Seligman wrote:
On 3/15/12 6:05 PM, William Seligman wrote:
On 3/15/12 4:57 PM, emmanuel segura wrote:
we can try to understand what happen when clvm hang
:04PM -0400, William Seligman wrote:
On 3/16/12 12:12 PM, William Seligman wrote:
On 3/16/12 7:02 AM, Andreas Kurz wrote:
On 03/15/2012 11:50 PM, William Seligman wrote:
On 3/15/12 6:07 PM, William Seligman wrote:
On 3/15/12 6:05 PM, William Seligman wrote:
On 3/15/12 4:57 PM, emmanuel
How do you configure clvmd?
with cman or with pacemaker?
Il giorno 23 marzo 2012 22:14, William Seligman selig...@nevis.columbia.edu
ha scritto:
On 3/23/12 5:03 PM, emmanuel segura wrote:
Sorry but i would to know if can show me your /etc/cluster/cluster.conf
Here it is: http
a problems
Il giorno 24 marzo 2012 13:09, William Seligman selig...@nevis.columbia.edu
ha scritto:
On 3/24/12 4:47 AM, emmanuel segura wrote:
How do you configure clvmd?
with cman or with pacemaker?
Pacemaker. Here's the output of 'crm configure show':
http://pastebin.com/426CdVwN
Il
:
On 3/26/12 4:28 PM, emmanuel segura wrote:
Sorry Willian i can't post my config now because i'm at home now not in
my
job
I think it's no a problem if clvm start before drbd, because clvm not
needed and devices to start
This it's the point, i hope to be clear
The introduction
William :-)
So now your cluster it's OK?
Il giorno 27 marzo 2012 00:33, William Seligman selig...@nevis.columbia.edu
ha scritto:
On 3/26/12 5:31 PM, William Seligman wrote:
On 3/26/12 5:17 PM, William Seligman wrote:
On 3/26/12 4:28 PM, emmanuel segura wrote:
and i suggest you to start
If i rember well, for use Ilo 3 card you sould use the cluster agent
ipmilan
Il giorno 11 aprile 2012 23:00, Pham, Tom tom.p...@viasat.com ha scritto:
Hi everyone,
I try to test a 2 nodes cluster with stonith resource using external/ipmi
( I tried external/riloe first but it does not seem
Maybe the problem it's the primitive nfsserver lsb:nfs-kernel-server, i
think this primitive was stoped befoure exportfs-admin
ocf:heartbeat:exportfs
And if i rember the lsb:nfs-kernel-server and exportfs agent does the same
thing
the first use the os scripts and the second the cluster agents
Why you are using cman corosync together?
I think you should use cman+pacemaker or corosync+pacemaker
2012/6/9 Yount, William D yount.will...@menloworldwide.com
I have two servers which are both Dell 990's. Each server has two 1tb hard
drives configured in RAID0. I have installed CentOS on
First of all the parameter 201 it must be diferent for every resource
2012/6/19 Martin Marji Cermak cerm...@gmail.com
Hello guys,
I have 3 questions if you please.
I have a HA NFS cluster - Centos 6.2, pacemaker, corosync, two NFS nodes
plus 1 quorum node, in semi Active-Active
Do you have a stonith configured?
2012/7/2 EXTERNAL Konold Martin (erfrakon, RtP2/TEF72)
external.martin.kon...@de.bosch.com
Hi,
when a split brain (drbd) happens mount.ocfs2 remains hanging unkillable
in D-state.
rt-lxcl9a:~ # ps aux | grep ocf
root 347 0.0 0.0 10468 740 ?
remove the standby on node node rt-lxcl9a
2012/7/2 EXTERNAL Konold Martin (erfrakon, RtP2/TEF72)
external.martin.kon...@de.bosch.com
Hi,
Do you have a stonith configured?
Yes. Though a hanging mount does not cause stonith to become activated.
node rt-lxcl9a \
attributes
are you using ext3 for drbd active/active? UM
2012/8/3 Elvis Altherr elvis.alth...@gmail.com
Hello together
On my gentoo servers (2 Node Cluster with kernel 3.x) i use heartbeat
3.0.5 and DRBD 8.4.0 for block replication between the two machines
which served apache, mysql and samba
i know the drbd primary to primary it's for use ocfs/gfs, so for have the
filesystem read write on both nodes, why you still using heartbeat 1.X
2012/8/3 Elvis Altherr elvis.alth...@gmail.com
Am 03.08.2012 09:32, schrieb emmanuel segura:
are you using ext3 for drbd active/active? UM
1: check if the services are configured to start at boot time
2:without info nobody can help you
2012/7/21 Chirag Vaishnav chirag.vaish...@saicare.com
Hi,
We are HA between two nodes, everything is configured as per standard
example file (using haresources) and everything works well
I don't know if ping is rigth for your case, try to look here
http://doc.opensuse.org/products/draft/SLE-HA/SLE-ha-guide_sd_draft/cha.ha.geo.html
2013/1/31 Robinson, Eric eric.robin...@psmnv.com
We have this configuration:
NodeA is located in DataCenterA. NodeB is located in (geographically
Hello Fedrik
Why you have a clone of cl_exportfs_root and you have ext4 filesystem, and
i think this order is not correct
order o_drbd_before_nfs inf: ms_drbd_nfs:promote g_nfs:start
order o_root_before_nfs inf: cl_exportfs_root g_nfs:start
I think like that you try to start g_nfs twice
Hello Nick
Try to use nic=eth0 instead of nic=eth0:3
thanks
2013/3/24 Nick Walke tubaguy50...@gmail.com
Thanks for the tip, however, it did not work. That's actually a /116. So
I put in 2600:3c00::0034:c007/116 and am getting the same error. I
requested that it restart the resource
Try look here
http://www.hastexo.com/resources/hints-and-kinks/whats-totem-retransmit-list-all-about-corosync
2013/4/3 Ulrich Windl ulrich.wi...@rz.uni-regensburg.de
Hi!
I have a simple question: Is it possible that DLM or OCFS2 causes
corosync/TOTEM retransmit messages? I have the feeling
create a partition on /dev/sdd and you that
2013/4/9 Fredrik Hudner fredrik.hud...@gmail.com
Hi,
I have a (for now) two node HA cluster with sbd as stonith mechanism.
I have followed the installation and configuration of sbd from
http://www.linux-ha.org/wiki/SBD_Fencing.
For one reason or
Sorry
create a partition on /dev/sdd and you use that
2013/4/9 emmanuel segura emi2f...@gmail.com
create a partition on /dev/sdd and you that
2013/4/9 Fredrik Hudner fredrik.hud...@gmail.com
Hi,
I have a (for now) two node HA cluster with sbd as stonith mechanism.
I have followed
Hello Angel
In this thread
http://comments.gmane.org/gmane.linux.redhat.release.rhel5/6395 you can
find the answer to your question
Thanks
2013/4/22 Angel L. Mateo ama...@um.es
Hello,
I'm deploying a clustered pop/imap server with mailboxes stored in
a
SAN connected with fibre
maybe you can use openvswitch
2013/4/30 Lang, David david_l...@intuit.com
I've thought about this for a few years, but have not yet implemented it.
What I would look at is setting up a new virtual network that trunks your
two physical networks together and you can then use the IP on that
group astorage drbd5_fs drbd8_fs nfs-common nfs-kernel-server astorage_ip
order drbd5_fs_after_drbd5 inf: ma-ms-drbd5:promote drbd5_fs:start
order drbd8_fs_after_drbd8 inf: ma-ms-drbd8:promote drbd8_fs:start
should be
group astorage drbd5_fs drbd8_fs nfs-common nfs-kernel-server astorage_ip
sorry
it should be
group astorage drbd5_fs drbd8_fs nfs-common nfs-kernel-server astorage_ip
order drbd_fs_after_drbd inf: ma-ms-drbd5:promote ma-ms-drbd8:promote
astorage:start
2013/6/6 emmanuel segura emi2f...@gmail.com
group astorage drbd5_fs drbd8_fs nfs-common nfs-kernel-server
Hello Thomas
Sorry i can't give you any explain, because i don't see any sense in your
config
Sorry
2013/6/6 Thomas Glanzmann tho...@glanzmann.de
Hello,
on Debian Wheezy (7.0) I installed pacemaker with heartbeat. When
putting multiple filesystems which depend on multiple drbd promotions,
Hello Parkirat
can you share with us what was the problem? maybe this can help others
persons
Thanks
2013/6/16 Parkirat parkiratba...@gmail.com
I figured out the problem.
Thanks and Regards,
Parkirat Singh Bagga.
--
View this message in context:
Hello Parkirat
Thank you very much
2013/6/17 Parkirat parkiratba...@gmail.com
Thanks Ulrich,
I have figured out the problem.
The actual problem was in the configuration file for the resource httpd. It
was correct in the Master node but the configuration was missing in the
standby node,
Hello
Can you show us crm configure show?
thanks
2013/7/27 Miles Fidelman mfidel...@meetinghouse.net
Hi Folks,
Dual-node, pacemaker cluster, DRBD-backed xen virtual machines - one of
our VMs will run on one node, but not the other, and crm status yields a
failure message saying that
Put your cluster node hostnames in /etc/hosts and i think you are missing
cman two_node=1 expected_votes=1/ in cluster.conf
2013/8/23 Jakob Curdes j...@info-systems.de
Hmmm, the problem turns out to DNS-related. At startup, some of the
virtual interfaces are inactive and the DNS servers are
yum install corosync pacemaker
2013/8/23 Francis SOUYRI francis.sou...@apec.fr
Hi,
Thank you but I do not find on yum OpenAIS or an rpm for OpenAIS for fc18,
do you know where I can search.
Best regards.
Francis
On 08/23/2013 03:58 PM, Nick Cameo wrote:
Pacemaker+Corosync/OpenAIS
maybe you are missing the uuid library
2013/12/1 John Williams john.1...@yahoo.com
I'm trying to install heartbeat and I'm getting the following error with
the cluster glue components during the make part of the build:
/bin/sh ../../libtool --tag=CC --tag=CC --mode=link gcc -std=gnu99
The idea behind use exclusive volume activation mode with clvmd was(i
think), have a vg active active and lvs opens just in one node, more lvm
metadata replicated on all cluster nodes, when you'll do a change like lvm
resize.
I have a redhat cluster with clvmd with vg active in exclusive mode, if
you are missingo cluster configuration and sbd configuration and multipath
config
2014-04-22 20:21 GMT+02:00 Tom Parker tpar...@cbnco.com:
Has anyone seen this? Do you know what might be causing the flapping?
Apr 21 22:03:03 qaxen6 sbd: [12962]: info: Watchdog enabled.
Apr 21 22:03:03
(loop) : 1
Timeout (msgwait) : 90
==Header on disk /dev/mapper/qa-xen-sbd is dumped
On 22/04/14 02:30 PM, emmanuel segura wrote:
you are missingo cluster configuration and sbd configuration and
multipath
config
2014-04-22 20:21 GMT+02:00 Tom Parker tpar...@cbnco.com:
Has
no logs!
2014-06-05 14:56 GMT+02:00 Bart Coninckx bart.conin...@telenet.be:
Hi all,
I have some DRBD resources on CentoOS 6.5 which refuse to start. A message I
get in Hawk and in /var/log/messages is:
Failed op: node=storage3, resource=p_drbd_ws021, call-id=73,
If you have the ClustetIP resource in g_mysql, i think you don't need
order order_mysql_before_ip Mandatory: p_mysql ClusterIP because the
group is ordered by default, if you wanna mysql running on all boxes,
use clone resource and a colocation constraint to put ip on a box with
a mysql instance
depmod -a modprobe drbd ?
2014-07-18 13:05 GMT+02:00 willi.feh...@t-online.de willi.feh...@t-online.de:
Hello,
I'm trying to use DRBD on CentOS7. It looks like RedHat hasn't compiled DRBD
into the Kernel.
So I downloaded the source rpm from Fedora 19 and created my own rpm.
[root@centos7
rpm -ql drbd-8.4.3-2.el7.centos.x86_64
2014-07-18 16:31 GMT+02:00 Alessandro Baggi alessandro.ba...@gmail.com:
I'm new to CentOS, and more new on CentOS 7.
Maybe you have not compiled drbd module. Reading on drbd site you must
prepare kernel source tree and supply --with-km to compile also
but the nfs failover works now?
2014-07-22 2:10 GMT+02:00 Charles Taylor chas...@ufl.edu:
On Jul 21, 2014, at 10:40 AM, Charles Taylor wrote:
As I write this, I'm thinking that perhaps the way to achieve this is to
change the order of the services so that the VIP is started last and stopped
i'm using resource-agents-3.9.2-0.25.5 on Suse 11 Sp2 and i don't have any
. /usr/lib/ocf/lib/heartbeat/ora-common.sh maybe you need to create a
new database user
++ local 'conn_s=connect OCFMON/OCFMON'
++ shift 1
++ local func
++ echo 'connect OCFMON/OCFMON'
2014-08-01 18:40 GMT+02:00 Wendt
logs?
2014-12-29 6:54 GMT+01:00 Marlon Guao marlon.g...@gmail.com:
Hi,
just want to ask regarding the LVM resource agent on pacemaker/corosync.
I setup 2 nodes cluster (opensuse13.2 -- my config below). The cluster
works as expected, like doing a manual failover (via crm resource move),
; then
ocf_exit_reason WARNING:
$OCF_RESKEY_volgrpname is active without the cluster tag, \$OUR_TAG\
On Mon, Dec 29, 2014 at 3:36 PM, emmanuel segura emi2f...@gmail.com
wrote:
logs?
2014-12-29 6:54 GMT+01:00 Marlon Guao marlon.g...@gmail.com:
Hi,
just want to ask regarding the LVM resource
, Dec 29, 2014 at 4:42 PM, emmanuel segura emi2f...@gmail.com
wrote:
please use pastebin and show your whole logs
2014-12-29 9:06 GMT+01:00 Marlon Guao marlon.g...@gmail.com:
by the way.. just to note that.. for a normal testing (manual failover,
rebooting the active node)... the cluster
marlon.g...@gmail.com:
ok, sorry for that.. please use this instead.
http://pastebin.centos.org/14771/
thanks.
On Mon, Dec 29, 2014 at 5:25 PM, emmanuel segura emi2f...@gmail.com wrote:
Sorry,
But your paste is empty.
2014-12-29 10:19 GMT+01:00 Marlon Guao marlon.g...@gmail.com:
hi
/cib/cib-0.raw: Operation not permitted (1)
2014-12-29 10:33 GMT+01:00 emmanuel segura emi2f...@gmail.com:
Hi,
You have a problem with the cluster stonithd:error: crm_abort:
crm_glib_handler: Forked child 6186 to record non-fatal assert at
logging.c:73
Try to post your cluster version
to poweroff the active node.. and tried pvscan on the
passive.. and yes.. it didn't worked --- it doesn't return to the shell.
So, the problem is on DLM?
On Mon, Dec 29, 2014 at 5:51 PM, emmanuel segura emi2f...@gmail.com wrote:
Power off the active node and after one seconde try to use one lvm
-b4ee-385e83e6abcc
Number of slots: 255
Sector size: 512
Timeout (watchdog) : 5
Timeout (allocate) : 2
Timeout (loop) : 1
Timeout (msgwait) : 10
On Mon, Dec 29, 2014 at 6:42 PM, emmanuel segura emi2f...@gmail.com wrote:
Dlm isn't the problem, but i think is your fencing, when
On Mon, Dec 29, 2014 at 9:24 PM, Marlon Guao marlon.g...@gmail.com wrote:
interesting, i'm using the newer pacemaker version..
pacemaker-1.1.12.git20140904.266d5c2-1.5.x86_64
On Mon, Dec 29, 2014 at 8:11 PM, emmanuel segura emi2f...@gmail.com
wrote:
https://bugzilla.redhat.com/show_bug.cgi?id
when a node goes down, you will see the node in unclean state, how you
see in your logs, forming new configuration(corosync) - stonith
reboot request - and you are using sbd and the node become offline
after thet msgwait is expired, when msgwait is expired pacemaker knows
the node is dead and than
68 matches
Mail list logo