Hi Richard

I am using PVE + DRBD (primary-primary, protocol C) + HA (rgmanager) without problems since many years ago (With and without BBU in the RAID cards, always with the recommended tuning by the DRBD web portal).

Now i am using:
- 2 NICs exclusive for DRBD in bonding-rr (NIC-to-NIC, some pair of nodes with 10 Gb/s, and others with 1 Gb/s, always with 9000 mtu)
- Programs:
drbd-utils 8.9.(1+linbit-1)
drbd8-module-source (2:8.4.4-0)
- Kernel: 3.10
pve-kernel-3.10.0-5-pve (3.10.0-19)
- LVM on top of DRBD

Hardware:
For better performance, will be better that you enable in the Bios Hardware:
- I/OAT DMA engine (in all the servers with Intel Processors, old or new, always works well with the NICs)
- Best performance (not save energy)
PVE (old and new versions) always works better with this Hardware configuration.

----- Original Message ----- From: "Lechner Richard" <[email protected]>
To: <[email protected]>
Sent: Thursday, January 22, 2015 3:02 PM
Subject: Re: [DRBD-user] split brain issues


Am Donnerstag 22 Januar 2015, 18:51:52 schrieb Yannis Milios:

Hello Yannis,

Not sure what caused Inconsistent state on your drbd resource on Node1, but
my guess is that it experienced some kind of low level corruption on it's
backing storage (hard disks) and auto sync was initiated from Node2.

Hard disc and md devices all are still running without any error.
At the moment i run it as Primary/Secondary and status is Up2date/Up2date.
So it looks like Eric's mention is right, pri/pri make some issues.
But before it was ok so i think it must be one of the latest kernel updates.
Somthing changed in the drbd code.

Are you using linux raid? Then probably you don't have a battery backed
raid controller, so it would be wise to remove the following entries since
they could cause data loss in your case:

                no-disk-flushes;
                no-md-flushes;
                no-disk-barrier;

Without this 3 options (and the right buffers) i get a realy bad performance, it's a mailserver with a lot of small files and with drbd protocol C sometimes
it takes to long to respond for some client's.
Biggest mailbox has 127 GB and 204.000 messages inside.
Every time the mailserver build's a new mailbox index, u can imagine! :-(

I do a lot of test's and it works fine with my config since Mar 2014, only the
last few weeks drbd start's hurting me. :-(

Finally from my experince proxmox does not work well with HA enabled (it's
not aware of underlying drbd resource) so it could cause frequent
split-brains to occur.Use DRBD without HA enabled on Proxmox , in
dual-primary mode (so you don't loose live migration capability on proxmox).

At the moment i stoped ha and run it with pri/sec but a want to get back my
ha!

You can also create separate drbd resources for each proxmox mode so you
can better handle split brains.For example drbd0 -> drbdvg0 -> mounted
always on Node1 and drbd1 -> drbdvg1 -> mounted on Node2.
This way you will always know that vms running on Node1 are located on
drbd0 and vms running on Node2 are located on drbd1.

Dont need this because i run only 1 VM in this cluster.

Regards

Richard

_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

Reply via email to