On 6/25/19 6:00 AM, Oleksiy Evin wrote:
> However, could you please clarify, what are the limitations/drawbacks
> of a single volume use?
Regarding huge volumes, the most obvious one is reaching the size limit,
which is currently 1 PiB.
But there are other limitations as well, e.g.:
- a single hug
BD -> LVM -> XFS I guess it would be required to
recover LVM volume changing DRBD to the physical disks in meta data.
-Original Message-
From: Adam Goryachev
To: drbd-user@lists.linbit.com
Subject: Re: [DRBD-user] PingAck did not arrive in time.
Date: Tue, 25 Jun 2019 09:12:17 +1000
de to DRBD 8, and 2 nodes
setup.
I'll give another try with increased "ping-timeout" shortly and let you know if
it resolved the problem. Anyway, thank you for all your assistance!
-Original Message-
From: Robert Altnoeder
To: drbd-user@lists.linbit.com
Subject: Re: [DR
On 25/6/19 12:58 am, Robert Altnoeder wrote:
On 6/24/19 3:39 PM, Oleksiy Evin wrote:
At this moment we're able to achieve about 1.2GB while replication
which is little bit slower then a single enclosure speed, but enough
for all our needs. Potentially we may grow up above PB next year or two.
I
On 6/24/19 3:39 PM, Oleksiy Evin wrote:
> At this moment we're able to achieve about 1.2GB while replication
> which is little bit slower then a single enclosure speed, but enough
> for all our needs. Potentially we may grow up above PB next year or two.
I don't think so, at least not with a singl
We are using this volume for video editing, mainly it's video/audio/photo
storage plus some project files replicated twice (previously 3 times). On a
physical level single node contains 6x12 bay RAID0 enclosures joined into a
logical volume. I've been playing earlier with a different configurati
On Fri, Jun 21, 2019 at 02:33:08PM +0200, Robert Altnoeder wrote:
> On 6/21/19 9:16 AM, Oleksiy Evin wrote:
> > I've tried to update the kernel and DRBD 9 to the last available
> > version, but nothing helped with the PingAck issue. So I had to
> > downgrade to DRBD 8.4 which is started to replicat
On 6/21/19 9:16 AM, Oleksiy Evin wrote:
> I've tried to update the kernel and DRBD 9 to the last available
> version, but nothing helped with the PingAck issue. So I had to
> downgrade to DRBD 8.4 which is started to replicate fine
> [...]
> Can anyone help me on the "PingAck did not arrive in time
I've tried to update the kernel and DRBD 9 to the last available version, but
nothing helped with the PingAck issue. So I had to downgrade to DRBD 8.4 which
is started to replicate fine except the following messages in the log:
[139267.930516] block drbd0: BAD! enr=34406392 rs_left=-4 rs_failed=
For those who are interested, or those who might stumble on this thread
through Google:
After fiddling around a bit longer without any reliable results, I've
decided to install DRBD8 instead of DRBD9, as Yannis suggested earlier.
Worked straight away, and after a bit of tuning it works awesome.
Hello,
Thank you for your suggestion. The MTU is 1500 on both nodes. I had it
at 9000, but reverted everything to 'normal' to debug this problem.
Pinging as in your example works fine.
Cheers,
Dirk
On 23-05-18 21:22, Nelson Hicks wrote:
> Is there any chance this could be an MTU mismatch betwee
Is there any chance this could be an MTU mismatch between the two nodes?
If you use ping with varying packet sizes from one node to the other, do
they stop working above a specific size? Does ifconfig report the same
MTU size for the interface on both nodes?
Examples:
ifconfig | grep MTU
pin
Hi,
Thank you for your reply.
I am / was under the impression that DRBD9 is the new and improved DRBD,
so I figured to use this version. But this is not the case? Could
somebody enlighten me a bit?
I already have disabled all bonding and other fancy network stuff, so
I'm usingĀ 1 nic currently.
Two things:
- I would use drbd8 instead of drbd9 for a 2 node setup.
- I would first test with 1 nic instead of 2.
On Wed, May 23, 2018 at 11:01 AM, Dirk Bonenkamp - ProActive <
d...@proactive.nl> wrote:
> Hi List,
>
> I'm struggling with a new DRBD9 setup. It's a simple Master/Slave setup.
> I'
14 matches
Mail list logo