-----Original Message-----
From: Florian Haas [mailto:[email protected]]
Sent: Sunday, October 07, 2012 5:47 AM
To: Dan Barker
Cc: drbd List
Subject: Re: [DRBD-user] Does oversize disk hurt anything?

>On Fri, Oct 5, 2012 at 7:59 PM, Dan Barker <[email protected]> wrote:
>> I just lost a disk on my secondary node. I looked EVERYWHERE and can't
>> find the spare disks I bought for such an occurrence. So, I put in a
>> handy disk, twice the size.
>>
>> drbdadm create-md r1
>> drbdadm attach r1
>>
>> and off we go.
>>
>> If memory serves, create-md will build a meta-data at the END of the disk.
>> Won't that cause a lot of seek to the hub when seeking to about the
>> middle of the platters would have done the trick, had the metadata
>> been at the same offset as the primary?
>
>Well if you had created a partition (/dev/sdc1) rather than use the full disk 
>(/dev/sdc), then you could have set up that partition to match the size of the 
>disk on your primary.

Partition. Great idea. If I had thought of that, I'd have bought only one new 
500G disk instead of two. Thanks for the hint. 1T disks cost the same as 500G 
these days.

>Besides, if you're using a RAID controller with a battery/flashed-backed write 
>cache then it won't matter much. I wrote about this years ago on my blog:
>http://fghaas.wordpress.com/2009/08/20/internal-metadata-and-why-we-recommend-it/

It's not RAID, it's simply a single SATA disk controlled by the motherboard 
chips.

>
>> version: 8.4.0 (api:1/proto:86-100)
>> GIT-hash: 28753f559ab51b549d16bcf487fe625d5919c49c build by
>> root@DrbdR0,
>> 2012-05-28 12:09:30 (Yes, I know. I need to upgrade).
>
>True. Rather urgently if you're on 8.4.0.
>

Agreed.

>> Failed disk: WD 500G
>> Replaced by: WD 1T
>> On server: DrbdR0
>>
>> cat /etc/drbd.d/r1.res
>> resource r1 {
>>     on DrbdR0 {
>>         volume 0 {
>>             device       /dev/drbd1 minor 1;
>>             disk         /dev/sdc;
>>             meta-disk    internal;
>>         }
>>         address          ipv4 10.20.30.46:7790;
>>     }
>>     on DrbdR1 {
>>         volume 0 {
>>             device       /dev/drbd1 minor 1;
>>             disk         /dev/sdc;
>>             meta-disk    internal;
>>         }
>>         address          ipv4 10.20.30.47:7790;
>>     }
>>     startup {
>>         become-primary-on DrbdR1;
>
>Why? Your cluster manager (typically Pacemaker) should take care of that for 
>you.

No cluster manager, no NA. Easy manual failover. This is a lab environment and 
HA is not really needed. The users of drbd storage are ESXi hosts. To "take" 
the primary server off line I:
DrbdR0: drbdadm primary all (allow dual primaries is on)
DrbdR0: start iet
ESXi (all): verify all four paths to both drbd are online
Drbdr1: stop iet
Drbdr0: stop drbd
Drbdr0: <perform maintenance, or whatever>

I think I'll add a new, first step: Verify no drbd resource is diskless! (and 
all are uptodate).

There is a slight possibility of a problem when running dual primaries since 
all my ESXi allocations are thin, but it's remote enough to me for this 
purpose. I don't run dual-primary for more than a few minutes and certainly not 
while any heavy allocations are being done by the ESXi guests.

>
>Cheers,
>Florian
>

Thanks for the help.

Dan in Atlanta

_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

Reply via email to