Ok I didn't realise this was that kind of an issue, I just assumed DRBD 
wasn't happy with my setup due to the second node not being online.
The partition /dev/sdb1 is only set to about 40GB while I'm testing the 
setup, once I've nutted everything the machine will be rebuilt to have 
it's full capacity. (around 7TB) it's an LVM partition

I created a single Logical Volume around 20GB have it exported via 
iscsi-scst. 

Single VMWare ESXi server (again this is all testing this phase), setup 
iscsi to connect to the SAN (testing MPIO) as soon as I "scan" from the 
storage adapter I start getting the 
[28609.797320] block drbd1: al_complete_io() called on inactive extent 623 

ete..

see http://www.genis-x.com/files/syslog

Where the log ends is where I shutdown the ESXi server

If there are any additional debugging steps/options I can enable that 
would help, let me know.
I haven't done anything with /dev/drdb0 yet this is going to hold config 
files etc.

Background:
Ubuntu 10.4 with recompiled linux-source (2.6.32.15) kernel due to 
iscsi-scst patches.
iscsi-scst as target engine
heartbeat (not yet though just getting the base working, drbd + scst) then 
I'll integrate heartbeat
lvm2 on top of drbd

r...@storage00:/etc# cat drbd.conf
global {
        usage-count no;
}

common {
        syncer { rate 110M; }
}
resource meta {
        protocol C;
        handlers {
         pri-on-incon-degr "echo O > /proc/sysrq-trigger ; halt -f";
         pri-lost-after-sb "echo O > /proc/sysrq-trigger ; halt -f";
         local-io-error "echo O > /proc/sysrq-trigger ; halt -f";
        }

        startup {
         # wfc-timeout 0;
         degr-wfc-timeout 120; # 2 minutes.
        }
        disk {
         on-io-error detach;
        }

        net {
         after-sb-0pri disconnect;
         after-sb-1pri disconnect;
         after-sb-2pri disconnect;
         rr-conflict disconnect;
        }

        syncer {
         # rate 110M;
         # after "r2";
         al-extents 257;
        }

        on storage00 {
         device /dev/drbd0;
         disk /dev/sda3;
         address 192.168.1.10:7788;
         meta-disk internal;
        }

        on storage01 {
         device /dev/drbd0;
         disk /dev/sda3;
         address 192.168.1.11:7788;
         meta-disk internal;
        }
}

resource data {
        protocol C;
        startup {
         wfc-timeout 0; ## Infinite!
         degr-wfc-timeout 120; ## 2 minutes.
        }

        disk {
         on-io-error detach;
        }

        net {
         # timeout 60;
         # connect-int 10;
         # ping-int 10;
         # max-buffers 2048;
         # max-epoch-size 2048;
        }

        syncer {
         after "meta";
        }

        on storage00 {
         device /dev/drbd1;
         disk /dev/sdb1;
         address 192.168.1.10:7789;
         meta-disk internal;
        }
        on storage01 {
         device /dev/drbd1;
         disk /dev/sdb1;
         address 192.168.1.11:7789;
         meta-disk internal;
        }
}

r...@storage00:/etc# fdisk -l

Disk /dev/sda: 21.5 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00004968

   Device Boot      Start         End      Blocks   Id  System
/dev/sda1   *           1        2432    19530752   83  Linux
/dev/sda2            2432        2494      499712   82  Linux swap / 
Solaris
/dev/sda3            2494        2611      939008   83  Linux

Disk /dev/sdb: 42.9 GB, 42949672960 bytes
255 heads, 63 sectors/track, 5221 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00064540

   Device Boot      Start         End      Blocks   Id  System
/dev/sdb1               1        5222    41940992   8e  Linux LVM

Regards,
Adam Niedzwiedzki

c: eye media australia pty ltd
a: suite 109, 685 burke rd, camberwell, victoria, 3124, au, earth
p: +61 (0)3 8080 5850
d: +61 (0)3 8080 5674
f: +61 (0)3 8080 5840
m: +61 (0)439 646 941
w: http://www.eyemedia.com.au
pgp: http://www.eyemedia.com.au/adamn.asc



From:
Lars Ellenberg <[email protected]>
To:
[email protected]
Date:
14/07/2010 23:17
Subject:
Re: [DRBD-user] 1 Primary Node
Sent by:
[email protected]



On Wed, Jul 14, 2010 at 03:54:48PM +1000, [email protected] wrote:
> Hi all,
> 
> I have a physical server that I'm setting up as a SAN, the second node 
> won't be available to put online for at least a week. (migrating FROM it 

> to the NEW SAN setup)
> 
> How do I tell DRBD to I guess stop trying to synchronise? I get pages 
and 
> pages of
> 
> [28609.791929] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.797320] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.801638] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.807021] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.811691] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.816680] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.821531] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.826190] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.829968] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.833730] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.837759] block drbd1: al_complete_io() called on inactive extent 
623

Is that the only message that gets logged there?
Nothing before?

"That should not happen", and actually if it does happen, either
something is broken in your setup, or DRBD has a bug.

> I'm running version: 8.3.8.1 (api:88/proto:86-94)
> 
> So in short I'm wanting to put my primary node online, migrate a bunch 
of 
> data from a "current" SAN (it's NOT running DRBD, around 7TB of data) to 

> the primary, format rebuild the current THEN sync.

That plan sounds ok, and should just work.

-- 
: Lars Ellenberg
: LINBIT | Your Way to High Availability
: DRBD/HA support and consulting http://www.linbit.com

DRBD® and LINBIT® are registered trademarks of LINBIT, Austria.
__
please don't Cc me, but send to list   --   I'm subscribed
_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

Reply via email to