Have you test these values ?

https://drbd.linbit.com/users-guide/s-throughput-tuning.html


El 26/05/15 a las 13:16, Ben RUBSON escribió:
RAID controller is OK yes.

Here is a 4 steps example of the issue :



### 1 - initial state :

Source :
- sdb read MB/s      : 0
- sdb write MB/s     : 0
- eth1 incoming MB/s : 0
- eth1 outgoing MB/s : 0
Target :
- sdb read MB/s      : 0
- sdb write MB/s     : 0
- eth1 incoming MB/s : 0
- eth1 outgoing MB/s : 0



### 2 - dd if=/dev/zero of=bigfile :

Source :
- sdb read MB/s      : 0
- sdb write MB/s     : 670
- eth1 incoming MB/s : 1
- eth1 outgoing MB/s : 670
Target :
- sdb read MB/s      : 0
- sdb write MB/s     : 670
- eth1 incoming MB/s : 670
- eth1 outgoing MB/s : 1



### 3 - disable the link between the 2 nodes :

Source :
- sdb read MB/s      : 0
- sdb write MB/s     : 670
- eth1 incoming MB/s : 0
- eth1 outgoing MB/s : 0
Target :
- sdb read MB/s      : 0
- sdb write MB/s     : 0
- eth1 incoming MB/s : 0
- eth1 outgoing MB/s : 0



### 4 - re-enable the link between the 2 nodes :

Source :
- sdb read MB/s      : ~20
- sdb write MB/s     : ~670
- eth1 incoming MB/s : 1
- eth1 outgoing MB/s : 670
Target :
- sdb read MB/s      : 0
- sdb write MB/s     : 670
- eth1 incoming MB/s : 670
- eth1 outgoing MB/s : 1
DRBD resource :
 1: cs:SyncTarget ro:Secondary/Primary ds:Inconsistent/UpToDate C r-----
ns:62950732 nr:1143320132 dw:1206271712 dr:1379744185 al:9869 bm:6499 lo:2 pe:681 ua:1 ap:0 ep:1 wo:d oos:11883000
    [>...................] sync'ed:  6.9% (11604/12448)M
    finish: 0:34:22 speed: 5,756 (6,568) want: 696,320 K/sec



### values I would have expected in step 4 :

Source :
- sdb read MB/s      : ~400 (because of c-min-rate 400M)
- sdb write MB/s     : ~370
- eth1 incoming MB/s : 1
- eth1 outgoing MB/s : 670
Target :
- sdb read MB/s      : 0
- sdb write MB/s     : 670
- eth1 incoming MB/s : 670
- eth1 outgoing MB/s : 1

Why resync is totally ignored and application (dd here in the example) still consumes all available IOs / bandwidth ?



Thank you,

Ben



2015-05-25 16:50 GMT+02:00 A.Rubio <[email protected] <mailto:[email protected]>>:

    Cache settings an I/O in RAID controller is optimal ???
    Write-back, write-through, cache enablad, I/O direct, ...

    El 25/05/15 a las 11:50, Ben RUBSON escribió:

        The link between nodes is a 10Gb/s link.
        The DRBD resource is a RAID-10 array which is able to resync
        at up to 800M (as you can see I have lowered it to 680M in my
        configuration file).

        The "issue" here seems to be a prioritization "issue" between
        application IOs and resync IOs.
        Perhaps I miss-configured something ?
        Goal is to have resync rate up to 680M, with a minimum of
        400M, even if there are application IOs.
        This in order to be sure to complete the resync even if there
        are a lot of write IOs from the application.

        With my simple test below, this is not the case, dd still
        writes at its best throughput, lowering resync rate which
        can’t reach 400M at all.

        Thank you !

            Le 25 mai 2015 à 11:18, A.Rubio <[email protected]
            <mailto:[email protected]>> a écrit :

            the link between nodes is ???  1Gb/s ??? , 10Gb/s ??? ...

            the Hard Disks are ??? SATA 7200rpm ???, 10000rpm ???, SAS
            ???,
            SSD ???...

            400M to 680M with a 10Gb/s link and SAS 15.000 rpm is OK
            but less ...

                Le 12 avr. 2014 à 17:23, Ben RUBSON
                <[email protected] <mailto:[email protected]>> a
                écrit :

                Hello,

                Let's assume the following configuration :
                disk {
                    c-plan-ahead 0;
                    resync-rate 680M;
                    c-min-rate 400M;
                }

                Both nodes are uptodate, and on the primary, I have a
                test IO burst running, using dd.

                I then cut replication link for a few minutes so that
                secondary node will be several GB behind primary node.

                I then re-enable replication link.
                What I expect here according to the configuration is
                that secondary node will fetch missing GB at a 400
                MB/s throughput.
                DRBD should then prefer resync IOs over application
                (dd here) IOs.

                However, it does not seems to work.
                dd still writes at its best throughput, meanwhile
                reads are made from the primary disk between 30 and 60
                MB/s to complete the resync.
                Of course this is not the expected behaviour.

Did I miss something ?



_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user

Reply via email to