I am testing a DRBD 9 and 8.4 in simple 2 node active - passive clusters with 
NFS.

Copying files form a third server to the NFS share using dd, I typically see an 
average of 20% CPU load (with v9) on the primary during transfer of larger 
files, testing with 0,5 and 2 GB.

At the very end of the transfer DRBD process briefly peaks at 70 - 100% CPU. 

This causes occasional problems with Corosync believing the node is down. 
Increasing the token time for Corosync to 2000 ms fixes the symptom but I am 
wondering about the root cause and any possible fixes? 

This is the DRBD configuration.

resource san_data {
  protocol C;
  meta-disk internal;
  device /dev/drbd1;
  disk   /dev/nfs/share;
  net {
    verify-alg sha1;
    cram-hmac-alg sha1; 
    shared-secret ”****************";
    after-sb-0pri discard-zero-changes;
    after-sb-1pri discard-secondary;
    after-sb-2pri disconnect;
  }
  on san1 { 
    address  192.168.1.86:7789;
  }
  on san2 { 
    address  192.168.1.87:7789;
  }
}

The nodes are two VM on different ESXi hosts (Dell T620). Hosts are very 
lightly loaded. Network is 1 Gb at the moment through a Catalyst switch. 
Network appears not saturated.

BTW when can we expect a DRBD resource agent for v9? It took me a while to 
figure out why DRBD9 was not working with Pacemaker and then finding a patch to 
the agent :) 

Cheers Mats
_______________________________________________
drbd-user mailing list
drbd-user@lists.linbit.com
http://lists.linbit.com/mailman/listinfo/drbd-user

Reply via email to