Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
I am testing a DRBD 9 and 8.4 in simple 2 node active - passive clusters with NFS.
Copying files form a third server to the NFS share using dd, I typically see an average of 20% CPU load (with v9) on the primary during transfer of larger files, testing with 0,5 and 2 GB.
At the very end of the transfer DRBD process briefly peaks at 70 - 100% CPU.
This causes occasional problems with Corosync believing the node is down. Increasing the token time for Corosync to 2000 ms fixes the symptom but I am wondering about the root cause and any possible fixes?
This is the DRBD configuration.
resource san_data {
protocol C;
meta-disk internal;
device /dev/drbd1;
disk /dev/nfs/share;
net {
verify-alg sha1;
cram-hmac-alg sha1;
shared-secret ”****************";
after-sb-0pri discard-zero-changes;
after-sb-1pri discard-secondary;
after-sb-2pri disconnect;
}
on san1 {
address 192.168.1.86:7789;
}
on san2 {
address 192.168.1.87:7789;
}
}
The nodes are two VM on different ESXi hosts (Dell T620). Hosts are very lightly loaded. Network is 1 Gb at the moment through a Catalyst switch. Network appears not saturated.
BTW when can we expect a DRBD resource agent for v9? It took me a while to figure out why DRBD9 was not working with Pacemaker and then finding a patch to the agent :)
Cheers Mats