[DRBD-user] Extremely high latency problem

Bret Mette bret.mette at dbihosting.com
Wed Jun 4 17:31:30 CEST 2014

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Hello,

I started looking at DRBD as a HA ISCSI target. I am experiencing very poor
performance and decided to run some tests. My current setup is as follows:

Intel(R) Xeon(R) CPU E3-1230 V2 @ 3.30GH
CentoS 6.5 - 2.6.32-431.17.1.el6.x86_64
drbd version: 8.3.16 (api:88/proto:86-97)
md RAID10 using 7200rpm drives

The 2 drbd nodes are synced using an intel  82579LM Gigabit card

I have created an logical drive using LVM and configured a couple drbd
resources on top of that. drbd0 is my iscsi configuration file, which is
shared between the 2 nodes and drbd1 is a 1.75TB ISCSI target.

I run heartbeat on the two nodes and expose a virtual IP to the ISCSI
initiators.

Originally I was running ISCSI with write-cache off (for data integrity
reasons) but have recently switched to write-cache on during testing (with
little to no gain).

My major concern is the extremely high latency test results I got when when
dd against drbd0 mounted on the primary node.

dd if=/dev/zero of=./testbin  bs=512 count=1000 oflag=direct
512000 bytes (512 kB) copied, 32.3254 s, 15.8 kB/s

I have pinged the second node as a very basic network latency test and get
0.209ms response time. I have also run the same test on both nodes with
drbd disconnected (or on partitions not associated with drbd) and get
typical results:

node1
dd if=/dev/zero of=./testbin  bs=512 count=1000 oflag=direct
12000 bytes (512 kB) copied, 0.153541 s, 3.3 MB/s

node2
dd if=/dev/zero of=~/testbin  bs=512 count=1000 oflag=direct
512000 bytes (512 kB) copied, 0.864994 s, 592 kB/s
512000 bytes (512 kB) copied, 0.328994 s, 1.6 MB/s

node2's latency (without drbd connected) is inconsistent but always falls
between those two ranges.

These tests were run with no ISCSI targets exposed, no initiators
connected, essentially on an idle system.

My question is why are my drbd connected latency tests showing results 35
to 100 times slower than my results when dbrd is not connected (or against
partitions not backed by drbd)?

This seems to be the source of my horrible performance on the ISCSI targs
(300~900 K/sec dd writes on the initiators) and very high iowait (35-75%)
on mildly busy initiators.


Any advice pointers, etc. would be highly appreciated. I have already tried
numerous performance tuning settings (suggested by the drbd manual). But I
am open to any suggestion and will try anything again if it might solve my
problem.

Here are the important bits of my current drbd.conf

        net {
        cram-hmac-alg sha1;
        shared-secret "password";
        after-sb-0pri disconnect;
        after-sb-1pri disconnect;
        after-sb-2pri disconnect;
        rr-conflict disconnect;
        max-buffers 8000;
        max-epoch-size 8000;
        sndbuf-size 0;
        }

        syncer {
        rate 100M;
        verify-alg sha1;
        al-extents 3389;
        }

I've played with the watermark setting and a few others and latency only
seems to get worse or stay where it's at.


Thank you,
Bret
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20140604/6dfc5c6d/attachment.htm>


More information about the drbd-user mailing list