[DRBD-user] Only 25% performance on a single node DRBD?

Brian Walters brian at penguinpeople.net
Tue Feb 10 21:52:17 CET 2015

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Greetings,

I'm currently working on a new server that I plan to use for an iSCSI 
SAN.  Because of a hardware availability issue I have to initially 
configure this as a single node DRBD system until I complete the 
migration of data from another server configured with similar hardware.

In a nutshell, my problem is that I'm seeing horrible write performance 
when writing to the DRBD device versus writing to the underlying Raid array.

root at san01-a:~# dd if=/dev/zero of=/dev/drbd1 bs=1M count=10000 oflag=direct
10485760000 bytes (10 GB) copied, 25.5897 s, 410 MB/s

root at san01-a:~# dd if=/dev/zero of=/dev/sda2 bs=1M count=10000 oflag=direct
10485760000 bytes (10 GB) copied, 7.07579 s, 1.5 GB/s

Quick Hardware description:
SuperMicro Mainboard w/ Xeon E5-2603
16GB Ram
Dual 120G SSD drives in a Linux Raid 1 for OS - Debian Wheezy
Adaptec ASR7805 w/BBU - latest firmware (32033)   w/ BBU
22 Toshiba 1TB SAS drives (MG03SCA100) in a Raid 60 array w/ 128k stripe
Dual 10G links (Intel card) to the second server when it is ready

drbdsetup show:
root at san01-a:~# drbdsetup show
resource meta {
     options {
         cpu-mask            "ff";
     }
     net {
         max-epoch-size      20000;
         max-buffers         131072;
         unplug-watermark    131072;
         after-sb-0pri       discard-least-changes;
         after-sb-1pri       consensus;
     }
     _remote_host {
         address            ipv4 1.1.200.2:7788;
     }
     _this_host {
         address            ipv4 1.1.200.1:7788;
         volume 0 {
             device            minor 0;
             disk            "/dev/sda1";
             meta-disk            internal;
             disk {
                 disk-flushes        no;
                 md-flushes          no;
                 al-extents          3389;
             }
         }
     }
}

resource data {
     options {
         cpu-mask            "ff";
     }
     net {
         max-epoch-size      20000;
         max-buffers         131072;
         unplug-watermark    131072;
         after-sb-0pri       discard-least-changes;
         after-sb-1pri       consensus;
     }
     _remote_host {
         address            ipv4 1.1.200.2:7789;
     }
     _this_host {
         address            ipv4 1.1.200.1:7789;
         volume 0 {
             device            minor 1;
             disk            "/dev/sda2";
             meta-disk            internal;
             disk {
                 disk-flushes        no;
                 md-flushes          no;
                 al-extents          3389;
             }
         }
     }
}
=====================

I've tried this with various settings and about the only setting that 
really makes a difference is cpu-mask ff;  The others (md-flushes, 
al-extents, disk-flushes, max-buffers) can all be removed with no real 
change in performance.

DRBD and drbd-utils were built from the git code
version: 8.4.5 (api:1/proto:86-101)
GIT-hash: 1d360bde0e095d495786eaeb2a1ac76888e4db96 build by root at san01, 
2015-02-05 11:09:59
  0: cs:WFConnection ro:Primary/Unknown ds:UpToDate/DUnknown C r----s
     ns:0 nr:0 dw:0 dr:664 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:d 
oos:511948
  1: cs:WFConnection ro:Primary/Unknown ds:UpToDate/DUnknown C r----s
     ns:0 nr:0 dw:10240000 dr:996 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 
wo:d oos:17570968856


root at san01-a:~# drbdadm  -v
Version: 8.9.2rc2 (api:1)
GIT-hash: faeb645ecbf334347e0512b4fa2d7549543b5b50 build by root at san01, 
2015-02-10 13:41:16

root at san01-a:~# uname -a
Linux san01 3.2.0-4-amd64 #1 SMP Debian 3.2.65-1+deb7u1 x86_64 GNU/Linux


The strangest thing is that I built a system with all the same hardware 
8 months ago and  I was seeing 700MB/s writes. The difference was that I 
had both nodes to work on while creating the DRBD.  This time I only 
have 1 node to start with.

-- 
Brian
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20150210/9870cd8f/attachment.htm>


More information about the drbd-user mailing list