[DRBD-user] Slower disk throughput on DRBD partition

Frederic DeMarcy fred.demarcy.ml at gmail.com
Wed Feb 1 17:19:30 CET 2012

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Hi Pascal

1) Both vdisks for /dev/sda and /dev/sdb are on the same datastore which is
made of the entire RAID5 array capacity (7xHDs + 1 spare). Minus space used
by the ESXi installation.
2) HD1 (/dev/sda) is SCSI (0:1) and HD2 (/dev/sdb) is SCSI (0:2). Both
initialized with Thick Provisioning Eager Zeroed. The SCSI controller type
is paravirtual.

Fred

On Wed, Feb 1, 2012 at 2:13 PM, Pascal BERTON (EURIALYS) <
pascal.berton at eurialys.fr> wrote:

> Frederic,
>
> Let's take care of the virtualisation layer wich might induce significant
> side effects
> Are sda and sdb :
> 1) vdisk files located on the same datastore ?
> 2) vdisks plugged on the same virtual SCSI interface ? What type of SCSI
> interface ?
>
> Best regards,
>
> Pascal.
>
> -----Message d'origine-----
> De : drbd-user-bounces at lists.linbit.com
> [mailto:drbd-user-bounces at lists.linbit.com] De la part de Frederic DeMarcy
> Envoyé : mercredi 1 février 2012 13:05
> À : drbd-user at lists.linbit.com
> Objet : Re: [DRBD-user] Slower disk throughput on DRBD partition
>
> Hi
>
> Note 1:
> Scientific Linux 6.1 with kernel 2.6.32-220.4.1.el6.x86_64
> DRBD 8.4.1 compiled from source
>
> Note 2:
> server1 and server2 are 2 VMware VMs on top of ESXi 5. However they reside
> on different physical 2U servers.
> The specs for the 2U servers are identical:
>  - HP DL380 G7 (2U)
>  - 2 x Six Core Intel Xeon X5680 (3.33GHz)
>  - 24GB RAM
>  - 8 x 146 GB SAS HD's (7xRAID5 + 1s)
>  - Smart Array P410i with 512MB BBWC
>
> Note 3:
> I've tested the network throughput with iperf which yields close to 1Gb/s
> [root at server1 ~]# iperf -c 192.168.111.11 -f g
> ------------------------------------------------------------
> Client connecting to 192.168.111.11, TCP port 5001
> TCP window size: 0.00 GByte (default)
> ------------------------------------------------------------
> [  3] local 192.168.111.10 port 54330 connected with 192.168.111.11 port
> 5001
> [ ID] Interval       Transfer     Bandwidth
> [  3]  0.0-10.0 sec  1.10 GBytes  0.94 Gbits/sec
>
> [root at server2 ~]# iperf -s -f g
> ------------------------------------------------------------
> Server listening on TCP port 5001
> TCP window size: 0.00 GByte (default)
> ------------------------------------------------------------
> [  4] local 192.168.111.11 port 5001 connected with 192.168.111.10 port
> 54330
> [ ID] Interval       Transfer     Bandwidth
> [  4]  0.0-10.0 sec  1.10 GBytes  0.94 Gbits/sec
>
> Scp'ing a large file from server1 to server2 yields ~ 57MB/s but I guess
> it's due to the encryption overhead.
>
> Note 4:
> MySQL was not running.
>
>
>
> Base DRBD config:
> resource mysql {
>  startup {
>    wfc-timeout 3;
>    degr-wfc-timeout 2;
>    outdated-wfc-timeout 1;
>  }
>  net {
>    protocol C;
>    verify-alg sha1;
>    csums-alg sha1;
>    data-integrity-alg sha1;
>    cram-hmac-alg sha1;
>    shared-secret "MySecret123";
>  }
>  on server1 {
>    device    /dev/drbd0;
>    disk      /dev/sdb;
>    address   192.168.111.10:7789;
>    meta-disk internal;
>  }
>  on server2 {
>    device    /dev/drbd0;
>    disk      /dev/sdb;
>    address   192.168.111.11:7789;
>    meta-disk internal;
>  }
> }
>
>
> After any change in the /etc/drbd.d/mysql.res file I issued a "drbdadm
> adjust mysql" on both nodes.
>
> Test #1
> DRBD partition on primary (secondary node disabled)
> Using Base DRBD config
> # dd if=/dev/zero of=/var/lib/mysql/TMP/disk-test.xxx bs=1M count=4096
> oflag=direct
> Throughput ~ 420MB/s
>
> Test #2
> DRBD partition on primary (secondary node enabled)
> Using Base DRBD config
> # dd if=/dev/zero of=/var/lib/mysql/TMP/disk-test.xxx bs=1M count=4096
> oflag=direct
> Throughput ~ 61MB/s
>
> Test #3
> DRBD partition on primary (secondary node enabled)
> Using Base DRBD config with:
>  Protocol B;
> # dd if=/dev/zero of=/var/lib/mysql/TMP/disk-test.xxx bs=1M count=4096
> oflag=direct
> Throughput ~ 68MB/s
>
> Test #4
> DRBD partition on primary (secondary node enabled)
> Using Base DRBD config with:
>  Protocol A;
> # dd if=/dev/zero of=/var/lib/mysql/TMP/disk-test.xxx bs=1M count=4096
> oflag=direct
> Throughput ~ 94MB/s
>
> Test #5
> DRBD partition on primary (secondary node enabled)
> Using Base DRBD config with:
>  disk {
>    disk-barrier no;
>    disk-flushes no;
>    md-flushes no;
>  }
> # dd if=/dev/zero of=/var/lib/mysql/TMP/disk-test.xxx bs=1M count=4096
> oflag=direct
> Disk throughput ~ 62MB/s
>
> No difference from Test #2 really. Also cat /proc/drbd still shows wo:b in
> both cases so I'm not even sure
> these disk {..} parameters have been taken into account...
>
> Test #6
> DRBD partition on primary (secondary node enabled)
> Using Base DRBD config with:
>  Protocol B;
>  disk {
>    disk-barrier no;
>    disk-flushes no;
>    md-flushes no;
>  }
> # dd if=/dev/zero of=/var/lib/mysql/TMP/disk-test.xxx bs=1M count=4096
> oflag=direct
> Disk throughput ~ 68MB/s
>
> No difference from Test #3 really. Also cat /proc/drbd still shows wo:b in
> both cases so I'm not even sure
> these disk {..} parameters have been taken into account...
>
>
> What else can I try?
> Is it worth trying DRBD 8.3.x?
>
> Thx.
>
> Fred
>
>
>
>
>
>
> On 1 Feb 2012, at 08:35, James Harper wrote:
>
> >> Hi
> >>
> >> I've configured DRBD with a view to use it with MySQL (and later on
> >> Pacemaker + Corosync) in a 2 nodes primary/secondary
> >> (master/slave) setup.
> >>
> >> ...
> >>
> >> No replication over the 1Gb/s crossover cable is taking place since the
> >> secondary node is down yet there's x2 lower disk performance.
> >>
> >> I've tried to add:
> >>  disk {
> >>    disk-barrier no;
> >>    disk-flushes no;
> >>    md-flushes no;
> >>  }
> >> to the config but it didn't seem to change anything.
> >>
> >> Am I missing something here?
> >> On another note is 8.4.1 the right version to use?
> >>
> >
> > If you can do it just for testing, try changing to protocol B with one
> primary and one secondary and see how that impacts your performance, both
> with barrier/flushes on and off. I'm not sure if it will help but if
> protocol B makes things faster then it might hint as to where to start
> looking...
> >
> > James
>
> _______________________________________________
> drbd-user mailing list
> drbd-user at lists.linbit.com
> http://lists.linbit.com/mailman/listinfo/drbd-user
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20120201/fa50d9b7/attachment.htm>


More information about the drbd-user mailing list