[DRBD-user] slow i/o performance

EDV at bauerfunken.de EDV at bauerfunken.de
Mon Apr 15 12:47:22 CEST 2013

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Hello all,

I'm getting problems with the performance of my drbd installation.

I'm using two servers with drbd 8.3.7 in primary/slave configuration installed on openfiler 2.3.

When I'm copying one file of 2GB directly on the drbd-device I'm getting an i/o of 13MB/s as maximum.

This drbd-device is also used as XEN-shared storage with 8 VM's on it. Now in normal usage the i/o of the
same testing gets 2.6MB/s.

I'm monitored the direct network connection for syncing both nodes and get 236MBit/s or 29.5MB/s
as maximum. This connection uses 2 1GBit network adaptors with crosslink-cable.

My configuration file drbd.conf

global {
 # minor-count 64;
 # dialog-refresh 5; # 5 seconds
 # disable-ip-verification;
 usage-count ask;
}

common {
 syncer { rate 100M; }
}

resource cluster_metadata {
 protocol C;
 handlers {
  pri-on-incon-degr "echo O > /proc/sysrq-trigger ; halt -f";
  pri-lost-after-sb "echo O > /proc/sysrq-trigger ; halt -f";
  local-io-error "echo O > /proc/sysrq-trigger ; halt -f";
  # outdate-peer "/usr/sbin/drbd-peer-outdater";
 }

 startup {
  # wfc-timeout 0;
  degr-wfc-timeout 120; # 2 minutes.
 }

 disk {
  on-io-error detach;
 }

 net {
  after-sb-0pri disconnect;
  after-sb-1pri disconnect;
  after-sb-2pri disconnect;
  rr-conflict disconnect;
 }

 syncer {
  # rate 10M;
  # after "r2";
  al-extents 257;
 }

 on filer1 {
  device /dev/drbd0;
  disk /dev/sdc1;
  address 10.3.2.1:7788;
  meta-disk internal;
 }

 on filer2 {
  device /dev/drbd0;
  disk /dev/sdc1;
  address 10.3.2.2:7788;
  meta-disk internal;
 }
}

resource vg0drbd {
 protocol C;
 startup {
  wfc-timeout 0; ## Infinite!
  degr-wfc-timeout 120; ## 2 minutes.
 }

 disk {
  no-disk-flushes;
  on-io-error detach;
 }

 net {
  # timeout 60;
  # connect-int 10;
  # ping-int 10;
  # max-buffers 2048;
  # max-epoch-size 2048;
  sndbuf-size 0;
 }

 syncer {
  after "cluster_metadata";
 }

 on filer1 {
  device /dev/drbd1;
  disk /dev/sdc2;
  address 10.3.2.1:7789;
  meta-disk internal;
 }

 on filer2 {
  device /dev/drbd1;
  disk /dev/sdc2;
  address 10.3.2.2:7789;
  meta-disk internal;
 }
}

drbdsetup drbd1 show says:

disk {
      size              0s _is_default; # bytes
      on-io-error       detach;
      fencing           dont-care _is_default;
      no-disk-flushes ;
      max-bio-bvecs     0 _is_default;
}
net {
      timeout           60 _is_default; # 1/10 seconds
      max-epoch-size    2048 _is_default;
      max-buffers       2048 _is_default;
      unplug-watermark  128 _is_default;
      connect-int       10 _is_default; # seconds
      ping-int          10 _is_default; # seconds
      sndbuf-size       0 _is_default; # bytes
      rcvbuf-size       0 _is_default; # bytes
      ko-count          0 _is_default;
      after-sb-0pri     disconnect _is_default;
      after-sb-1pri     disconnect _is_default;
      after-sb-2pri     disconnect _is_default;
      rr-conflict       disconnect _is_default;
      ping-timeout      5 _is_default; # 1/10 seconds
}
syncer {
      rate              102400k; # bytes/second
      after             0;
      al-extents        127 _is_default;
}
protocol C;
_this_host {
      device                  minor 1;
      disk              "/dev/sdc2";
      meta-disk         internal;
      address                 ipv4 10.3.2.1:7789;
}
_remote_host {
      address                 ipv4 10.3.2.2:7789;
}

Under my volume drbd1 I'm using a RAID5-system with an Adaptec AAC-RAID
controller and 1.8TB disc-capacity.

Can anyone help me to increase the i/o performance ?

Best regards

Andreas Weigand

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20130415/2e425c93/attachment.htm>


More information about the drbd-user mailing list