[DRBD-user] slow sync speed

Adam Weremczuk adamw at matrixscience.com
Thu Oct 11 15:43:39 CEST 2018


Hi Oleksiy,

Here is my entire current config:

global { usage-count no; }
common { protocol C; }
resource r0 {
   disk {
     on-io-error detach;
     no-disk-flushes ;
     no-disk-barrier;
     c-plan-ahead 0;
     c-fill-target 24M;
     c-min-rate 80M;
     c-max-rate 720M;
   }
   net {
     max-buffers 36k;
     sndbuf-size 1024k;
     rcvbuf-size 2048k;
   }
   on node1 {
     device /dev/drbd0;
     disk /dev/sdb1;
     address 192.168.200.1:7788;
     meta-disk internal;
   }
   on node2 {
     device /dev/drbd0;
     disk /dev/sdb1;
     address 192.168.200.2:7788;
     meta-disk internal;
   }
}

...and the speed is still dreadfully slow, even though the link can 
easily do 150MB/s:

cat /proc/drbd
version: 8.4.10 (api:1/proto:86-101)
srcversion: 17A0C3A0AF9492ED4B9A418
  0: cs:SyncSource ro:Primary/Secondary ds:UpToDate/Inconsistent C r-----
     ns:46712 nr:0 dw:0 dr:47528 al:8 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:d 
oos:3898265768
     [>....................] sync'ed:  0.1% (3806900/3806944)M
     finish: 3867:19:37 speed: 264 (260) K/sec

Anything else to try before giving up on DRBD?

Thanks,
Adam


On 11/10/18 07:45, Oleksiy Evin wrote:
> You may try to disable dynamic sync rate by setting "c-plan-ahead" to 
> 0 and increase "max-buffers". That's the only way which helped me to 
> get reasonable sync rate with 100GigE connection.
>
> net {
> ...
> max-buffers 32K;
> # max-epoch-size 18K;
> }
>
> disk {
> ...
> c-plan-ahead 0;
> }
>
> You can find some more info here:
> https://serverfault.com/questions/740311/drbd-terrible-sync-performance-on-10gige/740370
>
>
> //OE
>
> -----Original Message-----
> *From*: Adam Weremczuk <adamw at matrixscience.com 
> <mailto:Adam%20Weremczuk%20%3cadamw at matrixscience.com%3e>>
> *To*: drbd-user at lists.linbit.com <mailto:drbd-user at lists.linbit.com>
> *Subject*: [DRBD-user] slow sync speed
> *Date*: Wed, 10 Oct 2018 14:57:02 +0100
>
> Hi all,
> I'm trying out DRBD Pacemaker HA Cluster on Proxmox 5.2
> I have 2 identical servers connected with 2 x 1 Gbps links in bond_mode
> balance-rr.
> The bond is working fine; I get a transfer rate of 150 MB/s with scp.
> Following this guide:
> https://www.theurbanpenguin.com/drbd-pacemaker-ha-cluster-ubuntu-16-04/  
> was going  smoothly up until:
> drbdadm -- --overwrite-data-of-peer primary r0/0
> cat /proc/drbd
> version: 8.4.10 (api:1/proto:86-101)
> srcversion: 17A0C3A0AF9492ED4B9A418
>    0: cs:SyncSource ro:Primary/Secondary ds:UpToDate/Inconsistent C r-----
>       ns:10944 nr:0 dw:0 dr:10992 al:8 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:f
> oos:3898301536
>       [>....................] sync'ed:  0.1% (3806932/3806944)M
>       finish: 483:25:13 speed: 2,188 (2,188) K/sec
> The transfer rate is horribly slow and at this pace it's going to take
> 20 days for two 4 TB volumes to sync!
> That's almost 15 times slower comparing with the guide video (8:30):
> https://www.youtube.com/watch?v=WQGi8Nf0kVc
> The volumes have been zeroed and contain no live data yet.
> My sdb disks are logical drives (hardware RAID) set up as RAID50 with
> the defaults:
> Strip size: 128 KB
> Access policy: RW
> Read policy: Normal
> Write policy: Write Back with BBU
> IO policy: Direct
> Drive Cache: Disable
> Disable BGI: No
> Performance looks good when tested with hdparm:
> hdparm -tT /dev/sdb1
> /dev/sdb1:
>    Timing cached reads:   15056 MB in  1.99 seconds = 7550.46 MB/sec
>    Timing buffered disk reads: 2100 MB in  3.00 seconds = 699.81 MB/sec
> The volumes have been zeroed and contain no live data yet.
> It seems to be a problem with default DRBD settings.
> Can anybody recommend optimal tweaks specific to my environment?
> Regards,
> Adam
> _______________________________________________
> drbd-user mailing list
> drbd-user at lists.linbit.com <mailto:drbd-user at lists.linbit.com>
> http://lists.linbit.com/mailman/listinfo/drbd-user

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20181011/a710bc7f/attachment.htm>


More information about the drbd-user mailing list