[DRBD-user] slow sync speed
Adam Weremczuk
adamw at matrixscience.com
Thu Oct 11 15:43:39 CEST 2018
Hi Oleksiy,
Here is my entire current config:
global { usage-count no; }
common { protocol C; }
resource r0 {
disk {
on-io-error detach;
no-disk-flushes ;
no-disk-barrier;
c-plan-ahead 0;
c-fill-target 24M;
c-min-rate 80M;
c-max-rate 720M;
}
net {
max-buffers 36k;
sndbuf-size 1024k;
rcvbuf-size 2048k;
}
on node1 {
device /dev/drbd0;
disk /dev/sdb1;
address 192.168.200.1:7788;
meta-disk internal;
}
on node2 {
device /dev/drbd0;
disk /dev/sdb1;
address 192.168.200.2:7788;
meta-disk internal;
}
}
...and the speed is still dreadfully slow, even though the link can
easily do 150MB/s:
cat /proc/drbd
version: 8.4.10 (api:1/proto:86-101)
srcversion: 17A0C3A0AF9492ED4B9A418
0: cs:SyncSource ro:Primary/Secondary ds:UpToDate/Inconsistent C r-----
ns:46712 nr:0 dw:0 dr:47528 al:8 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:d
oos:3898265768
[>....................] sync'ed: 0.1% (3806900/3806944)M
finish: 3867:19:37 speed: 264 (260) K/sec
Anything else to try before giving up on DRBD?
Thanks,
Adam
On 11/10/18 07:45, Oleksiy Evin wrote:
> You may try to disable dynamic sync rate by setting "c-plan-ahead" to
> 0 and increase "max-buffers". That's the only way which helped me to
> get reasonable sync rate with 100GigE connection.
>
> net {
> ...
> max-buffers 32K;
> # max-epoch-size 18K;
> }
>
> disk {
> ...
> c-plan-ahead 0;
> }
>
> You can find some more info here:
> https://serverfault.com/questions/740311/drbd-terrible-sync-performance-on-10gige/740370
>
>
> //OE
>
> -----Original Message-----
> *From*: Adam Weremczuk <adamw at matrixscience.com
> <mailto:Adam%20Weremczuk%20%3cadamw at matrixscience.com%3e>>
> *To*: drbd-user at lists.linbit.com <mailto:drbd-user at lists.linbit.com>
> *Subject*: [DRBD-user] slow sync speed
> *Date*: Wed, 10 Oct 2018 14:57:02 +0100
>
> Hi all,
> I'm trying out DRBD Pacemaker HA Cluster on Proxmox 5.2
> I have 2 identical servers connected with 2 x 1 Gbps links in bond_mode
> balance-rr.
> The bond is working fine; I get a transfer rate of 150 MB/s with scp.
> Following this guide:
> https://www.theurbanpenguin.com/drbd-pacemaker-ha-cluster-ubuntu-16-04/
> was going smoothly up until:
> drbdadm -- --overwrite-data-of-peer primary r0/0
> cat /proc/drbd
> version: 8.4.10 (api:1/proto:86-101)
> srcversion: 17A0C3A0AF9492ED4B9A418
> 0: cs:SyncSource ro:Primary/Secondary ds:UpToDate/Inconsistent C r-----
> ns:10944 nr:0 dw:0 dr:10992 al:8 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:f
> oos:3898301536
> [>....................] sync'ed: 0.1% (3806932/3806944)M
> finish: 483:25:13 speed: 2,188 (2,188) K/sec
> The transfer rate is horribly slow and at this pace it's going to take
> 20 days for two 4 TB volumes to sync!
> That's almost 15 times slower comparing with the guide video (8:30):
> https://www.youtube.com/watch?v=WQGi8Nf0kVc
> The volumes have been zeroed and contain no live data yet.
> My sdb disks are logical drives (hardware RAID) set up as RAID50 with
> the defaults:
> Strip size: 128 KB
> Access policy: RW
> Read policy: Normal
> Write policy: Write Back with BBU
> IO policy: Direct
> Drive Cache: Disable
> Disable BGI: No
> Performance looks good when tested with hdparm:
> hdparm -tT /dev/sdb1
> /dev/sdb1:
> Timing cached reads: 15056 MB in 1.99 seconds = 7550.46 MB/sec
> Timing buffered disk reads: 2100 MB in 3.00 seconds = 699.81 MB/sec
> The volumes have been zeroed and contain no live data yet.
> It seems to be a problem with default DRBD settings.
> Can anybody recommend optimal tweaks specific to my environment?
> Regards,
> Adam
> _______________________________________________
> drbd-user mailing list
> drbd-user at lists.linbit.com <mailto:drbd-user at lists.linbit.com>
> http://lists.linbit.com/mailman/listinfo/drbd-user
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20181011/a710bc7f/attachment.htm>
More information about the drbd-user
mailing list