Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Did you test your network connection to make sure that it can transfer
at a greater speed than that? Maybe it is the bottleneck - jumbo frames on?
On 25/04/12 15:40, Chris Dickson wrote:
> Also use oflag=direct in both tests and perform them a few times,
> sometimes high speeds are the result of caching.
>
> On Wed, Apr 25, 2012 at 10:35 AM, Chris Dickson <chrisd1100 at gmail.com
> <mailto:chrisd1100 at gmail.com>> wrote:
>
> Try turning off disk-barrier and disk-flushes and see if that
> makes a difference.
>
>
> 2012/4/25 feng zheng <zf5984599 at gmail.com
> <mailto:zf5984599 at gmail.com>>
>
> hi, dear all:
>
> When I use drbd, I found the write performance very slow, against
> testing without drbd module.
>
> 1. the environment:
> -) CentOS 5.6
> -) 2.6.18 kernel
> -) drbd 8.4.1
> -) drbd.conf:
> resource r0
> {
> protocol B;
>
> net
> {
> max-buffers 8000;
> max-epoch-size 8000;
> sndbuf-size 512K;
> }
>
> disk
> {
> al-extents 3389;
> }
>
>
> on OSS211
> {
> device /dev/drbd0;
> disk /dev/sdb1;
> address *MailScanner warning: numerical links are often
> malicious:* 192.168.100.231:7788 <http://192.168.100.231:7788>;
> meta-disk internal;
> }
>
> on OSS213
> {
> device /dev/drbd0;
> disk /dev/sde1;
> address *MailScanner warning: numerical links are often
> malicious:* 192.168.100.213:7788 <http://192.168.100.213:7788>;
> meta-disk internal;
> }
>
> }
>
>
> 2. Test scenario:
> *) without drbd module,
> dd to write 1G stream into one disk, which formatted to ext3:
> [para]# !echo
> echo 3 > /proc/sys/vm/drop_caches
> [para]# !dd
> dd if=/dev/zero of=test1 bs=1M count=1000 conv=fdatasync
> 1000+0 records in
> 1000+0 records out
> 1048576000 bytes (1.0 GB) copied, 10.9905 seconds, 95.4 MB/s
>
> *) with drbd module,
> dd 1G stream to the disk, which is ext3 too:
> [para]# cat /proc/drbd
> version: 8.4.1 (api:1/proto:86-100)
> GIT-hash: 91b4c048c1a0e06777b5f65d312b38d47abaea80 build by
> root at OSS213, 2012-04-16 21:38:36
> 0: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate B
> r-----
> ns:1260036 nr:0 dw:1260036 dr:297 al:330 bm:0 lo:0 pe:0
> ua:0 ap:0
> ep:1 wo:b oos:0
> [para]#
> [para]# dd if=/dev/zero of=test1 bs=1M count=1000 conv=fdatasync
> 1000+0 records in
> 1000+0 records out
> 1048576000 bytes (1.0 GB) copied, 26.7392 seconds, 39.2 MB/s
> [para]# cat /proc/drbd
> version: 8.4.1 (api:1/proto:86-100)
> GIT-hash: 91b4c048c1a0e06777b5f65d312b38d47abaea80 build by
> root at OSS213, 2012-04-16 21:38:36
> 0: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate B
> r-----
>
> All the upper test writing disk are same. From the upper
> result, if i
> use DRBD to test, the performance
> is 39 MB/s; while if i do not use, the performance is about 95M/s.
>
> 3. My question is:
> -) this write performance decays so large is normal or not?
> I had read the following from the DRBD website:
> "15.1. Hardware considerations:
> .... A single, reasonably recent, SCSI or SAS disk will
> typically allow streaming writes of roughly 40MB/s to the
> single disk."
> But this is very slow.
>
> -)if this is not normal, how can i turn this? is the config file
> something not correct?
>
> thanks a lot
> BRs,
> feng
> _______________________________________________
> drbd-user mailing list
> drbd-user at lists.linbit.com <mailto:drbd-user at lists.linbit.com>
> http://lists.linbit.com/mailman/listinfo/drbd-user
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20120425/d917e602/attachment.htm>