Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hello, Getting some "strange" results on write testing on DRBD Primary. Every time I get more data written than 1Gb link can handle. Get about 133 MB/s with 1Gb link saturated and both Nodes in sync. Also If I make a test with files of size less that 1GB ( for example 900MB ) I always get results between 650-750MB/s and does not meter which sync protocol I use in DRBD. Does this have something to do with DRBD realization ? Buffers or something ... Here is my configuration file: global { usage-count no; } resource repdata { protocol B; startup { wfc-timeout 0; degr-wfc-timeout 100; } disk { on-io-error detach; no-disk-barrier; no-disk-flushes; no-md-flushes; } net { max-buffers 20000; max-epoch-size 20000; unplug-watermark 1024; sndbuf-size 0; cram-hmac-alg "sha1"; shared-secret "secret"; data-integrity-alg "crc32c"; } syncer { rate 35M; verify-alg "crc32c"; csums-alg "crc32c"; } on node1 { device /dev/drbd0; disk /dev/sda9; address 10.10.1.141:7788; meta-disk internal; } on node2 { device /dev/drbd0; disk /dev/sda8; address 10.10.1.142:7788; meta-disk internal; } } OS: CentOS 5.6 x86_64 RAID controller H700 with 1GB cache. -- Maxim Ianoglo