Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Am Dienstag, den 03.04.2012, 12:03 +0200 schrieb Felix Frank: > Hi, > > On 04/03/2012 11:53 AM, Lukas Gradl wrote: > > For use with KVM with automatic failover I need a primary/primary setup, > > so AFAIK protocol C is required. > > For dual-primary it is required, yes. You do need dual-primary for live > migrations. You do *not* need it for automatic failover (in failure > scenarios, live migration won't do you any good, anyway). > > If live migration isn't an issue for you, single-primary is perfectly > fine! You still want protocol C though :-) > > > According to my benchmarks DRBD is much slower in that setup than native > > HDD performance and changing the Network-Setup from 1GBit direct link to > > 2 bonded interfaces doesn't increase speed. > > Have you identified the exact bottleneck inside your DRBD setup? > Have you done analysis according to > http://www.drbd.org/users-guide/ch-benchmark.html? Yes. I benchmarked exactly as described in that doc. Throughput values don't change really - 85MB/s on the raw device, 83MB/s on the drbd-device. (average of 5 times "dd if=/dev/zero of=/dev/drbd1 bs=512M count=1 oflag=direct") But latency drops: for the 1000 512B blocks it took 0,05397s to write on the raw device and 12,757s on the drbd device (average of 5 times "dd if=/dev/zero of=/dev/drbd1 bs=512 count=1000 oflag=direct") Additionally I tried drbd with internal metadata on the sata-disk and external metadata on the Boot-SSD - there were no significant changes. My drbd.conf looks like this: global { usage-count no; } common { protocol C; syncer { rate 120M; al-extents 3389; } startup { wfc-timeout 15; degr-wfc-timeout 60; become-primary-on both; } net { cram-hmac-alg sha1; shared-secret "secret"; allow-two-primaries; after-sb-0pri discard-zero-changes; after-sb-1pri discard-secondary; after-sb-2pri disconnect; sndbuf-size 512k; } } resource r0 { on vm01 { device /dev/drbd0; disk /dev/sdb3; address 10.254.1.101:7780; meta-disk /dev/sda3[0]; } on vm02 { device /dev/drbd0; disk /dev/sdb3; address 10.254.1.102:7780; meta-disk /dev/sda3[0]; } } resource r1 { on vm01 { device /dev/drbd1; disk /dev/sdb1; address 10.254.1.101:7781; meta-disk internal; } on vm02 { device /dev/drbd1; disk /dev/sdb1; address 10.254.1.102:7781; meta-disk internal; } } The both nodes are linkes by a direct Gigabit connection used by drbd exclusively. > > > What do the experts think: Should this be sufficient to get the > > perfomance of a single SATA-Disk without DRBD? > > I don't really feel addressed ;-) but here's my 2 cents: > > If DRBD performance with rotational disks is dissatisfactory, I wouldn't > count on faster disks somehow solving the problem. You *may* save enough > latency to make the setup worthwhile, but myself, I'd rather keep trying > to root out the main problem. I would like to do so - but I've no real idea what the problem might be. regards Lukas -- -------------------------- software security networks Lukas Gradl <proxmox#ssn.at> Eduard-Bodem-Gasse 6 A - 6020 Innsbruck Tel: +43-512-214040-0 Fax: +43-512-214040-21 --------------------------