Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hi, i've got a third setup running with drbd9 right now and i'm wondering that performance in this new version of drbd is that slow. I see a lot of traffic going on and then suddenly the traffic is slowing down. drbdadm show then "congested:yes" on Secondard and "blocker:lower" on primary device. the network is a gigabit network and with the same servers and drbd8.3 i was able to get at least 80mb/s out of it. When the status is like this the performance is near to 2mb/s for arround 2-5 minutes, then it boosts up for 20sec and slowing down again. I also tested it with a crosslink connection. The setup is quit normal and nothing tweaked yet. Couldn't find the bottleneck. global_common.conf: global { #usage-count yes; # minor-count dialog-refresh disable-ip-verification # cmd-timeout-short 5; cmd-timeout-medium 121; cmd-timeout-long 600; } common { handlers { # These are EXAMPLE handlers only. # They may have severe implications, # like hard resetting the node under certain circumstances. # Be careful when chosing your poison. # pri-on-incon-degr "/usr/lib/drbd/notify-pri-on-incon-degr.sh; /usr/lib/drbd/notify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot -f"; # pri-lost-after-sb "/usr/lib/drbd/notify-pri-lost-after-sb.sh; /usr/lib/drbd/notify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot -f"; # local-io-error "/usr/lib/drbd/notify-io-error.sh; /usr/lib/drbd/notify-emergency-shutdown.sh; echo o > /proc/sysrq-trigger ; halt -f"; # fence-peer "/usr/lib/drbd/crm-fence-peer.sh"; # split-brain "/usr/lib/drbd/notify-split-brain.sh root"; # out-of-sync "/usr/lib/drbd/notify-out-of-sync.sh root"; # before-resync-target "/usr/lib/drbd/snapshot-resync-target-lvm.sh -p 15 -- -c 16k"; # after-resync-target /usr/lib/drbd/unsnapshot-resync-target-lvm.sh; } startup { wfc-timeout 20; degr-wfc-timeout 20; # outdated-wfc-timeout wait-after-sb } options { # cpu-mask on-no-data-accessible } disk { # size on-io-error fencing disk-barrier disk-flushes # disk-drain md-flushes resync-rate resync-after al-extents # c-plan-ahead c-delay-target c-fill-target c-max-rate # c-min-rate disk-timeout } net { # protocol timeout max-epoch-size max-buffers unplug-watermark # connect-int ping-int sndbuf-size rcvbuf-size ko-count # allow-two-primaries cram-hmac-alg shared-secret after-sb-0pri # after-sb-1pri after-sb-2pri always-asbp rr-conflict # ping-timeout data-integrity-alg tcp-cork on-congestion # congestion-fill congestion-extents csums-alg verify-alg # use-rle protocol C; } } r0.res: resource r0 { net { cram-hmac-alg sha1; shared-secret "Cwzvaerfyficjdh6"; } volume 0 { device /dev/drbd0; disk /dev/sda3; meta-disk internal; } on node1 { node-id 0; address 10.0.0.1:7000; } on node2 { node-id 1; address 10.0.0.2:7000; } connection { host node1 port 7000; host node2 port 7000; net { protocol C; } } } would be nice if anyone can help me with this issue. kind regards