Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Forgot to repost configs:
drbd.d/global_common.conf:
global {
usage-count yes;
}
common {
protocol C;
handlers {
pri-on-incon-degr
"/usr/lib/drbd/notify-pri-on-incon-degr.sh;
/usr/lib/drbd/notify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ;
reboot -f";
pri-lost-after-sb
"/usr/lib/drbd/notify-pri-lost-after-sb.sh;
/usr/lib/drbd/notify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ;
reboot -f";
local-io-error "/usr/lib/drbd/notify-io-error.sh;
/usr/lib/drbd/notify-emergency-shutdown.sh; echo o > /proc/sysrq-trigger ;
halt -f";
}
startup {
# wfc-timeout degr-wfc-timeout outdated-wfc-timeout
wait-after-sb
}
disk {
on-io-error detach;
}
net {
after-sb-0pri discard-zero-changes;
after-sb-1pri consensus;
after-sb-2pri disconnect;
ko-count 6;
}
syncer {
rate 50M;
csums-alg crc32c;
verify-alg md5;
use-rle;
}
}
drbd.conf:
# You can find an example in /usr/share/doc/drbd.../drbd.conf.example
include "drbd.d/global_common.conf";
include "drbd.d/*.res";
resource meta_lower {
disk /dev/backingvg/metabacking;
device /dev/drbd0;
meta-disk internal;
disk {
fencing resource-only;
}
handlers {
fence-peer "/usr/lib/drbd/crm-fence-peer.sh";
after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh";
}
on openfiler1 {
address 10.50.153.1:7788;
}
on openfiler2 {
address 10.50.153.2:7788;
}
}
resource data4_lower {
device /dev/drbd4;
disk /dev/backingvg/1024data4backing;
meta-disk internal;
disk {
fencing resource-only;
}
handlers {
fence-peer "/usr/lib/drbd/crm-fence-peer.sh";
after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh";
}
on openfiler1 {
address 10.50.153.1:7792;
}
on openfiler2 {
address 10.50.153.2:7792;
}
}
resource data5_lower {
device /dev/drbd5;
disk /dev/backingvg/2048data5backing;
meta-disk internal;
disk {
fencing resource-only;
}
handlers {
fence-peer "/usr/lib/drbd/crm-fence-peer.sh";
after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh";
}
on openfiler1 {
address 10.50.153.1:7793;
}
on openfiler2 {
address 10.50.153.2:7793;
}
}
resource meta {
protocol A;
device /dev/drbd10;
meta-disk internal;
handlers {
before-resync-target "/usr/lib/drbd/snapshot-resync-target-lvm.sh";
after-resync-target "/usr/lib/drbd/unsnapshot-resync-target-lvm.sh";
}
net {
sndbuf-size 512k;
on-congestion pull-ahead;
congestion-fill 500k;
}
syncer {
rate 1000k;
}
stacked-on-top-of meta_lower {
address 10.50.150.101:7788;
}
on openfiler3 {
disk /dev/backingvg/metabacking;
address 10.50.250.4:7788;
}
}
resource data4 {
protocol A;
device /dev/drbd14;
meta-disk internal;
handlers {
before-resync-target "/usr/lib/drbd/snapshot-resync-target-lvm.sh";
after-resync-target "/usr/lib/drbd/unsnapshot-resync-target-lvm.sh";
}
net {
data-integrity-alg md5;
sndbuf-size 512k;
on-congestion pull-ahead;
congestion-fill 500k;
}
syncer {
rate 2500k;
}
stacked-on-top-of data4_lower {
address 10.50.150.101:7792;
}
on openfiler3 {
disk /dev/backingvg/1024data4backing;
address 10.50.250.4:7792;
}
}
resource data5 {
protocol A;
device /dev/drbd15;
meta-disk internal;
handlers {
before-resync-target "/usr/lib/drbd/snapshot-resync-target-lvm.sh";
after-resync-target "/usr/lib/drbd/unsnapshot-resync-target-lvm.sh";
}
net {
sndbuf-size 512k;
on-congestion pull-ahead;
congestion-fill 500k;
}
syncer {
rate 2500k;
}
stacked-on-top-of data5_lower {
address 10.50.150.101:7793;
}
on openfiler3 {
disk /dev/backingvg/2048data5backing;
address 10.50.250.4:7793;
}
}
--
View this message in context: http://old.nabble.com/drbd-resource-ahead---behind-problem-tp33454636p33544661.html
Sent from the DRBD - User mailing list archive at Nabble.com.