Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
I'm in the process of building my cluster. I have one node up named a.example.org. However, when I boot this machine (a), the boot process blocks with this message: DRBD's startup script waits for the peer node(s) to appear. - In case this node was already a degraded cluster before the reboot the timeout is 120 seconds. [degr-wfc-timeout] - If the peer was available before the reboot the timeout will expire after 0 seconds. [wfc-timeout] (These values are for resource 'vm1-root'; 0 sec -> wait forever) To abort waiting enter 'yes' [ 340]: In my case, the peer (I assume b is the peer) has never been available. Therefore, this node should have been a degraded cluster before the reboot, hence, the timeout should be 120 seconds, but the counter keeps on ticking and ticking, so it seems there's a problem here. Did I do something wrong? How do I convince a that it is a degraded cluster and to come up with my drbd devices ready? drbdadm -v reports this: Version: 8.0.13 (api:86) GIT-hash: ee3ad77563d2e87171a3da17cc002ddfd1677dbe build by phil at fat-tyre, 2008-08-04 15:28:07 I have the following in my drbd.conf file (taken from the sample.conf): resource vm2-root { protocol C; handlers { pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f"; pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f"; local-io-error "echo o > /proc/sysrq-trigger ; halt -f"; outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater -t 5"; } startup { } disk { on-io-error detach; } net { after-sb-0pri disconnect; after-sb-1pri disconnect; after-sb-2pri disconnect; rr-conflict disconnect; } syncer { rate 100M; al-extents 257; } on a.example.org { device /dev/drbd1; disk /dev/b/root2; address 10.0.0.1:7790; meta-disk internal; } on b.example.org { device /dev/drbd1; disk /dev/b/root2; address 10.0.0.2:7790; meta-disk internal; } }