Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
On Thu, 02 Dec 2010 13:36:25 -0600, J. Ryan Earl wrote:
> If you "gracefully" stop DRBD on one
> node, it's not "degraded."  Degraded is from like a non-graceful
> separation due to a crash, power-outage, network issue, etc where one
> end detects the other is gone instead of being told to gracefully close
> connection between nodes.
I issued a "stop" (a graceful shutdown) only after I broke the DRBD 
connection by blocking the relevant packets.  So before the stop, the 
cluster was in a degraded state:
 1: cs:WFConnection ro:Primary/Unknown ds:UpToDate/DUnknown C r----
    ns:0 nr:0 dw:0 dr:0 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:b oos:0
Using "stop" still causes a clean shutdown which then avoids degr-wfc-
timeout?
Is there any way that a network issue, or anything else short of a crash 
of the system, can invoke degr-wfc-timeout?  I've even tried 'kill -9' of 
the drbd processes, but they seems immune to this.
I can force a system crash if I have to, but that's something of pain in 
the neck so I'd prefer another option if one is available.
Or have I misunderstood?  I've been assuming that degr-wfc-timeout 
applies only to the WFC at startup (because the timeout value is in the 
startup block of the configuration file).  Is this controlling some other 
WFC period?
When I break the connection (and with no extra fencing logic specified), 
I see that both nodes go into a WFC state.  But this is lasting well 
longer than the 60 seconds I have defined in degr-wfc-timeout.
Thanks...
	Andrew