Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
I am doing some testing with drbd9 and drbdmanage and am seeing some
behaviour that I do not understand.
I have three nodes in a cluster. Node names are kvm09, kvm10 and kvm11.
kvm09 is the leader. All three systems are up to date Centos 7.3 with
drbd 9.0.6, drbd-utils 8.9.10 and drbdmanage 0.98.2
If I shutdown a node, drbdmanage commands executed on the leader now
take a "long time" to complete.
For instance if I shutdown kvm11 and run on the leader (kvm09) I see
[root at kvm09 ~]# time drbdmanage role
Operation completed successfully
leader
real 0m42.975s
user 0m0.084s
sys 0m0.019s
The same operation prior to shutting down kvm11 (times are similar on
all three nodes)
[root at kvm09 ~]# time drbdmanage role
Operation completed successfully
leader
real 0m0.120s
user 0m0.088s
sys 0m0.031s
On the non-leader node (kvm10) in this scenario the commands always
execute "quickly", but do not show up to date information.
For instance on the leader (which is correct and takes a long time)
[root at kvm09 ~]# time drbdmanage list-nodes
+---------------------------------------------------------------------------------------------------+
| Name | Pool Size | Pool Free |
| State |
|---------------------------------------------------------------------------------------------------|
| kvm09.devnull.net.nz | 1668076 | 1114992 |
| ok |
| kvm10.devnull.net.nz | 1668076 | 1069948 |
| ok |
| kvm11.devnull.net.nz | 1668076 | 1069948 |
| OFFLINE |
+---------------------------------------------------------------------------------------------------+
real 0m37.045s
user 0m0.079s
sys 0m0.026s
But on kvm10 (which is incorrect and does NOT take a long time)
[[root at kvm10 ~]# time drbdmanage list-nodes
+------------------------------------------------------------------------------------------------------------+
| Name | Pool Size | Pool Free |
| State |
|------------------------------------------------------------------------------------------------------------|
| kvm09.devnull.net.nz | 1668076 | 1114992 |
| ok |
| kvm10.devnull.net.nz | 1668076 | 1069948 |
| ok |
| kvm11.devnull.net.nz | 1668076 | 1069948 |
| ok |
+------------------------------------------------------------------------------------------------------------+
real 0m0.101s
user 0m0.071s
sys 0m0.025s
Is this expected behaviour in this type of situation?
Regards
Mike