Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hi everyone, I have a question about removing the primary node with DRBD9. "Is it normal that we only got 20% success rate of removing the primary node?" Version of DRBD kernel source is the newest version(9.0.4-1).Version of DRBD utils is 8.9.6. Description: 3 nodes, one of the nodes is primary,disk state is UpToDate. I got an error message 'State change failed: (-2) Need access to UpToDate data' when executing the command 'drbdadm down <res-name>' on the primary node. After excuting the command for the first time, the status of the primary node is the following: # drbdadm status r0 r0 role:Secondary disk:UpToDate After executing the command for the second time,the status of the primary node is empty, then the primary node is removed successfully the time. # drbdadm status r0 <NOTHING TO PRINT>. Analysis: We found the following code that report the error. __is_valid_soft_transition() { ... if (device->open_ro_cnt && any_disk_up_to_date[OLD] && !any_disk_up_to_date[NEW]){ return SS_NO_UP_TO_DATE_DISK; } ... } Then we done 20 runs of the test: 1)start the node 2)promote the node 3)remove the node(drbdadm down <res-name>) Only got a 20% success rate of step 3). Fail happens: drbd_open(device->open_ro_cnt is set to non-zero)-->if condition is satisfied-->drbd_release(device->open_ro_cnt is set to zero) Success happends: drbd_open(device->open_ro_cnt is set to non-zero)-->drbd_release(device->open_ro_cnt is set to zero)-->if condition is not satisfied -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20160910/c77558ba/attachment.htm>