Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hi, I'm trying to build a new cluster of servers using debian plus DRBD9 and drbdmanage. After a number of attempts, I thought I had everything right, and it's all been "ok" for a couple of weeks. Today, I rebooted both machines (new power being installed for the UPS), and then I tried to create a new volume of 700GB. Here is what I did: san2:~# drbdmanage add-volume oldNFS 700GB --deploy 2 Operation completed successfully Operation completed successfully san2:~# drbdmanage list-nodes +------------------------------------------------------------------------------------------------------------+ | Name | Pool Size | Pool Free | | State | |------------------------------------------------------------------------------------------------------------| | san2.websitemanagers.com.au | 3777040 | 3109316 | | ok | | san3 | 1830932 | 1830924 | | ok | +------------------------------------------------------------------------------------------------------------+ san2:~# drbdmanage list-volumes --show Port +------------------------------------------------------------------------------------------------------------+ | Name | Vol ID | Size | Minor | Port | | State | |------------------------------------------------------------------------------------------------------------| | oldNFS | 0 | 667572 | 101 | 7001 | | ok | | test1 | 0 | 9536 | 100 | 7000 | | ok | +------------------------------------------------------------------------------------------------------------+ san2:~# lvs LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert .drbdctrl_0 drbdpool -wi-ao---- 4.00m .drbdctrl_1 drbdpool -wi-ao---- 4.00m oldNFS_00 drbdpool -wi-ao---- 652.07g san2:~# dpkg -l | grep drbd ii drbd-utils 8.9.6-1 amd64 RAID 1 over TCP/IP for Linux (user utilities) ii python-drbdmanage 0.94-1 all DRBD distributed resource management utility san2:~# cat /proc/drbd version: 9.0.1-1 (api:2/proto:86-111) GIT-hash: f57acfc22d29a95697e683fb6bbacd9a1ad4348e build by root at san2.websitemanagers.com.au, 2016-03-01 00:38:53 Transports (api:14): tcp (1.0.0) *So far, everything looks good, so I thought to check out the other node, and see what is happening there....* root at san3:~# lvs LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert .drbdctrl_0 drbdpool -wi-ao---- 4.00m .drbdctrl_1 drbdpool -wi-ao---- 4.00m backup_system_20141006_193935 san1small -wi-a----- 8.00g swap san1small -wi-ao---- 3.72g system san1small -wi-ao---- 13.97g *Hmmm, that's strange, we don't have any new LV **here?* root at san3:~# drbdmanage list-nodes +------------------------------------------------------------------------------------------------------------+ | Name | Pool Size | Pool Free | | State | |------------------------------------------------------------------------------------------------------------| | san2.websitemanagers.com.au | 3777040 | 3109316 | | ok | | san3 | 1830932 | 1830924 | | ok | +------------------------------------------------------------------------------------------------------------+ root at san3:~# drbdmanage list-volumes --show Port +------------------------------------------------------------------------------------------------------------+ | Name | Vol ID | Size | Minor | Port | | State | |------------------------------------------------------------------------------------------------------------| | oldNFS | 0 | 667572 | 101 | 7001 | | ok | | test1 | 0 | 9536 | 100 | 7000 | | ok | +------------------------------------------------------------------------------------------------------------+ root at san3:~# dpkg -l | grep drbd ii drbd-utils 8.9.6-1 amd64 RAID 1 over TCP/IP for Linux (user utilities) ii python-drbdmanage 0.94-1 all DRBD distributed resource management utility root at san3:~# cat /proc/drbd version: 9.0.1-1 (api:2/proto:86-111) GIT-hash: f57acfc22d29a95697e683fb6bbacd9a1ad4348e build by root at san1, 2016-03-01 00:38:33 Transports (api:14): tcp (1.0.0) Reading more docs, I then find this section: http://www.drbd.org/doc/users-guide-90/s-check-status san2:~# drbd-overview 0:.drbdctrl/0 Connected(2*) Secondary(2*) UpToDa/UpToDa 1:.drbdctrl/1 Connected(2*) Secondary(2*) UpToDa/UpToDa 101:oldNFS/0 Connec/C'ting Second/Unknow UpToDa/DUnkno root at san3:/etc/drbd.d# drbd-overview 0:.drbdctrl/0 Connected(2*) Secondary(2*) UpToDa/UpToDa 1:.drbdctrl/1 Connected(2*) Secondary(2*) UpToDa/UpToDa So, it would seem that the problem is the config hasn't been sent to the other node, and it just doesn't know anything about it..... san2:~# drbdadm status oldNFS --verbose drbdsetup status oldNFS oldNFS role:Secondary disk:UpToDate san3 connection:Connecting Can anyone help advise where I should look, or what I might need to do to get this working? Thanks, Adam -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20160407/1e34d989/attachment.htm>