<html>
  <head>

    <meta http-equiv="content-type" content="text/html; charset=utf-8">
  </head>
  <body bgcolor="#FFFFFF" text="#000000">
    Hi,<br>
    <br>
    I'm trying to build a new cluster of servers using debian plus DRBD9
    and drbdmanage.<br>
    <br>
    After a number of attempts, I thought I had everything right, and
    it's all been "ok" for a couple of weeks.<br>
    <br>
    Today, I rebooted both machines (new power being installed for the
    UPS), and then I tried to create a new volume of 700GB.<br>
    <br>
    Here is what I did:<br>
    <tt>san2:~# drbdmanage add-volume oldNFS 700GB --deploy 2<br>
      Operation completed successfully<br>
      Operation completed successfully<br>
      san2:~# drbdmanage list-nodes<br>
+------------------------------------------------------------------------------------------------------------+<br>
      | Name                        | Pool Size | Pool Free
      |                                              | State |<br>
|------------------------------------------------------------------------------------------------------------|<br>
      | san2.websitemanagers.com.au |   3777040 |   3109316
      |                                              |    ok |<br>
      | san3                        |   1830932 |   1830924
      |                                              |    ok |<br>
+------------------------------------------------------------------------------------------------------------+<br>
      san2:~# drbdmanage list-volumes --show Port<br>
+------------------------------------------------------------------------------------------------------------+<br>
      | Name   | Vol ID |   Size | Minor | Port
      |                                                          | State
      |<br>
|------------------------------------------------------------------------------------------------------------|<br>
      | oldNFS |      0 | 667572 |   101 | 7001
      |                                                          |    ok
      |<br>
      | test1  |      0 |   9536 |   100 | 7000
      |                                                          |    ok
      |<br>
+------------------------------------------------------------------------------------------------------------+<br>
      san2:~# lvs<br>
        LV          VG       Attr       LSize   Pool Origin Data% 
      Meta%  Move Log Cpy%Sync Convert<br>
        .drbdctrl_0 drbdpool -wi-ao----  
      4.00m                                                    <br>
        .drbdctrl_1 drbdpool -wi-ao----  
      4.00m                                                    <br>
        oldNFS_00   drbdpool -wi-ao----
      652.07g                                                    <br>
      san2:~# dpkg -l | grep drbd<br>
      ii  drbd-utils                       8.9.6-1                      
      amd64        RAID 1 over TCP/IP for Linux (user utilities)<br>
      ii  python-drbdmanage                0.94-1                       
      all          DRBD distributed resource management utility<br>
      san2:~# cat /proc/drbd<br>
      version: 9.0.1-1 (api:2/proto:86-111)<br>
      GIT-hash: f57acfc22d29a95697e683fb6bbacd9a1ad4348e build by
      <a class="moz-txt-link-abbreviated" href="mailto:root@san2.websitemanagers.com.au">root@san2.websitemanagers.com.au</a>, 2016-03-01 00:38:53<br>
      Transports (api:14): tcp (1.0.0)<br>
      <br>
    </tt><b>So far, everything looks good, so I thought to check out the
      other node, and see what is happening there....</b><tt><br>
      <br>
      root@san3:~# lvs<br>
        LV                            VG        Attr       LSize  Pool
      Origin Data%  Meta%  Move Log Cpy%Sync Convert<br>
        .drbdctrl_0                   drbdpool  -wi-ao---- 
      4.00m                                                    <br>
        .drbdctrl_1                   drbdpool  -wi-ao---- 
      4.00m                                                    <br>
        backup_system_20141006_193935 san1small -wi-a----- 
      8.00g                                                    <br>
        swap                          san1small -wi-ao---- 
      3.72g                                                    <br>
        system                        san1small -wi-ao----
      13.97g                                                    <br>
      <br>
    </tt><b>Hmmm, that's strange, we don't have any new LV </b><b>here?</b><tt><br>
      <br>
      root@san3:~# drbdmanage list-nodes<br>
+------------------------------------------------------------------------------------------------------------+<br>
      | Name                        | Pool Size | Pool Free
      |                                              | State |<br>
|------------------------------------------------------------------------------------------------------------|<br>
      | san2.websitemanagers.com.au |   3777040 |   3109316
      |                                              |    ok |<br>
      | san3                        |   1830932 |   1830924
      |                                              |    ok |<br>
+------------------------------------------------------------------------------------------------------------+<br>
      root@san3:~# drbdmanage list-volumes --show Port<br>
+------------------------------------------------------------------------------------------------------------+<br>
      | Name   | Vol ID |   Size | Minor | Port
      |                                                          | State
      |<br>
|------------------------------------------------------------------------------------------------------------|<br>
      | oldNFS |      0 | 667572 |   101 | 7001
      |                                                          |    ok
      |<br>
      | test1  |      0 |   9536 |   100 | 7000
      |                                                          |    ok
      |<br>
+------------------------------------------------------------------------------------------------------------+<br>
      root@san3:~# dpkg -l | grep drbd<br>
      ii  drbd-utils                       8.9.6-1                      
      amd64        RAID 1 over TCP/IP for Linux (user utilities)<br>
      ii  python-drbdmanage                0.94-1                       
      all          DRBD distributed resource management utility<br>
      root@san3:~# cat /proc/drbd<br>
      version: 9.0.1-1 (api:2/proto:86-111)<br>
      GIT-hash: f57acfc22d29a95697e683fb6bbacd9a1ad4348e build by
      root@san1, 2016-03-01 00:38:33<br>
      Transports (api:14): tcp (1.0.0)<br>
      <br>
      Reading more docs, I then find this section:<br>
      <a class="moz-txt-link-freetext" href="http://www.drbd.org/doc/users-guide-90/s-check-status">http://www.drbd.org/doc/users-guide-90/s-check-status</a><br>
      san2:~# drbd-overview <br>
        0:.drbdctrl/0  Connected(2*) Secondary(2*) UpToDa/UpToDa <br>
        1:.drbdctrl/1  Connected(2*) Secondary(2*) UpToDa/UpToDa <br>
      101:oldNFS/0     Connec/C'ting Second/Unknow UpToDa/DUnkno <br>
      <br>
      <a class="moz-txt-link-abbreviated" href="mailto:root@san3:/etc/drbd.d#">root@san3:/etc/drbd.d#</a> drbd-overview <br>
       0:.drbdctrl/0  Connected(2*) Secondary(2*) UpToDa/UpToDa <br>
       1:.drbdctrl/1  Connected(2*) Secondary(2*) UpToDa/UpToDa <br>
      <br>
      So, it would seem that the problem is the config hasn't been sent
      to the other node, and it just doesn't know anything about it.....<br>
      <br>
      san2:~# drbdadm status oldNFS --verbose<br>
      drbdsetup status oldNFS <br>
      oldNFS role:Secondary<br>
        disk:UpToDate<br>
        san3 connection:Connecting<br>
      <br>
      Can anyone help advise where I should look, or what I might need
      to do to get this working?<br>
      <br>
      Thanks,<br>
      Adam<br>
      <br>
    </tt>
  </body>
</html>