[DRBD-user] Can't find LV from second node on a Primary/Primary DRBD Device

markus at marcap.de markus at marcap.de
Mon Oct 3 18:18:06 CEST 2011

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Hi,

thanks for the answer. No I don't use clustered LVM. I tried to do  
deactivating and activating from the second node (client01) but  
nothing changed:

client01:~# vgscan
   Reading all physical volumes.  This may take a while...
   Found volume group "pve" using metadata type lvm2
   Found volume group "replicated" using metadata type lvm2
client01:~# vgchange -an
   Can't deactivate volume group "pve" with 4 open logical volume(s)
client01:~# vgchange -an replicated
   0 logical volume(s) in volume group "replicated" now active
client01:~# vgscan
   Reading all physical volumes.  This may take a while...
   Found volume group "pve" using metadata type lvm2
   Found volume group "replicated" using metadata type lvm2
client01:~# vgchange -ay replicated
   2 logical volume(s) in volume group "replicated" now active
client01:~# lvdisplay
output shows just local lv's

---------------------------------------

If I do lvdisplay on the first node (client00) I see the local LV's  
and the DRBD-LV's:
   --- Logical volume ---
   LV Name                /dev/replicated/vm-101-disk-1
   VG Name                replicated
   LV UUID                i7NM1a-heRf-ZufV-5pVJ-NsU8-0xHh-d2vALH
   LV Write Access        read/write
   LV Status              available
   # open                 0
   LV Size                10.00 GB
   Current LE             2560
   Segments               1
   Allocation             inherit
   Read ahead sectors     auto
   - currently set to     256
   Block device           253:4

   --- Logical volume ---
   LV Name                /dev/replicated/vm-104-disk-1
   VG Name                replicated
   LV UUID                eEdV1h-42uJ-GH56-o0f8-Nk2i-DaMm-G1er2X
   LV Write Access        read/write
   LV Status              available
   # open                 0
   LV Size                20.00 GB
   Current LE             5120
   Segments               1
   Allocation             inherit
   Read ahead sectors     auto
   - currently set to     256
   Block device           253:5

I think it should work without clustered LVM?

Best regards,
Markus

Zitat von David Coulson <david at davidcoulson.net>:

>  Are you using clustered LVM? If not, you probably should be.
>
> You could try deactivating the VG then reactivating it and seeing if  
> the new primary sees the LV.
>
> On 10/3/11 6:02 AM, markus at marcap.de wrote:
>> Dear all,
>>
>> I want to build an HA setup with virtual machines on a two hardware  
>> node setup. I already setup a nested LVM DRBD configuration. But  
>> from the second node I can't see the lv's from the drbd device;  
>> pvscan displays it.
>>
>> Has anybody an idea whats wrong with my setup? I think I need a  
>> primary/primary drbd device because if one hardware node fails I  
>> want to start the KVM container from the other one; therefore I  
>> need the latest data of the KVM container.
>>
>> Please find some logs below...
>>
>> Thanks a lot for your help in advance!
>>
>> Best regards,
>> Markus
>>
>> Client00:
>> ---------
>> client00:~# lvscan
>>  ACTIVE            '/dev/pve/swap' [15.00 GB] inherit
>>  ACTIVE            '/dev/pve/root' [96.00 GB] inherit
>>  ACTIVE            '/dev/pve/data' [1.32 TB] inherit
>>  ACTIVE            '/dev/pve/r0' [60.00 GB] inherit
>>  ACTIVE            '/dev/replicated/vm-101-disk-1' [10.00 GB] inherit
>> client00:~# pvscan
>>  PV /dev/sda2    VG pve          lvm2 [1.82 TB / 343.99 GB free]
>>  PV /dev/drbd0   VG replicated   lvm2 [60.00 GB / 50.00 GB free]
>>  Total: 2 [1.88 TB] / in use: 2 [1.88 TB] / in no VG: 0 [0   ]
>> client00:~# cat /proc/drbd
>> version: 8.3.10 (api:88/proto:86-96)
>> GIT-hash: 5c0b0469666682443d4785d90a2c603378f9017b build by  
>> phil at fat-tyre, 2011-01-28 12:17:35
>> 0: cs:Connected ro:Primary/Primary ds:UpToDate/UpToDate C r-----
>>    ns:81968 nr:28 dw:28 dr:84752 al:0 bm:15 lo:0 pe:0 ua:0 ap:0  
>> ep:1 wo:b oos:0
>>
>> Client01:
>> ---------
>> client01:~# lvscan
>>  ACTIVE            '/dev/pve/swap' [4.00 GB] inherit
>>  ACTIVE            '/dev/pve/root' [96.00 GB] inherit
>>  ACTIVE            '/dev/pve/data' [194.14 GB] inherit
>>  ACTIVE            '/dev/pve/r0' [60.00 GB] inherit
>> client01:~# pvscan
>>  PV /dev/sda2    VG pve          lvm2 [698.13 GB / 343.99 GB free]
>>  PV /dev/drbd0   VG replicated   lvm2 [60.00 GB / 50.00 GB free]
>>  Total: 2 [758.13 GB] / in use: 2 [758.13 GB] / in no VG: 0 [0   ]
>> client01:~# cat /proc/drbd
>> version: 8.3.10 (api:88/proto:86-96)
>> GIT-hash: 5c0b0469666682443d4785d90a2c603378f9017b build by  
>> phil at fat-tyre, 2011-01-28 12:17:35
>> 0: cs:Connected ro:Primary/Primary ds:UpToDate/UpToDate C r-----
>>    ns:28 nr:81968 dw:81948 dr:2096 al:2 bm:14 lo:0 pe:0 ua:0 ap:0  
>> ep:1 wo:b oos:0
>> _______________________________________________
>> drbd-user mailing list
>> drbd-user at lists.linbit.com
>> http://lists.linbit.com/mailman/listinfo/drbd-user
>





More information about the drbd-user mailing list