[DRBD-user] Can't find LV from second node on a Primary/Primary DRBD Device

David Coulson david at davidcoulson.net
Mon Oct 3 18:19:42 CEST 2011

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


You could try doing a vgexport then a vgimport. You probably need to 
blow away your lvm cache too.

There are ways to make it work without clvmd, but you can't have the VG 
imported on both hosts at the same time.

On 10/3/11 12:18 PM, markus at marcap.de wrote:
> Hi,
>
> thanks for the answer. No I don't use clustered LVM. I tried to do 
> deactivating and activating from the second node (client01) but 
> nothing changed:
>
> client01:~# vgscan
>   Reading all physical volumes.  This may take a while...
>   Found volume group "pve" using metadata type lvm2
>   Found volume group "replicated" using metadata type lvm2
> client01:~# vgchange -an
>   Can't deactivate volume group "pve" with 4 open logical volume(s)
> client01:~# vgchange -an replicated
>   0 logical volume(s) in volume group "replicated" now active
> client01:~# vgscan
>   Reading all physical volumes.  This may take a while...
>   Found volume group "pve" using metadata type lvm2
>   Found volume group "replicated" using metadata type lvm2
> client01:~# vgchange -ay replicated
>   2 logical volume(s) in volume group "replicated" now active
> client01:~# lvdisplay
> output shows just local lv's
>
> ---------------------------------------
>
> If I do lvdisplay on the first node (client00) I see the local LV's 
> and the DRBD-LV's:
>   --- Logical volume ---
>   LV Name                /dev/replicated/vm-101-disk-1
>   VG Name                replicated
>   LV UUID                i7NM1a-heRf-ZufV-5pVJ-NsU8-0xHh-d2vALH
>   LV Write Access        read/write
>   LV Status              available
>   # open                 0
>   LV Size                10.00 GB
>   Current LE             2560
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     256
>   Block device           253:4
>
>   --- Logical volume ---
>   LV Name                /dev/replicated/vm-104-disk-1
>   VG Name                replicated
>   LV UUID                eEdV1h-42uJ-GH56-o0f8-Nk2i-DaMm-G1er2X
>   LV Write Access        read/write
>   LV Status              available
>   # open                 0
>   LV Size                20.00 GB
>   Current LE             5120
>   Segments               1
>   Allocation             inherit
>   Read ahead sectors     auto
>   - currently set to     256
>   Block device           253:5
>
> I think it should work without clustered LVM?
>
> Best regards,
> Markus
>
> Zitat von David Coulson <david at davidcoulson.net>:
>
>>  Are you using clustered LVM? If not, you probably should be.
>>
>> You could try deactivating the VG then reactivating it and seeing if 
>> the new primary sees the LV.
>>
>> On 10/3/11 6:02 AM, markus at marcap.de wrote:
>>> Dear all,
>>>
>>> I want to build an HA setup with virtual machines on a two hardware 
>>> node setup. I already setup a nested LVM DRBD configuration. But 
>>> from the second node I can't see the lv's from the drbd device; 
>>> pvscan displays it.
>>>
>>> Has anybody an idea whats wrong with my setup? I think I need a 
>>> primary/primary drbd device because if one hardware node fails I 
>>> want to start the KVM container from the other one; therefore I need 
>>> the latest data of the KVM container.
>>>
>>> Please find some logs below...
>>>
>>> Thanks a lot for your help in advance!
>>>
>>> Best regards,
>>> Markus
>>>
>>> Client00:
>>> ---------
>>> client00:~# lvscan
>>>  ACTIVE            '/dev/pve/swap' [15.00 GB] inherit
>>>  ACTIVE            '/dev/pve/root' [96.00 GB] inherit
>>>  ACTIVE            '/dev/pve/data' [1.32 TB] inherit
>>>  ACTIVE            '/dev/pve/r0' [60.00 GB] inherit
>>>  ACTIVE            '/dev/replicated/vm-101-disk-1' [10.00 GB] inherit
>>> client00:~# pvscan
>>>  PV /dev/sda2    VG pve          lvm2 [1.82 TB / 343.99 GB free]
>>>  PV /dev/drbd0   VG replicated   lvm2 [60.00 GB / 50.00 GB free]
>>>  Total: 2 [1.88 TB] / in use: 2 [1.88 TB] / in no VG: 0 [0   ]
>>> client00:~# cat /proc/drbd
>>> version: 8.3.10 (api:88/proto:86-96)
>>> GIT-hash: 5c0b0469666682443d4785d90a2c603378f9017b build by 
>>> phil at fat-tyre, 2011-01-28 12:17:35
>>> 0: cs:Connected ro:Primary/Primary ds:UpToDate/UpToDate C r-----
>>>    ns:81968 nr:28 dw:28 dr:84752 al:0 bm:15 lo:0 pe:0 ua:0 ap:0 ep:1 
>>> wo:b oos:0
>>>
>>> Client01:
>>> ---------
>>> client01:~# lvscan
>>>  ACTIVE            '/dev/pve/swap' [4.00 GB] inherit
>>>  ACTIVE            '/dev/pve/root' [96.00 GB] inherit
>>>  ACTIVE            '/dev/pve/data' [194.14 GB] inherit
>>>  ACTIVE            '/dev/pve/r0' [60.00 GB] inherit
>>> client01:~# pvscan
>>>  PV /dev/sda2    VG pve          lvm2 [698.13 GB / 343.99 GB free]
>>>  PV /dev/drbd0   VG replicated   lvm2 [60.00 GB / 50.00 GB free]
>>>  Total: 2 [758.13 GB] / in use: 2 [758.13 GB] / in no VG: 0 [0   ]
>>> client01:~# cat /proc/drbd
>>> version: 8.3.10 (api:88/proto:86-96)
>>> GIT-hash: 5c0b0469666682443d4785d90a2c603378f9017b build by 
>>> phil at fat-tyre, 2011-01-28 12:17:35
>>> 0: cs:Connected ro:Primary/Primary ds:UpToDate/UpToDate C r-----
>>>    ns:28 nr:81968 dw:81948 dr:2096 al:2 bm:14 lo:0 pe:0 ua:0 ap:0 
>>> ep:1 wo:b oos:0
>>> _______________________________________________
>>> drbd-user mailing list
>>> drbd-user at lists.linbit.com
>>> http://lists.linbit.com/mailman/listinfo/drbd-user
>>
>
>



More information about the drbd-user mailing list