[DRBD-user] Having Trouble with LVM on DRBD

Eric Robinson eric.robinson at psmnv.com
Sat Feb 27 01:18:31 CET 2016

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Sadly, it still isn’t working.

Here is my crm config...

node ha13a
node ha13b
primitive p_drbd0 ocf:linbit:drbd \
        params drbd_resource=ha01_mysql \
        op monitor interval=31s role=Slave \
        op monitor interval=30s role=Master
primitive p_drbd1 ocf:linbit:drbd \
        params drbd_resource=ha02_mysql \
        op monitor interval=29s role=Slave \
        op monitor interval=28s role=Master
primitive p_fs_clust17 Filesystem \
        params device="/dev/vg_drbd0/lv_drbd0" directory="/ha01_mysql" fstype=ext3 options=noatime
primitive p_fs_clust18 Filesystem \
        params device="/dev/vg_drbd1/lv_drbd1" directory="/ha02_mysql" fstype=ext3 options=noatime
primitive p_lvm_drbd0 LVM \
        params volgrpname=vg_drbd0
primitive p_lvm_drbd1 LVM \
        params volgrpname=vg_drbd1
primitive p_vip_clust17 IPaddr2 \
        params ip=192.168.9.104 cidr_netmask=32 \
        op monitor interval=30s
primitive p_vip_clust18 IPaddr2 \
        params ip=192.168.9.105 cidr_netmask=32 \
        op monitor interval=30s
ms ms_drbd0 p_drbd0 \
        meta master-max=1 master-node-max=1 clone-max=2 clone-node-max=1 notify=true target-role=Master
ms ms_drbd1 p_drbd1 \
        meta master-max=1 master-node-max=1 clone-max=2 clone-node-max=1 notify=true target-role=Master
colocation c_clust17 inf: p_vip_clust17 p_lvm_drbd0 ms_drbd0:Master
colocation c_clust18 inf: p_vip_clust18 p_lvm_drbd1 ms_drbd1:Master
order o_clust17 inf: ms_drbd0:promote p_lvm_drbd0:start p_vip_clust17
order o_clust18 inf: ms_drbd1:promote p_lvm_drbd1:start p_vip_clust18
property cib-bootstrap-options: \
        dc-version=1.1.11-97629de \
        cluster-infrastructure="classic openais (with plugin)" \
        no-quorum-policy=ignore \
        stonith-enabled=false \
        maintenance-mode=false \
        expected-quorum-votes=2 \
        last-lrm-refresh=1456529727
# vim: set filetype=pcmk:

Here is what my filter looks like...

    filter = [ "a|/dev/sda*|", "a|/dev/drbd*|", "r|.*|" ]
    write_cache_state = 0
    volume_list = [ "vg00", "vg_drbd0", "vg_drbd1" ]


Here is what lvdisplay shows on node ha13a...

 [root at ha13a ~]# lvdisplay
  --- Logical volume ---
  LV Path                /dev/vg00/lv00
  LV Name                lv00
  VG Name                vg00
  LV UUID                BfYyBv-VPNI-2f5s-0kVZ-AoSr-dGcY-gojAzs
  LV Write Access        read/write
  LV Creation host, time ha13a.mycharts.md, 2014-01-23 03:38:38 -0800
  LV Status              available
  # open                 1
  LV Size                78.12 GiB
  Current LE             20000
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     256
  Block device           253:0

  --- Logical volume ---
  LV Path                /dev/vg_drbd1/lv_drbd1
  LV Name                lv_drbd1
  VG Name                vg_drbd1
  LV UUID                HLVYSz-mZbQ-rCUm-OMBg-a1G9-vqdg-FwRp5S
  LV Write Access        read/write
  LV Creation host, time ha13b, 2016-02-26 13:48:51 -0800
  LV Status              NOT available
  LV Size                1.00 TiB
  Current LE             262144
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto

  --- Logical volume ---
  LV Path                /dev/vg_drbd0/lv_drbd0
  LV Name                lv_drbd0
  VG Name                vg_drbd0
  LV UUID                2q0e0v-P2g1-inu4-GKDN-cTyn-e2L7-jCJ1BY
  LV Write Access        read/write
  LV Creation host, time ha13a, 2016-02-26 13:48:06 -0800
  LV Status              available
  # open                 1
  LV Size                1.00 TiB
  Current LE             262144
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     256
  Block device           253:1

And here is what it shows on ha13b...

[root at ha13b ~]# lvdisplay
  --- Logical volume ---
  LV Path                /dev/vg_drbd1/lv_drbd1
  LV Name                lv_drbd1
  VG Name                vg_drbd1
  LV UUID                HLVYSz-mZbQ-rCUm-OMBg-a1G9-vqdg-FwRp5S
  LV Write Access        read/write
  LV Creation host, time ha13b, 2016-02-26 13:48:51 -0800
  LV Status              available
  # open                 1
  LV Size                1.00 TiB
  Current LE             262144
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     256
  Block device           253:1

  --- Logical volume ---
  LV Path                /dev/vg_drbd0/lv_drbd0
  LV Name                lv_drbd0
  VG Name                vg_drbd0
  LV UUID                2q0e0v-P2g1-inu4-GKDN-cTyn-e2L7-jCJ1BY
  LV Write Access        read/write
  LV Creation host, time ha13a, 2016-02-26 13:48:06 -0800
  LV Status              NOT available
  LV Size                1.00 TiB
  Current LE             262144
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto

  --- Logical volume ---
  LV Path                /dev/vg00/lv00
  LV Name                lv00
  VG Name                vg00
  LV UUID                lIJWiz-2Y9j-cq2G-Ie4f-9wVK-xJbu-2s1f23
  LV Write Access        read/write
  LV Creation host, time ha13b.mycharts.md, 2014-01-23 10:01:36 -0800
  LV Status              available
  # open                 1
  LV Size                78.12 GiB
  Current LE             20000
  Segments               1
  Allocation             inherit
  Read ahead sectors     auto
  - currently set to     256
  Block device           253:0

And here is crm_mon...

Last updated: Fri Feb 26 16:16:30 2016
Last change: Fri Feb 26 16:12:39 2016
Stack: classic openais (with plugin)
Current DC: ha13a - partition with quorum
Version: 1.1.11-97629de
2 Nodes configured, 2 expected votes
10 Resources configured


Online: [ ha13a ha13b ]

 Master/Slave Set: ms_drbd0 [p_drbd0]
     Masters: [ ha13a ]
     Slaves: [ ha13b ]
 Master/Slave Set: ms_drbd1 [p_drbd1]
     Masters: [ ha13a ]
     Slaves: [ ha13b ]
p_vip_clust17   (ocf::heartbeat:IPaddr2):       Started ha13a
p_fs_clust17    (ocf::heartbeat:Filesystem):    Started ha13a
p_fs_clust18    (ocf::heartbeat:Filesystem):    Started ha13b
p_lvm_drbd0     (ocf::heartbeat:LVM):   Started ha13a
p_lvm_drbd1     (ocf::heartbeat:LVM):   FAILED ha13b (unmanaged)

Failed actions:
    p_lvm_drbd1_stop_0 on ha13b 'unknown error' (1): call=104, status=complete, last-rc-change='Fri Feb 26 16:12:29 2016', queued=0ms, exec=10447ms
    p_lvm_drbd1_stop_0 on ha13b 'unknown error' (1): call=104, status=complete, last-rc-change='Fri Feb 26 16:12:29 2016', queued=0ms, exec=10447ms
    p_fs_clust17_start_0 on ha13b 'not installed' (5): call=110, status=complete, last-rc-change='Fri Feb 26 16:12:40 2016', queued=0ms, exec=46ms

I've tried cleaning up all the resources, but this is what I get. Sometimes if I mess around enough, I can get everything up, but as soon as I tried to fail over one of the cluster IPs or filesystems, the whole thing goes to crap.

Do you see any potential causes?

--Eric




More information about the drbd-user mailing list