Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
On Sat, Feb 27, 2016 at 11:18 AM, Eric Robinson <eric.robinson at psmnv.com>
wrote:
> Sadly, it still isn’t working.
>
> Here is my crm config...
>
> node ha13a
> node ha13b
> primitive p_drbd0 ocf:linbit:drbd \
> params drbd_resource=ha01_mysql \
> op monitor interval=31s role=Slave \
> op monitor interval=30s role=Master
> primitive p_drbd1 ocf:linbit:drbd \
> params drbd_resource=ha02_mysql \
> op monitor interval=29s role=Slave \
> op monitor interval=28s role=Master
> primitive p_fs_clust17 Filesystem \
> params device="/dev/vg_drbd0/lv_drbd0" directory="/ha01_mysql"
> fstype=ext3 options=noatime
> primitive p_fs_clust18 Filesystem \
> params device="/dev/vg_drbd1/lv_drbd1" directory="/ha02_mysql"
> fstype=ext3 options=noatime
> primitive p_lvm_drbd0 LVM \
> params volgrpname=vg_drbd0
> primitive p_lvm_drbd1 LVM \
> params volgrpname=vg_drbd1
> primitive p_vip_clust17 IPaddr2 \
> params ip=192.168.9.104 cidr_netmask=32 \
> op monitor interval=30s
> primitive p_vip_clust18 IPaddr2 \
> params ip=192.168.9.105 cidr_netmask=32 \
> op monitor interval=30s
> ms ms_drbd0 p_drbd0 \
> meta master-max=1 master-node-max=1 clone-max=2 clone-node-max=1
> notify=true target-role=Master
> ms ms_drbd1 p_drbd1 \
> meta master-max=1 master-node-max=1 clone-max=2 clone-node-max=1
> notify=true target-role=Master
> colocation c_clust17 inf: p_vip_clust17 p_lvm_drbd0 ms_drbd0:Master
> colocation c_clust18 inf: p_vip_clust18 p_lvm_drbd1 ms_drbd1:Master
> order o_clust17 inf: ms_drbd0:promote p_lvm_drbd0:start p_vip_clust17
> order o_clust18 inf: ms_drbd1:promote p_lvm_drbd1:start p_vip_clust18
> property cib-bootstrap-options: \
> dc-version=1.1.11-97629de \
> cluster-infrastructure="classic openais (with plugin)" \
> no-quorum-policy=ignore \
> stonith-enabled=false \
> maintenance-mode=false \
> expected-quorum-votes=2 \
> last-lrm-refresh=1456529727
> # vim: set filetype=pcmk:
>
> Here is what my filter looks like...
>
> filter = [ "a|/dev/sda*|", "a|/dev/drbd*|", "r|.*|" ]
> write_cache_state = 0
> volume_list = [ "vg00", "vg_drbd0", "vg_drbd1" ]
>
>
> Here is what lvdisplay shows on node ha13a...
>
> [root at ha13a ~]# lvdisplay
> --- Logical volume ---
> LV Path /dev/vg00/lv00
> LV Name lv00
> VG Name vg00
> LV UUID BfYyBv-VPNI-2f5s-0kVZ-AoSr-dGcY-gojAzs
> LV Write Access read/write
> LV Creation host, time ha13a.mycharts.md, 2014-01-23 03:38:38 -0800
> LV Status available
> # open 1
> LV Size 78.12 GiB
> Current LE 20000
> Segments 1
> Allocation inherit
> Read ahead sectors auto
> - currently set to 256
> Block device 253:0
>
> --- Logical volume ---
> LV Path /dev/vg_drbd1/lv_drbd1
> LV Name lv_drbd1
> VG Name vg_drbd1
> LV UUID HLVYSz-mZbQ-rCUm-OMBg-a1G9-vqdg-FwRp5S
> LV Write Access read/write
> LV Creation host, time ha13b, 2016-02-26 13:48:51 -0800
> LV Status NOT available
> LV Size 1.00 TiB
> Current LE 262144
> Segments 1
> Allocation inherit
> Read ahead sectors auto
>
> --- Logical volume ---
> LV Path /dev/vg_drbd0/lv_drbd0
> LV Name lv_drbd0
> VG Name vg_drbd0
> LV UUID 2q0e0v-P2g1-inu4-GKDN-cTyn-e2L7-jCJ1BY
> LV Write Access read/write
> LV Creation host, time ha13a, 2016-02-26 13:48:06 -0800
> LV Status available
> # open 1
> LV Size 1.00 TiB
> Current LE 262144
> Segments 1
> Allocation inherit
> Read ahead sectors auto
> - currently set to 256
> Block device 253:1
>
> And here is what it shows on ha13b...
>
> [root at ha13b ~]# lvdisplay
> --- Logical volume ---
> LV Path /dev/vg_drbd1/lv_drbd1
> LV Name lv_drbd1
> VG Name vg_drbd1
> LV UUID HLVYSz-mZbQ-rCUm-OMBg-a1G9-vqdg-FwRp5S
> LV Write Access read/write
> LV Creation host, time ha13b, 2016-02-26 13:48:51 -0800
> LV Status available
> # open 1
> LV Size 1.00 TiB
> Current LE 262144
> Segments 1
> Allocation inherit
> Read ahead sectors auto
> - currently set to 256
> Block device 253:1
>
> --- Logical volume ---
> LV Path /dev/vg_drbd0/lv_drbd0
> LV Name lv_drbd0
> VG Name vg_drbd0
> LV UUID 2q0e0v-P2g1-inu4-GKDN-cTyn-e2L7-jCJ1BY
> LV Write Access read/write
> LV Creation host, time ha13a, 2016-02-26 13:48:06 -0800
> LV Status NOT available
> LV Size 1.00 TiB
> Current LE 262144
> Segments 1
> Allocation inherit
> Read ahead sectors auto
>
> --- Logical volume ---
> LV Path /dev/vg00/lv00
> LV Name lv00
> VG Name vg00
> LV UUID lIJWiz-2Y9j-cq2G-Ie4f-9wVK-xJbu-2s1f23
> LV Write Access read/write
> LV Creation host, time ha13b.mycharts.md, 2014-01-23 10:01:36 -0800
> LV Status available
> # open 1
> LV Size 78.12 GiB
> Current LE 20000
> Segments 1
> Allocation inherit
> Read ahead sectors auto
> - currently set to 256
> Block device 253:0
>
> And here is crm_mon...
>
> Last updated: Fri Feb 26 16:16:30 2016
> Last change: Fri Feb 26 16:12:39 2016
> Stack: classic openais (with plugin)
> Current DC: ha13a - partition with quorum
> Version: 1.1.11-97629de
> 2 Nodes configured, 2 expected votes
> 10 Resources configured
>
>
> Online: [ ha13a ha13b ]
>
> Master/Slave Set: ms_drbd0 [p_drbd0]
> Masters: [ ha13a ]
> Slaves: [ ha13b ]
> Master/Slave Set: ms_drbd1 [p_drbd1]
> Masters: [ ha13a ]
> Slaves: [ ha13b ]
> p_vip_clust17 (ocf::heartbeat:IPaddr2): Started ha13a
> p_fs_clust17 (ocf::heartbeat:Filesystem): Started ha13a
> *p_fs_clust18 (ocf::heartbeat:Filesystem): Started ha13b*
>
^^^^^^^
how could
this happen
without
starting the LV first???
p_lvm_drbd0 (ocf::heartbeat:LVM): Started ha13a
> *p_lvm_drbd1 (ocf::heartbeat:LVM): FAILED ha13b (unmanaged)*
>
^^^^^^^^
failed, so no
way p_fs_clust18
should be
allowed to start
>
> Failed actions:
> p_lvm_drbd1_stop_0 on ha13b 'unknown error' (1): call=104,
> status=complete, last-rc-change='Fri Feb 26 16:12:29 2016', queued=0ms,
> exec=10447ms
> p_lvm_drbd1_stop_0 on ha13b 'unknown error' (1): call=104,
> status=complete, last-rc-change='Fri Feb 26 16:12:29 2016', queued=0ms,
> exec=10447ms
> p_fs_clust17_start_0 on ha13b 'not installed' (5): call=110,
> status=complete, last-rc-change='Fri Feb 26 16:12:40 2016', queued=0ms,
> exec=46ms
>
>
Can you please try following constraints instead the ones you have:
group g_drbd0 p_lvm_drbd0 p_fs_clust17 p_vip_clust17
group g_drbd1 p_lvm_drbd1 p_fs_clust18 p_vip_clust18
colocation c_clust17 inf: g_drbd0 ms_drbd0:Master
colocation c_clust18 inf: g_drbd1 ms_drbd1:Master
order o_clust17 inf: ms_drbd0:promote g_drbd0:start
order o_clust18 inf: ms_drbd1:promote g_drbd1:start
> I've tried cleaning up all the resources, but this is what I get.
> Sometimes if I mess around enough, I can get everything up, but as soon as
> I tried to fail over one of the cluster IPs or filesystems, the whole thing
> goes to crap.
>
> Do you see any potential causes?
>
> --Eric
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20160227/3a08d502/attachment.htm>