Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
At 2016-09-27 15:47:37, "Nick Wang" <nwang at suse.com> wrote:
>>>> On 2016-9-26 at 19:17, in message
><CACp6BS7W6PyW=453WkrRFGSZ+f0mqHqL2m9FjSXFicOCQ+wiwA at mail.gmail.com>, Igor
>Cicimov <igorc at encompasscorporation.com> wrote:
>> On 26 Sep 2016 7:26 pm, "mzlld1988" <mzlld1988 at 163.com> wrote:
>> >
>> > I apply the attached patch file to scripts/drbd.ocf,then pacemaker can
>> start drbd successfully,but only two nodes, the third node's drbd is
>> down,is it right?
>> Well you didnt say you have 3 nodes. Usually you use pacemaker with 2 nodes
>> and drbd.
>The patch suppose to help on 3(more) nodes scenario, as long as only one Primary.
>Is the 3 nodes DRBD cluster working without pacemaker? And how did you
>configure in pacemaker?
Accoring to http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf, I executed the following commands to configure drbd in pacemaker.
[root at pcmk-1 ~]# pcs cluster cib drbd_cfg
[root at pcmk-1 ~]# pcs -f drbd_cfg resource create WebData ocf:linbit:drbd \
drbd_resource=wwwdata op monitor interval=60s
[root at pcmk-1 ~]# pcs -f drbd_cfg resource master WebDataClone WebData \
master-max=1 master-node-max=1 clone-max=2 clone-node-max=1 \
notify=true
[root at pcmk-1 ~]# pcs cluster cib-push drbd_cfg
>> > And another question is ,can pacemaker successfully stop the slave node
>> ? My result is pacemaker can't sop the slave node.
>> >
>Yes, need to check the log on which resource prevent pacemaker to stop.
Pacemaker can't stop slave node's drbd, I think the reason may be the same as my previous email(see attached file) ,but no one reply that email.
[root at drbd ~]# pcs status
Cluster name: mycluster
Stack: corosync
Current DC: drbd.node103 (version 1.1.15-e174ec8) - partition with quorum
Last updated: Mon Sep 26 04:36:50 2016 Last change: Mon Sep 26 04:36:49 2016 by root via cibadmin on drbd.node101
3 nodes and 2 resources configured
Online: [ drbd.node101 drbd.node102 drbd.node103 ]
Full list of resources:
Master/Slave Set: WebDataClone [WebData]
Masters: [ drbd.node102 ]
Slaves: [ drbd.node101 ]
Daemon Status:
corosync: active/corosync.service is not a native service, redirecting to /sbin/chkconfig.
Executing /sbin/chkconfig corosync --level=5
enabled
pacemaker: active/pacemaker.service is not a native service, redirecting to /sbin/chkconfig.
Executing /sbin/chkconfig pacemaker --level=5
enabled
pcsd: active/enabled
-------------------------------------------------------------
Faile to execute ‘pcs cluster stop drbd.node101’
=Error message on drbd.node101(secondary node)
Sep 26 04:39:26 drbd lrmd[3521]: notice: WebData_stop_0:4726:stderr [ Command 'drbdsetup down r0' terminated with exit code 11 ]
Sep 26 04:39:26 drbd lrmd[3521]: notice: WebData_stop_0:4726:stderr [ r0: State change failed: (-10) State change was refused by peer node ]
Sep 26 04:39:26 drbd lrmd[3521]: notice: WebData_stop_0:4726:stderr [ additional info from kernel: ]
Sep 26 04:39:26 drbd lrmd[3521]: notice: WebData_stop_0:4726:stderr [ failed to disconnect ]
Sep 26 04:39:26 drbd lrmd[3521]: notice: WebData_stop_0:4726:stderr [ Command 'drbdsetup down r0' terminated with exit code 11 ]
Sep 26 04:39:26 drbd crmd[3524]: error: Result of stop operation for WebData on drbd.node101: Timed Out | call=12 key=WebData_stop_0 timeout=100000ms
Sep 26 04:39:26 drbd crmd[3524]: notice: drbd.node101-WebData_stop_0:12 [ r0: State change failed: (-10) State change was refused by peer node\nadditional info from kernel:\nfailed to disconnect\nCommand 'drbdsetup down r0' terminated with exit code 11\nr0: State change failed: (-10) State change was refused by peer node\nadditional info from kernel:\nfailed to disconnect\nCommand 'drbdsetup down r0' terminated with exit code 11\nr0: State change failed: (-10) State change was refused by peer node\nadditional info from kernel:\nfailed t
=Error message on drbd.node102(primary node)
Sep 26 04:39:25 drbd kernel: drbd r0 drbd.node101: Preparing remote state change 3578772780 (primary_nodes=4, weak_nodes=FFFFFFFFFFFFFFFB)
Sep 26 04:39:25 drbd kernel: drbd r0: State change failed: Refusing to be Primary while peer is not outdated
Sep 26 04:39:25 drbd kernel: drbd r0: Failed: susp-io( no -> fencing)
Sep 26 04:39:25 drbd kernel: drbd r0 drbd.node101: Failed: conn( Connected -> TearDown ) peer( Secondary -> Unknown )
Sep 26 04:39:25 drbd kernel: drbd r0/0 drbd1 drbd.node101: Failed: pdsk( UpToDate -> DUnknown ) repl( Established -> Off )
Sep 26 04:39:25 drbd kernel: drbd r0 drbd.node101: Aborting remote state change 3578772780
>> > I'm looking forward to your answers.Thanks.
>> >
>> Yes it works with 2 nodes drbd9 configured in standard way not via
>> drbdmanager. Haven't tried any other layout.
>>
>> >
>
>Best regards,
>Nick
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20160927/628f1619/attachment.htm>
-------------- next part --------------
An embedded message was scrubbed...
From: =?GBK?B?wfW1pA==?= <mzlld1988 at 163.com>
Subject: Is it normal that we can't directly remove the secondary node when
fencing is set?
Date: Sat, 10 Sep 2016 14:46:01 +0800 (CST)
Size: 15557
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20160927/628f1619/attachment.eml>