Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
On Sat, Oct 31, 2009 at 02:57:37AM +0100, Thomas Schneider wrote: > Hello, > > I'm trying to setup a cluster for shared storage with pacemaker, drbd, and > nfs. > I use two server and there is two network interface one each server, eth0 > connected to the network, and eth1 direct link betwene the two node for drbd > replication, > I can migrate the resource between the to node, but the problem is when i > make a hard power-off of server01 ( the server where is running the > resource) the second server (server02) doesn't make failover. (the resource > doesn't start) May you can take a look to my config file in the following > storage01:~# crm configure show > node storage01.myriapulse.local \ > attributes standby="off" > node storage02.myriapulse.local \ > attributes standby="off" > primitive drbd_nfs ocf:linbit:drbd \ > params drbd_resource="nfs" \ > op monitor interval="15s" \ > meta target-role="Started" > primitive fs_nfs ocf:heartbeat:Filesystem \ > params device="/dev/drbd/by-res/nfs" directory="/share" > fstype="ext3" \ > meta is-managed="true" > primitive ftp-server lsb:proftpd \ > op monitor interval="1min" > primitive ip_nfs ocf:heartbeat:IPaddr2 \ > params ip="10.1.1.69" nic="eth0" > primitive nfs-kernel-server lsb:nfs-kernel-server \ > op monitor interval="1min" > group nfs fs_nfs ip_nfs nfs-kernel-server \ > meta target-role="Started" > ms ms_drbd_nfs drbd_nfs \ > meta master-max="1" master-node-max="1" clone-max="2" > clone-node-max="1" notify="true" > location cli-standby-nfs nfs \ > rule $id="cli-standby-rule-nfs" -inf: #uname eq > storage02.myriapulse.local colocation ftp_on_nfs inf: ftp-server nfs > colocation nfs_on_drbd inf: nfs ms_drbd_nfs:Master order ftp_after_nfs inf: > nfs ftp-server order nfs_after_drbd inf: ms_drbd_nfs:promote nfs:start You really should get your line breaks right! > property $id="cib-bootstrap-options" \ > dc-version="1.0.5-unknown" \ > cluster-infrastructure="openais" \ > expected-quorum-votes="2" \ > no-quorum-policy="ignore" \ > stonith-enabled="false" > Oct 31 02:47:06 storage02 kernel: [419458.084631] block drbd0: helpercommand: /sbin/drbdadm fence-peer minor-0 > Oct 31 02:47:07 storage02 crm-fence-peer.sh[24594]: invoked for nfs > Oct 31 02:47:07 storage02 kernel: [419459.283186] block drbd0: helpercommand: /sbin/drbdadm fence-peer minor-0 exit code 4 (0x400) > Oct 31 02:47:07 storage02 kernel: [419459.283186] block drbd0: fence-peerhelper returned 4 (peer was fenced) > Oct 31 02:47:07 storage02 kernel: [419459.283186] block drbd0: role(Secondary -> Primary ) pdsk( DUnknown -> Outdated ) > Oct 31 02:47:07 storage02 kernel: [419459.343187] block drbd0: Creating newcurrent UUID > Oct 31 02:47:07 storage02 lrmd: [23822]: info: RA output:(drbd_nfs:0:promote:stdout) > Oct 31 02:47:08 storage02 crmd: [23825]: info: process_lrm_event: LRMoperation drbd_nfs:0_promote_0 (call=12, rc=0, cib-update=42,confirmed=true) complete ok DRBD is promoted just fine. The resources using it are not. Which is expected: As long as you explicitly forbid "nfs" to run on storage02, you should not complain about nfs not being started on storage02. Hint: > location cli-standby-nfs nfs \ > rule $id="cli-standby-rule-nfs" -inf: #uname eq storage02.myriapulse.local -- : Lars Ellenberg : LINBIT | Your Way to High Availability : DRBD/HA support and consulting http://www.linbit.com DRBD® and LINBIT® are registered trademarks of LINBIT, Austria. __ please don't Cc me, but send to list -- I'm subscribed