[DRBD-user] re: after drbd is up

Chris Harms chris at cmiware.com
Tue Jul 3 02:35:09 CEST 2007

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.

An embedded and charset-unspecified text was scrubbed...
Name: warning1.txt
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20070702/bc7e7ce9/attachment.txt>
-------------- next part --------------
I am running DRBD on RH cluster suite 5 presently.  So far it behaves 
fairly well for me, although I have run across a vexing issue which I 
will post in another message.  I set up individual resources for my DRBD 
patitions and use a script to manage the start, stop and status 
operations.  I cobbled one together from the ones distributed with 
Heartbeat.  The one for version 2 is much more complex, and I doubt RH 
Cluster can even make use of most of it.

I have linux bring up the kernel module, and start all resources 
secondary and have the cluster manager make them primary (start) and 
secondary (stop) as necessary.

I'm attaching my primary script in case anyone finds them useful.  
Disclaimer: I am not an experienced shell programmer.  No warranties 
etc.  Would love improvements if anyone has suggestions.  For each 
resource, I have an additional script called drbd-res.sh which includes 
the above, sets the resource name and provides start, stop and status 
operations as such:

. $(dirname $0)/drbd.sh


case "$CMD" in
        # exec, so the exit code of drbdadm propagates
        echo "Usage: drbddisk [resource] {start|stop|status}"
        exit 1

exit 0


 > Hey All,

 > Just wondering if someone could shed some light on my situation? I
 > currently have one cluster with 3 servers in it and I want to use drbd
 > between 2 of them. But instead of using heart beat I want to use the
 > failover option in cluster suite 5. This is what I have so far. My
 > question is once I have drbd up what do I do next? My next question is
 > will this work or do I have to have HeartBeat?

 > 1. Cluster is up and operating fine.
 > 2. rgmanager is up and monitoring the services
 > 3. clvmd is running
 > 4. gnbd_export is running on the 2 nodes using drbd and gnbd_import is
 > running on the other and is point to the vip of the 2 drbd nodes.
 > 5. I have a 10g test partition created on both of the drbd hosts.
 > 6. drbd is running on 2 of the nodes, meaning I have successfully issued
 > the drbdadm up all command with no errors.

More information about the drbd-user mailing list