Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hello,
I want to switch from one line to 2 lines. I have a dedicated line without
any switches now.
How do I have to change the heartbeat and drbd config to use both lines?
Can I just add the new IPs to the ha.cf? Hostname will be the same
#use_logd on
logfile /var/log/ha-log
debugfile /var/log/ha-debug
logfacility local0
keepalive 2
deadtime 10
warntime 3
initdead 20
udpport 694
ucast eth0 172.17.8.201 #old
ucast eth0 172.17.8.202 #old
ucast eth0 172.31.0.1 #new
ucast eth0 172.31.0.2 #new
node xen-a1
node xen-b1
auto_failback on
How do i have to change the drbd.conf, should I only use the dedicated LAN?
What happens when I only use the dedicated LAN an restart drbd/heartbeat,
will i
get a splitbrain because they see each other not during that restart?
resource bloon {
protocol C;
startup {
degr-wfc-timeout 120; # 2 minutes.
degr-wfc-timeout 120; ## 2 minutes.
}
disk {
on-io-error detach;
}
net {
after-sb-0pri disconnect;
after-sb-1pri disconnect;
after-sb-2pri disconnect;
# allow-two-primaries;
after-sb-0pri discard-zero-changes;
after-sb-1pri discard-secondary;
timeout 60;
# connect-int 10;
# ping-int 10;
# max-buffers 2048;
# max-epoch-size 2048;
}
syncer {
# group 100;
al-extents 257;
rate 20M;
}
on xen-a1 {
device /dev/drbd0;
disk /dev/VolGroup00/lv_bloon;
address 172.17.8.201:7788;
meta-disk internal;
}
on xen-b1 {
device /dev/drbd0;
disk /dev/VolGroup00/lv_bloon;
address 172.17.8.202:7788;
meta-disk internal;
}
}
greetings
.r
On Sat, Jul 4, 2009 at 3:58 AM, Scott Inderlied <scott at northwestcomputer.biz
> wrote:
> ha.cf
>
> #use_logd on
> logfile /var/log/ha-log
> debugfile /var/log/ha-debug
> logfacility local0
> keepalive 2
> deadtime 10
> warntime 3
> initdead 20
> udpport 694
> ucast eth0 172.20.1.1
> ucast eth0 172.20.2.1
>
> # using different subnets won’t make any difference. Install another nicand use a crossover or us a serial cable. What you experienced is by design
> to prevent splitbrain
> node xen-A1.fra1.mailcluster
> node xen-B1.fra1.mailcluster
> auto_failback on
>
>
>
> *From:* drbd-user-bounces at lists.linbit.com [mailto:
> drbd-user-bounces at lists.linbit.com] *On Behalf Of *Heiko
> *Sent:* Friday, July 03, 2009 09:01
> *To:* drbd-user at lists.linbit.com
> *Subject:* [DRBD-user] switch was down, all drbd machines rebootet
>
>
>
> Hello,
>
>
> i had an earlier discussion here where we came to the conclusion that using
> Protocol C can cause crashes.
> Yesterday we had problems with one of our switches and therefore the drbd
> enabled machines couldnt see each other,
> than all the machines did reboots, created splitbrains and a lot of work.
> Do you think the crashes/reboots are caused by the same problem or can we
> prevent this behavouir by optimizing our
> heartbeat drbd config? Ill attach a drbd config and the ha.cf
>
> ---------------------------------
> drbd.conf
>
> common {
> protocol C;
> }
>
>
>
> resource drbd_backend {
> startup {
> degr-wfc-timeout 120; # 2 minutes.
> }
> disk {
> on-io-error detach;
> }
> net {
> }
> syncer {
> rate 500M;
> al-extents 257;
> }
>
> on xen-B1.fra1.mailcluster {
> device /dev/drbd0;
> disk /dev/md3;
> address 172.20.2.1:7788;
> meta-disk internal;
> }
> on xen-A1.fra1.mailcluster {
> device /dev/drbd0;
> disk /dev/md3;
> address 172.20.1.1:7788;
> meta-disk internal;
> }
> }
>
> ---------------------------------------
> ha.cf
>
> #use_logd on
> logfile /var/log/ha-log
> debugfile /var/log/ha-debug
> logfacility local0
> keepalive 2
> deadtime 10
> warntime 3
> initdead 20
> udpport 694
> ucast eth0 172.20.1.1
> ucast eth0 172.20.2.1
> node xen-A1.fra1.mailcluster
> node xen-B1.fra1.mailcluster
> auto_failback on
>
>
>
> thnx a lot
>
>
> .r
>
> _______________________________________________
> drbd-user mailing list
> drbd-user at lists.linbit.com
> http://lists.linbit.com/mailman/listinfo/drbd-user
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20090707/0486f7aa/attachment.htm>