[DRBD-user] switch was down, all drbd machines rebootet

Heiko rupertt at gmail.com
Tue Jul 7 13:18:47 CEST 2009

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


On Tue, Jul 7, 2009 at 12:14 PM, Heiko <rupertt at gmail.com> wrote:

> Hello,
>
> I want to switch from one line to 2 lines. I have a dedicated line without
> any switches now.
> How do I have to change the heartbeat and drbd config to use both lines?
> Can I just add the new IPs to the ha.cf? Hostname will be the same
>
> #use_logd on
> logfile /var/log/ha-log
> debugfile /var/log/ha-debug
> logfacility local0
> keepalive 2
> deadtime 10
> warntime 3
> initdead 20
> udpport 694
> ucast eth0 172.17.8.201 #old
> ucast eth0 172.17.8.202 #old
> ucast eth0 172.31.0.1 #new
> ucast eth0 172.31.0.2 #new


of cource I meant "ucast eth1  172.31.0.1/2"

>
> node xen-a1
> node xen-b1
> auto_failback on
>
>
> How do i have to change the drbd.conf, should I only use the dedicated LAN?
> What happens when I only use the dedicated LAN an restart drbd/heartbeat,
> will i
> get a splitbrain because they see each other not during that restart?
>
> resource bloon {
>   protocol C;
>   startup {
>     degr-wfc-timeout 120;    # 2 minutes.
>     degr-wfc-timeout 120; ## 2 minutes.
>   }
>   disk {
>     on-io-error   detach;
>   }
>   net {
>         after-sb-0pri disconnect;
>         after-sb-1pri disconnect;
>         after-sb-2pri disconnect;
>
> #       allow-two-primaries;
>         after-sb-0pri discard-zero-changes;
>         after-sb-1pri discard-secondary;
>         timeout 60;
> #       connect-int 10;
> #       ping-int 10;
> #       max-buffers 2048;
> #       max-epoch-size 2048;
>   }
>   syncer {
> #    group 100;
>     al-extents 257;
>         rate 20M;
>   }
>
>   on xen-a1 {
>     device    /dev/drbd0;
>     disk      /dev/VolGroup00/lv_bloon;
>     address   172.17.8.201:7788;
>
>     meta-disk internal;
>   }
>
>   on xen-b1 {
>     device    /dev/drbd0;
>     disk      /dev/VolGroup00/lv_bloon;
>     address   172.17.8.202:7788;
>     meta-disk internal;
>   }
> }
>
>
> greetings
>
>
> .r
>
> On Sat, Jul 4, 2009 at 3:58 AM, Scott Inderlied <
> scott at northwestcomputer.biz> wrote:
>
>>  ha.cf
>>
>> #use_logd on
>> logfile /var/log/ha-log
>> debugfile /var/log/ha-debug
>> logfacility local0
>> keepalive 2
>> deadtime 10
>> warntime 3
>> initdead 20
>> udpport 694
>> ucast eth0 172.20.1.1
>> ucast eth0 172.20.2.1
>>
>> # using different subnets won’t make any difference. Install another nicand use a crossover or us a serial cable. What you experienced is by design
>> to prevent splitbrain
>> node xen-A1.fra1.mailcluster
>> node xen-B1.fra1.mailcluster
>> auto_failback on
>>
>>
>>
>> *From:* drbd-user-bounces at lists.linbit.com [mailto:
>> drbd-user-bounces at lists.linbit.com] *On Behalf Of *Heiko
>> *Sent:* Friday, July 03, 2009 09:01
>> *To:* drbd-user at lists.linbit.com
>> *Subject:* [DRBD-user] switch was down, all drbd machines rebootet
>>
>>
>>
>> Hello,
>>
>>
>> i had an earlier discussion here where we came to the conclusion that
>> using Protocol C can cause crashes.
>> Yesterday we had problems with one of our switches and therefore the drbd
>> enabled machines couldnt see each other,
>> than all the machines did reboots, created splitbrains and a lot of work.
>> Do you think the crashes/reboots are caused by the same problem or can we
>> prevent this behavouir by optimizing our
>> heartbeat drbd config? Ill attach a drbd config and the ha.cf
>>
>> ---------------------------------
>> drbd.conf
>>
>> common {
>>   protocol C;
>> }
>>
>>
>>
>> resource drbd_backend {
>>   startup {
>>     degr-wfc-timeout 120;    # 2 minutes.
>>   }
>>   disk {
>>     on-io-error   detach;
>>   }
>>   net {
>>   }
>>   syncer {
>>         rate 500M;
>>         al-extents 257;
>>   }
>>
>>   on xen-B1.fra1.mailcluster {
>>     device    /dev/drbd0;
>>     disk      /dev/md3;
>>     address   172.20.2.1:7788;
>>     meta-disk internal;
>>   }
>>   on xen-A1.fra1.mailcluster {
>>     device    /dev/drbd0;
>>     disk      /dev/md3;
>>     address   172.20.1.1:7788;
>>     meta-disk internal;
>>   }
>> }
>>
>> ---------------------------------------
>> ha.cf
>>
>> #use_logd on
>> logfile /var/log/ha-log
>> debugfile /var/log/ha-debug
>> logfacility local0
>> keepalive 2
>> deadtime 10
>> warntime 3
>> initdead 20
>> udpport 694
>> ucast eth0 172.20.1.1
>> ucast eth0 172.20.2.1
>> node xen-A1.fra1.mailcluster
>> node xen-B1.fra1.mailcluster
>> auto_failback on
>>
>>
>>
>> thnx a lot
>>
>>
>> .r
>>
>> _______________________________________________
>> drbd-user mailing list
>> drbd-user at lists.linbit.com
>> http://lists.linbit.com/mailman/listinfo/drbd-user
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20090707/23c37ddf/attachment.htm>


More information about the drbd-user mailing list