[DRBD-user] switch was down, all drbd machines rebootet

Heiko rupertt at gmail.com
Tue Jul 7 13:20:25 CEST 2009

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


does it matter that the hostname are in an DNS Server?
Can I just add 2 entries to my hosts file with the new IP addresses on eth1?
I just had another crash and would like to solve this todays, if possible.


thnx

.r


On Tue, Jul 7, 2009 at 1:18 PM, Heiko <rupertt at gmail.com> wrote:

>
>
> On Tue, Jul 7, 2009 at 12:14 PM, Heiko <rupertt at gmail.com> wrote:
>
>> Hello,
>>
>> I want to switch from one line to 2 lines. I have a dedicated line without
>> any switches now.
>> How do I have to change the heartbeat and drbd config to use both lines?
>> Can I just add the new IPs to the ha.cf? Hostname will be the same
>>
>> #use_logd on
>> logfile /var/log/ha-log
>> debugfile /var/log/ha-debug
>> logfacility local0
>> keepalive 2
>> deadtime 10
>> warntime 3
>> initdead 20
>> udpport 694
>> ucast eth0 172.17.8.201 #old
>> ucast eth0 172.17.8.202 #old
>> ucast eth0 172.31.0.1 #new
>> ucast eth0 172.31.0.2 #new
>
>
> of cource I meant "ucast eth1  172.31.0.1/2"
>
>>
>> node xen-a1
>> node xen-b1
>> auto_failback on
>>
>>
>> How do i have to change the drbd.conf, should I only use the dedicated
>> LAN?
>> What happens when I only use the dedicated LAN an restart drbd/heartbeat,
>> will i
>> get a splitbrain because they see each other not during that restart?
>>
>> resource bloon {
>>   protocol C;
>>   startup {
>>     degr-wfc-timeout 120;    # 2 minutes.
>>     degr-wfc-timeout 120; ## 2 minutes.
>>   }
>>   disk {
>>     on-io-error   detach;
>>   }
>>   net {
>>         after-sb-0pri disconnect;
>>         after-sb-1pri disconnect;
>>         after-sb-2pri disconnect;
>>
>> #       allow-two-primaries;
>>         after-sb-0pri discard-zero-changes;
>>         after-sb-1pri discard-secondary;
>>         timeout 60;
>> #       connect-int 10;
>> #       ping-int 10;
>> #       max-buffers 2048;
>> #       max-epoch-size 2048;
>>   }
>>   syncer {
>> #    group 100;
>>     al-extents 257;
>>         rate 20M;
>>   }
>>
>>   on xen-a1 {
>>     device    /dev/drbd0;
>>     disk      /dev/VolGroup00/lv_bloon;
>>     address   172.17.8.201:7788;
>>
>>     meta-disk internal;
>>   }
>>
>>   on xen-b1 {
>>     device    /dev/drbd0;
>>     disk      /dev/VolGroup00/lv_bloon;
>>     address   172.17.8.202:7788;
>>     meta-disk internal;
>>   }
>> }
>>
>>
>> greetings
>>
>>
>> .r
>>
>> On Sat, Jul 4, 2009 at 3:58 AM, Scott Inderlied <
>> scott at northwestcomputer.biz> wrote:
>>
>>>  ha.cf
>>>
>>> #use_logd on
>>> logfile /var/log/ha-log
>>> debugfile /var/log/ha-debug
>>> logfacility local0
>>> keepalive 2
>>> deadtime 10
>>> warntime 3
>>> initdead 20
>>> udpport 694
>>> ucast eth0 172.20.1.1
>>> ucast eth0 172.20.2.1
>>>
>>> # using different subnets won’t make any difference. Install another nicand use a crossover or us a serial cable. What you experienced is by design
>>> to prevent splitbrain
>>> node xen-A1.fra1.mailcluster
>>> node xen-B1.fra1.mailcluster
>>> auto_failback on
>>>
>>>
>>>
>>> *From:* drbd-user-bounces at lists.linbit.com [mailto:
>>> drbd-user-bounces at lists.linbit.com] *On Behalf Of *Heiko
>>> *Sent:* Friday, July 03, 2009 09:01
>>> *To:* drbd-user at lists.linbit.com
>>> *Subject:* [DRBD-user] switch was down, all drbd machines rebootet
>>>
>>>
>>>
>>> Hello,
>>>
>>>
>>> i had an earlier discussion here where we came to the conclusion that
>>> using Protocol C can cause crashes.
>>> Yesterday we had problems with one of our switches and therefore the drbd
>>> enabled machines couldnt see each other,
>>> than all the machines did reboots, created splitbrains and a lot of work.
>>> Do you think the crashes/reboots are caused by the same problem or can we
>>> prevent this behavouir by optimizing our
>>> heartbeat drbd config? Ill attach a drbd config and the ha.cf
>>>
>>> ---------------------------------
>>> drbd.conf
>>>
>>> common {
>>>   protocol C;
>>> }
>>>
>>>
>>>
>>> resource drbd_backend {
>>>   startup {
>>>     degr-wfc-timeout 120;    # 2 minutes.
>>>   }
>>>   disk {
>>>     on-io-error   detach;
>>>   }
>>>   net {
>>>   }
>>>   syncer {
>>>         rate 500M;
>>>         al-extents 257;
>>>   }
>>>
>>>   on xen-B1.fra1.mailcluster {
>>>     device    /dev/drbd0;
>>>     disk      /dev/md3;
>>>     address   172.20.2.1:7788;
>>>     meta-disk internal;
>>>   }
>>>   on xen-A1.fra1.mailcluster {
>>>     device    /dev/drbd0;
>>>     disk      /dev/md3;
>>>     address   172.20.1.1:7788;
>>>     meta-disk internal;
>>>   }
>>> }
>>>
>>> ---------------------------------------
>>> ha.cf
>>>
>>> #use_logd on
>>> logfile /var/log/ha-log
>>> debugfile /var/log/ha-debug
>>> logfacility local0
>>> keepalive 2
>>> deadtime 10
>>> warntime 3
>>> initdead 20
>>> udpport 694
>>> ucast eth0 172.20.1.1
>>> ucast eth0 172.20.2.1
>>> node xen-A1.fra1.mailcluster
>>> node xen-B1.fra1.mailcluster
>>> auto_failback on
>>>
>>>
>>>
>>> thnx a lot
>>>
>>>
>>> .r
>>>
>>> _______________________________________________
>>> drbd-user mailing list
>>> drbd-user at lists.linbit.com
>>> http://lists.linbit.com/mailman/listinfo/drbd-user
>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20090707/f2503e10/attachment.htm>


More information about the drbd-user mailing list