Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hello, I want to switch from one line to 2 lines. I have a dedicated line without any switches now. How do I have to change the heartbeat and drbd config to use both lines? Can I just add the new IPs to the ha.cf? Hostname will be the same #use_logd on logfile /var/log/ha-log debugfile /var/log/ha-debug logfacility local0 keepalive 2 deadtime 10 warntime 3 initdead 20 udpport 694 ucast eth0 172.17.8.201 #old ucast eth0 172.17.8.202 #old ucast eth0 172.31.0.1 #new ucast eth0 172.31.0.2 #new node xen-a1 node xen-b1 auto_failback on How do i have to change the drbd.conf, should I only use the dedicated LAN? What happens when I only use the dedicated LAN an restart drbd/heartbeat, will i get a splitbrain because they see each other not during that restart? resource bloon { protocol C; startup { degr-wfc-timeout 120; # 2 minutes. degr-wfc-timeout 120; ## 2 minutes. } disk { on-io-error detach; } net { after-sb-0pri disconnect; after-sb-1pri disconnect; after-sb-2pri disconnect; # allow-two-primaries; after-sb-0pri discard-zero-changes; after-sb-1pri discard-secondary; timeout 60; # connect-int 10; # ping-int 10; # max-buffers 2048; # max-epoch-size 2048; } syncer { # group 100; al-extents 257; rate 20M; } on xen-a1 { device /dev/drbd0; disk /dev/VolGroup00/lv_bloon; address 172.17.8.201:7788; meta-disk internal; } on xen-b1 { device /dev/drbd0; disk /dev/VolGroup00/lv_bloon; address 172.17.8.202:7788; meta-disk internal; } } greetings .r On Sat, Jul 4, 2009 at 3:58 AM, Scott Inderlied <scott at northwestcomputer.biz > wrote: > ha.cf > > #use_logd on > logfile /var/log/ha-log > debugfile /var/log/ha-debug > logfacility local0 > keepalive 2 > deadtime 10 > warntime 3 > initdead 20 > udpport 694 > ucast eth0 172.20.1.1 > ucast eth0 172.20.2.1 > > # using different subnets won’t make any difference. Install another nicand use a crossover or us a serial cable. What you experienced is by design > to prevent splitbrain > node xen-A1.fra1.mailcluster > node xen-B1.fra1.mailcluster > auto_failback on > > > > *From:* drbd-user-bounces at lists.linbit.com [mailto: > drbd-user-bounces at lists.linbit.com] *On Behalf Of *Heiko > *Sent:* Friday, July 03, 2009 09:01 > *To:* drbd-user at lists.linbit.com > *Subject:* [DRBD-user] switch was down, all drbd machines rebootet > > > > Hello, > > > i had an earlier discussion here where we came to the conclusion that using > Protocol C can cause crashes. > Yesterday we had problems with one of our switches and therefore the drbd > enabled machines couldnt see each other, > than all the machines did reboots, created splitbrains and a lot of work. > Do you think the crashes/reboots are caused by the same problem or can we > prevent this behavouir by optimizing our > heartbeat drbd config? Ill attach a drbd config and the ha.cf > > --------------------------------- > drbd.conf > > common { > protocol C; > } > > > > resource drbd_backend { > startup { > degr-wfc-timeout 120; # 2 minutes. > } > disk { > on-io-error detach; > } > net { > } > syncer { > rate 500M; > al-extents 257; > } > > on xen-B1.fra1.mailcluster { > device /dev/drbd0; > disk /dev/md3; > address 172.20.2.1:7788; > meta-disk internal; > } > on xen-A1.fra1.mailcluster { > device /dev/drbd0; > disk /dev/md3; > address 172.20.1.1:7788; > meta-disk internal; > } > } > > --------------------------------------- > ha.cf > > #use_logd on > logfile /var/log/ha-log > debugfile /var/log/ha-debug > logfacility local0 > keepalive 2 > deadtime 10 > warntime 3 > initdead 20 > udpport 694 > ucast eth0 172.20.1.1 > ucast eth0 172.20.2.1 > node xen-A1.fra1.mailcluster > node xen-B1.fra1.mailcluster > auto_failback on > > > > thnx a lot > > > .r > > _______________________________________________ > drbd-user mailing list > drbd-user at lists.linbit.com > http://lists.linbit.com/mailman/listinfo/drbd-user > > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20090707/0486f7aa/attachment.htm>