Hello,<br><br>i had an earlier discussion here where we came to the conclusion that using Protocol C can cause crashes.<br>Yesterday we had problems with one of our switches and therefore the drbd enabled machines couldnt see each other,<br>
than all the machines did reboots, created splitbrains and a lot of work.<br>Do you think the crashes/reboots are caused by the same problem or can we prevent this behavouir by optimizing our<br>heartbeat drbd config? Ill attach a drbd config and the <a href="http://ha.cf">ha.cf</a><br>
<br>---------------------------------<br>drbd.conf<br><br>common {<br> protocol C;<br>}<br><br><br><br>resource drbd_backend {<br> startup {<br> degr-wfc-timeout 120; # 2 minutes.<br> }<br> disk {<br> on-io-error detach;<br>
}<br> net {<br> }<br> syncer {<br> rate 500M;<br> al-extents 257;<br> }<br><br> on xen-B1.fra1.mailcluster {<br> device /dev/drbd0;<br> disk /dev/md3;<br> address <a href="http://172.20.2.1:7788">172.20.2.1:7788</a>;<br>
meta-disk internal;<br> }<br> on xen-A1.fra1.mailcluster {<br> device /dev/drbd0;<br> disk /dev/md3;<br> address <a href="http://172.20.1.1:7788">172.20.1.1:7788</a>;<br> meta-disk internal;<br>
}<br>}<br><br>---------------------------------------<br><a href="http://ha.cf">ha.cf</a><br><br>#use_logd on<br>logfile /var/log/ha-log<br>debugfile /var/log/ha-debug<br>logfacility local0<br>keepalive 2<br>deadtime 10<br>
warntime 3<br>initdead 20<br>udpport 694<br>ucast eth0 172.20.1.1<br>ucast eth0 172.20.2.1<br>node xen-A1.fra1.mailcluster<br>node xen-B1.fra1.mailcluster<br>auto_failback on<br><br><br><br>thnx a lot<br><br><br>.r<br>