[DRBD-user] Problem with often Split-brain

Rafael Emerick rafael.rezo at gmail.com
Sat Apr 11 20:56:55 CEST 2009

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Hi,

Im mounting a shared storad with "raid5 in hardware" -> lvm -> drbd -> ocfs2
with two identical servers.
Everthing its gonny rigths, but now, in test time, drbd is often in split
brain condicion.

The two server is connect by a cross cable with a gigabit ethernet
connection.

My messages log file:
Apr 11 14:05:06 node2 kernel: drbd10: Split-Brain detected, dropping
connection!
Apr 11 14:06:14 node2 kernel: drbd10: Split-Brain detected, dropping
connection!
Apr 11 15:13:41 node2 kernel: drbd10: Split-Brain detected, dropping
connection!
Apr 11 15:35:18 node2 kernel: drbd10: Split-Brain detected, 0 primaries,
automatically solved. Sync from peer node

My drbd.conf file:

global {
    usage-count yes;
}

common {

  syncer {
    rate 80M;
    #al-extents 257;
    al-extents 2401;
    verify-alg sha1;
  }

  protocol C;

 handlers {
    split-brain "/usr/lib/drbd/notify-split-brain.sh rafael.rezo at gmail.com";
    pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f";
    pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f";
    local-io-error "echo o > /proc/sysrq-trigger ; halt -f";
    outdate-peer "/usr/sbin/drbd-peer-outdater";
  }

  disk {
    on-io-error   detach;
  }

  net {
    allow-two-primaries;
    after-sb-0pri discard-younger-primary;
    after-sb-1pri consensus;
    after-sb-2pri disconnect;
    rr-conflict disconnect;
  }


}

resource r0 {

  on node1 {

    device     /dev/drbd0;
    disk       /dev/vg0/lvm1;
    address    192.168.1.1:7788;
    meta-disk  internal;

  }

  on node2 {

    device    /dev/drbd0;
    disk      /dev/vg0/lvm1;
    address   192.168.1.2:7788;
    meta-disk internal;
  }
}

resource r1 {

on node1 {
    device     /dev/drbd1;
    disk       /dev/vg0/lvm2;
    address    192.168.1.1:7789;
    meta-disk  internal;
  }

on node2 {
    device    /dev/drbd1;
    disk      /dev/vg0/lvm2;
    address   192.168.1.2:7789;
    meta-disk internal;
  }

}

resource r2 {

  on node1 {

    device     /dev/drbd2;
    disk       /dev/vg0/vg_ocfsteste;
    address    192.168.1.1:7790;
    meta-disk  internal;

  }

 on node2 {
    device    /dev/drbd2;
    disk      /dev/vg0/vg_ocfsteste;
    address   192.168.1.2:7790;
    meta-disk internal;
  }

}

resource r10{

 on node1 {

   device     /dev/drbd10;
   disk       /dev/vg0/lvm3;
   address    192.168.1.1:7791;
   meta-disk  internal;
 }

on node2 {
   device    /dev/drbd10;
   disk      /dev/vg0/lvm3;
   address   192.168.1.2:7791;
   meta-disk internal;
 }

}

Somebody can help me?

Thank very much


-- 
Rafael Emerick Z. de Oliveira
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20090411/f9286ce3/attachment.htm>


More information about the drbd-user mailing list