Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Thank you. I'll do so.
Greetings,
Oliver
> Am Samstag, 10. April 2010 14:50:30 schrieb Oliver Hoffmann:
> > Hi list,
> >
> > probably more a heartbeat question but maybe some-one can push me in
> > the right direction.
> >
> > I want two storage servers with a bunch of drbd-devices and LVM on
> > top in order to make snapshots and resizing disk space. The idea is
> > to add a HD on each node, make it a drbd-device and extend a
> > logical volume accordingly whenever I run out of space. heartbeat
> > (or pacemaker?) should take care and do a failover to node2 if
> > node1 dies. Finally clients should connect via iSCSI or smb or nfs.
> >
> > I successfully made drbd0 and drbd1 and created a volume group with
> > logical volumes. I resized my lv after adding drbd2 as well.
> > Furthermore one drbd (without LVM) and heatbeat works without
> > problems.
> >
> > So far so good, but how does heartbeat put all three drbd to
> > primary, do a vgchange -ay and finally mount my logical volumes
> > after a reboot or when node1 dies? It was even not possible to
> > handle just the three drbd-devices by heartbeat.
> >
> > I tried different haressources:
> >
> > node1 IPaddr2::external-IP/24/eth0 drbddisk::raid
> > Filesystem::/dev/drbd0
> >
> > I know that I need my mount point and ext4 here but I don't want it
> > to be mounted because a lv should be mounted. I only want it to be
> > primary.
> >
> > Now the setup with three devices.
> >
> > node1 IPaddr2::external-IP/24/eth0
> > node1 drbddisk::raid Filesystem::/dev/drbd0
> > node1 drbddisk::raid2 Filesystem::/dev/drbd1
> > node1 drbddisk::raid3 Filesystem::/dev/drbd2
> >
> > This doesn't work at all. After a reboot all devices are in state
> > secondary/unknown. It is possible to get them back to normal
> > (primary/secondary up-to-date) by hand though. I have to connect,
> > perform "drbdsetup /dev/drbdX primary -o" and restart drbd.
> >
> > drbd and LVM together:
> >
> > node1 drbddisk::raid LVM::drbd-vg
> > Filesystem::/dev/mapper/drbd-vg-lv1::/mnt/lv1::ext4
> >
> > Works not at all.
> >
> > Thanx a lot for guidance!
> >
> >
> > Oliver
> >
> >
> > ################### The other configs ###########################
> >
> > #/etc/drbd.conf
> > resource raid {
> >
> > protocol C;
> >
> > handlers {
> > pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f";
> > pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f";
> > local-io-error "echo o > /proc/sysrq-trigger ; halt -f";
> > outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater";
> > }
> >
> > startup { wfc-timeout 0; degr-wfc-timeout 120; }
> > disk { fencing resource-only; }
> > syncer { rate 50M; al-extents 257; }
> >
> > net {
> > after-sb-0pri discard-younger-primary;
> > after-sb-1pri consensus;
> > after-sb-2pri disconnect;
> > rr-conflict call-pri-lost;
> > ping-timeout 20;
> > }
> >
> > on node1 {
> > device /dev/drbd0;
> > disk /dev/sda6;
> > address 192.168.1.1:7788;
> > meta-disk internal;
> > }
> >
> > on node2 {
> > device /dev/drbd0;
> > disk /dev/sda6;
> > address 192.168.1.2:7788;
> > meta-disk internal;
> > }
> >
> > }
> >
> > resource raid2 {
> >
> > protocol C;
> >
> > handlers {
> > pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f";
> > pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f";
> > local-io-error "echo o > /proc/sysrq-trigger ; halt -f";
> > outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater";
> > }
> >
> > startup { wfc-timeout 0; degr-wfc-timeout 120; }
> > disk { fencing resource-only; }
> > syncer { rate 50M; al-extents 257; }
> >
> > net {
> > after-sb-0pri discard-younger-primary;
> > after-sb-1pri consensus;
> > after-sb-2pri disconnect;
> > rr-conflict call-pri-lost;
> > ping-timeout 20;
> > }
> >
> > on node1 {
> > device /dev/drbd1;
> > disk /dev/sda7;
> > address 192.168.2.1:7788;
> > meta-disk internal;
> > }
> >
> > on node2 {
> > device /dev/drbd1;
> > disk /dev/sda7;
> > address 192.168.2.2:7788;
> > meta-disk internal;
> > }
> >
> > }
> >
> > resource raid3 {
> >
> > protocol C;
> >
> > handlers {
> > pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f";
> > pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f";
> > local-io-error "echo o > /proc/sysrq-trigger ; halt -f";
> > outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater";
> > }
> >
> > startup { wfc-timeout 0; degr-wfc-timeout 120; }
> > disk { fencing resource-only; }
> > syncer { rate 50M; al-extents 257; }
> >
> > net {
> > after-sb-0pri discard-younger-primary;
> > after-sb-1pri consensus;
> > after-sb-2pri disconnect;
> > rr-conflict call-pri-lost;
> > ping-timeout 20;
> > }
> >
> > on node1 {
> > device /dev/drbd2;
> > disk /dev/sda8;
> > address 192.168.3.1:7788;
> > meta-disk internal;
> > }
> >
> > on node2 {
> > device /dev/drbd2;
> > disk /dev/sda8;
> > address 192.168.3.2:7788;
> > meta-disk internal;
> > }
> >
> > }
> >
> > #/etc/network/interfaces on node2
> > auto eth1
> > iface eth1 inet static
> > address 192.168.1.2
> > netmask 255.255.255.0
> > network 192.168.1.0
> > broadcast 192.168.1.255
> >
> > auto eth1:0
> > iface eth1:0 inet static
> > name Ethernet alias LAN card
> > address 192.168.2.2
> > netmask 255.255.255.0
> > broadcast 192.168.2.255
> > network 192.168.2.0
> >
> > auto eth1:1
> > iface eth1:1 inet static
> > name Ethernet alias LAN card
> > address 192.168.3.2
> > netmask 255.255.255.0
> > broadcast 192.168.3.255
> > network 192.168.3.0
> >
> >
> >
> > _______________________________________________
> > drbd-user mailing list
> > drbd-user at lists.linbit.com
> > http://lists.linbit.com/mailman/listinfo/drbd-user
>
>
> Hi,
>
> first advise:
>
> do not use heartbeat version 1 - style configurations any more. You
> will not get any support from anybody no more.
>
> Please use a distinct cluster manger like pacemaker. See
> www.clusterlabs,org.
>
> Greetings,
>
> Michael Schwartzkopff.
>
> _______________________________________________
> drbd-user mailing list
> drbd-user at lists.linbit.com
> http://lists.linbit.com/mailman/listinfo/drbd-user
>