Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Thank you. I'll do so. Greetings, Oliver > Am Samstag, 10. April 2010 14:50:30 schrieb Oliver Hoffmann: > > Hi list, > > > > probably more a heartbeat question but maybe some-one can push me in > > the right direction. > > > > I want two storage servers with a bunch of drbd-devices and LVM on > > top in order to make snapshots and resizing disk space. The idea is > > to add a HD on each node, make it a drbd-device and extend a > > logical volume accordingly whenever I run out of space. heartbeat > > (or pacemaker?) should take care and do a failover to node2 if > > node1 dies. Finally clients should connect via iSCSI or smb or nfs. > > > > I successfully made drbd0 and drbd1 and created a volume group with > > logical volumes. I resized my lv after adding drbd2 as well. > > Furthermore one drbd (without LVM) and heatbeat works without > > problems. > > > > So far so good, but how does heartbeat put all three drbd to > > primary, do a vgchange -ay and finally mount my logical volumes > > after a reboot or when node1 dies? It was even not possible to > > handle just the three drbd-devices by heartbeat. > > > > I tried different haressources: > > > > node1 IPaddr2::external-IP/24/eth0 drbddisk::raid > > Filesystem::/dev/drbd0 > > > > I know that I need my mount point and ext4 here but I don't want it > > to be mounted because a lv should be mounted. I only want it to be > > primary. > > > > Now the setup with three devices. > > > > node1 IPaddr2::external-IP/24/eth0 > > node1 drbddisk::raid Filesystem::/dev/drbd0 > > node1 drbddisk::raid2 Filesystem::/dev/drbd1 > > node1 drbddisk::raid3 Filesystem::/dev/drbd2 > > > > This doesn't work at all. After a reboot all devices are in state > > secondary/unknown. It is possible to get them back to normal > > (primary/secondary up-to-date) by hand though. I have to connect, > > perform "drbdsetup /dev/drbdX primary -o" and restart drbd. > > > > drbd and LVM together: > > > > node1 drbddisk::raid LVM::drbd-vg > > Filesystem::/dev/mapper/drbd-vg-lv1::/mnt/lv1::ext4 > > > > Works not at all. > > > > Thanx a lot for guidance! > > > > > > Oliver > > > > > > ################### The other configs ########################### > > > > #/etc/drbd.conf > > resource raid { > > > > protocol C; > > > > handlers { > > pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f"; > > pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f"; > > local-io-error "echo o > /proc/sysrq-trigger ; halt -f"; > > outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater"; > > } > > > > startup { wfc-timeout 0; degr-wfc-timeout 120; } > > disk { fencing resource-only; } > > syncer { rate 50M; al-extents 257; } > > > > net { > > after-sb-0pri discard-younger-primary; > > after-sb-1pri consensus; > > after-sb-2pri disconnect; > > rr-conflict call-pri-lost; > > ping-timeout 20; > > } > > > > on node1 { > > device /dev/drbd0; > > disk /dev/sda6; > > address 192.168.1.1:7788; > > meta-disk internal; > > } > > > > on node2 { > > device /dev/drbd0; > > disk /dev/sda6; > > address 192.168.1.2:7788; > > meta-disk internal; > > } > > > > } > > > > resource raid2 { > > > > protocol C; > > > > handlers { > > pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f"; > > pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f"; > > local-io-error "echo o > /proc/sysrq-trigger ; halt -f"; > > outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater"; > > } > > > > startup { wfc-timeout 0; degr-wfc-timeout 120; } > > disk { fencing resource-only; } > > syncer { rate 50M; al-extents 257; } > > > > net { > > after-sb-0pri discard-younger-primary; > > after-sb-1pri consensus; > > after-sb-2pri disconnect; > > rr-conflict call-pri-lost; > > ping-timeout 20; > > } > > > > on node1 { > > device /dev/drbd1; > > disk /dev/sda7; > > address 192.168.2.1:7788; > > meta-disk internal; > > } > > > > on node2 { > > device /dev/drbd1; > > disk /dev/sda7; > > address 192.168.2.2:7788; > > meta-disk internal; > > } > > > > } > > > > resource raid3 { > > > > protocol C; > > > > handlers { > > pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f"; > > pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f"; > > local-io-error "echo o > /proc/sysrq-trigger ; halt -f"; > > outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater"; > > } > > > > startup { wfc-timeout 0; degr-wfc-timeout 120; } > > disk { fencing resource-only; } > > syncer { rate 50M; al-extents 257; } > > > > net { > > after-sb-0pri discard-younger-primary; > > after-sb-1pri consensus; > > after-sb-2pri disconnect; > > rr-conflict call-pri-lost; > > ping-timeout 20; > > } > > > > on node1 { > > device /dev/drbd2; > > disk /dev/sda8; > > address 192.168.3.1:7788; > > meta-disk internal; > > } > > > > on node2 { > > device /dev/drbd2; > > disk /dev/sda8; > > address 192.168.3.2:7788; > > meta-disk internal; > > } > > > > } > > > > #/etc/network/interfaces on node2 > > auto eth1 > > iface eth1 inet static > > address 192.168.1.2 > > netmask 255.255.255.0 > > network 192.168.1.0 > > broadcast 192.168.1.255 > > > > auto eth1:0 > > iface eth1:0 inet static > > name Ethernet alias LAN card > > address 192.168.2.2 > > netmask 255.255.255.0 > > broadcast 192.168.2.255 > > network 192.168.2.0 > > > > auto eth1:1 > > iface eth1:1 inet static > > name Ethernet alias LAN card > > address 192.168.3.2 > > netmask 255.255.255.0 > > broadcast 192.168.3.255 > > network 192.168.3.0 > > > > > > > > _______________________________________________ > > drbd-user mailing list > > drbd-user at lists.linbit.com > > http://lists.linbit.com/mailman/listinfo/drbd-user > > > Hi, > > first advise: > > do not use heartbeat version 1 - style configurations any more. You > will not get any support from anybody no more. > > Please use a distinct cluster manger like pacemaker. See > www.clusterlabs,org. > > Greetings, > > Michael Schwartzkopff. > > _______________________________________________ > drbd-user mailing list > drbd-user at lists.linbit.com > http://lists.linbit.com/mailman/listinfo/drbd-user >