Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Am Samstag, 10. April 2010 14:50:30 schrieb Oliver Hoffmann: > Hi list, > > probably more a heartbeat question but maybe some-one can push me in > the right direction. > > I want two storage servers with a bunch of drbd-devices and LVM on top > in order to make snapshots and resizing disk space. The idea is to add a > HD on each node, make it a drbd-device and extend a logical volume > accordingly whenever I run out of space. heartbeat (or pacemaker?) > should take care and do a failover to node2 if node1 dies. Finally > clients should connect via iSCSI or smb or nfs. > > I successfully made drbd0 and drbd1 and created a volume group with > logical volumes. I resized my lv after adding drbd2 as well. Furthermore > one drbd (without LVM) and heatbeat works without problems. > > So far so good, but how does heartbeat put all three drbd to primary, > do a vgchange -ay and finally mount my logical volumes after a reboot > or when node1 dies? It was even not possible to handle just the three > drbd-devices by heartbeat. > > I tried different haressources: > > node1 IPaddr2::external-IP/24/eth0 drbddisk::raid > Filesystem::/dev/drbd0 > > I know that I need my mount point and ext4 here but I don't want it to > be mounted because a lv should be mounted. I only want it to be primary. > > Now the setup with three devices. > > node1 IPaddr2::external-IP/24/eth0 > node1 drbddisk::raid Filesystem::/dev/drbd0 > node1 drbddisk::raid2 Filesystem::/dev/drbd1 > node1 drbddisk::raid3 Filesystem::/dev/drbd2 > > This doesn't work at all. After a reboot all devices are in state > secondary/unknown. It is possible to get them back to normal > (primary/secondary up-to-date) by hand though. I have to connect, > perform "drbdsetup /dev/drbdX primary -o" and restart drbd. > > drbd and LVM together: > > node1 drbddisk::raid LVM::drbd-vg > Filesystem::/dev/mapper/drbd-vg-lv1::/mnt/lv1::ext4 > > Works not at all. > > Thanx a lot for guidance! > > > Oliver > > > ################### The other configs ########################### > > #/etc/drbd.conf > resource raid { > > protocol C; > > handlers { > pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f"; > pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f"; > local-io-error "echo o > /proc/sysrq-trigger ; halt -f"; > outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater"; > } > > startup { wfc-timeout 0; degr-wfc-timeout 120; } > disk { fencing resource-only; } > syncer { rate 50M; al-extents 257; } > > net { > after-sb-0pri discard-younger-primary; > after-sb-1pri consensus; > after-sb-2pri disconnect; > rr-conflict call-pri-lost; > ping-timeout 20; > } > > on node1 { > device /dev/drbd0; > disk /dev/sda6; > address 192.168.1.1:7788; > meta-disk internal; > } > > on node2 { > device /dev/drbd0; > disk /dev/sda6; > address 192.168.1.2:7788; > meta-disk internal; > } > > } > > resource raid2 { > > protocol C; > > handlers { > pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f"; > pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f"; > local-io-error "echo o > /proc/sysrq-trigger ; halt -f"; > outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater"; > } > > startup { wfc-timeout 0; degr-wfc-timeout 120; } > disk { fencing resource-only; } > syncer { rate 50M; al-extents 257; } > > net { > after-sb-0pri discard-younger-primary; > after-sb-1pri consensus; > after-sb-2pri disconnect; > rr-conflict call-pri-lost; > ping-timeout 20; > } > > on node1 { > device /dev/drbd1; > disk /dev/sda7; > address 192.168.2.1:7788; > meta-disk internal; > } > > on node2 { > device /dev/drbd1; > disk /dev/sda7; > address 192.168.2.2:7788; > meta-disk internal; > } > > } > > resource raid3 { > > protocol C; > > handlers { > pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f"; > pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f"; > local-io-error "echo o > /proc/sysrq-trigger ; halt -f"; > outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater"; > } > > startup { wfc-timeout 0; degr-wfc-timeout 120; } > disk { fencing resource-only; } > syncer { rate 50M; al-extents 257; } > > net { > after-sb-0pri discard-younger-primary; > after-sb-1pri consensus; > after-sb-2pri disconnect; > rr-conflict call-pri-lost; > ping-timeout 20; > } > > on node1 { > device /dev/drbd2; > disk /dev/sda8; > address 192.168.3.1:7788; > meta-disk internal; > } > > on node2 { > device /dev/drbd2; > disk /dev/sda8; > address 192.168.3.2:7788; > meta-disk internal; > } > > } > > #/etc/network/interfaces on node2 > auto eth1 > iface eth1 inet static > address 192.168.1.2 > netmask 255.255.255.0 > network 192.168.1.0 > broadcast 192.168.1.255 > > auto eth1:0 > iface eth1:0 inet static > name Ethernet alias LAN card > address 192.168.2.2 > netmask 255.255.255.0 > broadcast 192.168.2.255 > network 192.168.2.0 > > auto eth1:1 > iface eth1:1 inet static > name Ethernet alias LAN card > address 192.168.3.2 > netmask 255.255.255.0 > broadcast 192.168.3.255 > network 192.168.3.0 > > > > _______________________________________________ > drbd-user mailing list > drbd-user at lists.linbit.com > http://lists.linbit.com/mailman/listinfo/drbd-user Hi, first advise: do not use heartbeat version 1 - style configurations any more. You will not get any support from anybody no more. Please use a distinct cluster manger like pacemaker. See www.clusterlabs,org. Greetings, Michael Schwartzkopff.