Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
On Wednesday 02 June 2010 12:42:53 Ralf W. wrote: > Hello - I have the following network configuration on both storage servers: > bond0 --> 10.255.255.x/24 (shared storage network - here I have KVM node > who are exporting iscsi targets from the ha-cluster) bond1 --> > 10.13.0.x/24 (This should be the replication link, were both clusters > replicate/sync/update) eth0 --> 10.12.33.x/24 (only admin network) > eth1 --> 10.14.33.x/24 (here I want the ha heartbeat to go over) > > -----> crm configure show > node $id="e471b446-a7e2-4253-a257-bda343d7c13d" sm-storage-1b \ > attributes standby="off" > node $id="fe27f2e0-d551-4495-bfb9-819d31884a65" sm-storage-1a \ > attributes standby="off" > primitive ha_drbd ocf:linbit:drbd \ > params drbd_resource="vm3" drbdconf="/etc/drbd.conf" \ > op monitor interval="59s" role="Master" timeout="30s" \ > op monitor interval="60s" role="Slave" timeout="30s" \ > meta is-managed="true" > primitive ip_drbd ocf:heartbeat:IPaddr2 \ > params ip="10.255.255.205" nic="bond0" \ > meta is-managed="true" > primitive iscsi lsb:iscsi-target \ > meta is-managed="true" > primitive lvm_drbd ocf:heartbeat:LVM \ > params volgrpname="vg_ralf1" exclusive="true" \ > meta is-managed="true" > group drbdd lvm_drbd iscsi ip_drbd \ > meta target-role="Started" > ms ms_drbd_fail ha_drbd \ > meta master-max="1" master-node-max="1" clone-max="2" > clone-node-max="1" notify="true" target-role="Started" colocation col_drbd > inf: drbdd ms_drbd_fail:Master > order drbd_after inf: ms_drbd_fail:promote drbdd:start > property $id="cib-bootstrap-options" \ > dc-deadtime="60" \ > cluster-delay="60" \ > stonith-enabled="false" \ > default-action-timeout="20" \ > stonith-timeout="60" \ > dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \ > cluster-infrastructure="Heartbeat" \ > last-lrm-refresh="1275383370" > <---- end > > ---> ha.cf > debugfile /var/log/ha-debug > logfacility local0 > mcast eth1 225.0.0.1 694 1 0 > ping 10.14.33.118 > ping 10.14.33.119 > respawn hacluster /usr/lib/heartbeat/dopd > apiauth dopd gid=haclient uid=hacluster > node sm-storage-1a > node sm-storage-1b > crm yes > <--- > > > --> /etc/drbd.conf > global { usage-count yes; } > common { syncer { rate 512M; } } > resource vm3 { > protocol C; > startup { > wfc-timeout 0; > degr-wfc-timeout 120; > # become-primary-on both; > } > handlers { > fence-peer "/usr/lib/heartbeat/drbd-peer-outdater -t 5"; > } > > disk { > on-io-error detach; > fencing resource-only; > } > net { > cram-hmac-alg sha1; > # allow-two-primaries; # Wichtig für Primary/Primary Setup > } > > on sm-storage-1a { > device /dev/drbd0; > disk /dev/sde1; > address 10.255.255.203:7788; > meta-disk internal; > } > > on sm-storage-1b { > device /dev/drbd0; > disk /dev/sde1; > address 10.255.255.204:7788; > meta-disk internal; > } > } > <--- > > sm-storage-1a = 10.12.33.118(eth0), 10.255.255.203(bond0), > 10.13.0.118(bond1), 10.14.33.118(eth1) sm-storage-1b = 10.12.33.119(eth0), > 10.255.255.204(bond0), 10.13.0.119(bond1), 10.14.33.119(eth1) > > > Question: How can I divide shared storage network and drbd replication > network? Is this a crm configuration or a ha.cf configuration. I'm > confused. Thank you for your help. > > Ralf > > > > > _______________________________________________ > drbd-user mailing list > drbd-user at lists.linbit.com > http://lists.linbit.com/mailman/listinfo/drbd-user > If you connect to just the NICs that are meant for iSCSI, do you then need a further separation? IET listens to all interfaces, but as long you connect to certain IP addresses only those NICs suffer performance I guess. B.