Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hello - I have the following network configuration on both storage servers: bond0 --> 10.255.255.x/24 (shared storage network - here I have KVM node who are exporting iscsi targets from the ha-cluster) bond1 --> 10.13.0.x/24 (This should be the replication link, were both clusters replicate/sync/update) eth0 --> 10.12.33.x/24 (only admin network) eth1 --> 10.14.33.x/24 (here I want the ha heartbeat to go over) -----> crm configure show node $id="e471b446-a7e2-4253-a257-bda343d7c13d" sm-storage-1b \ attributes standby="off" node $id="fe27f2e0-d551-4495-bfb9-819d31884a65" sm-storage-1a \ attributes standby="off" primitive ha_drbd ocf:linbit:drbd \ params drbd_resource="vm3" drbdconf="/etc/drbd.conf" \ op monitor interval="59s" role="Master" timeout="30s" \ op monitor interval="60s" role="Slave" timeout="30s" \ meta is-managed="true" primitive ip_drbd ocf:heartbeat:IPaddr2 \ params ip="10.255.255.205" nic="bond0" \ meta is-managed="true" primitive iscsi lsb:iscsi-target \ meta is-managed="true" primitive lvm_drbd ocf:heartbeat:LVM \ params volgrpname="vg_ralf1" exclusive="true" \ meta is-managed="true" group drbdd lvm_drbd iscsi ip_drbd \ meta target-role="Started" ms ms_drbd_fail ha_drbd \ meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1" notify="true" target-role="Started" colocation col_drbd inf: drbdd ms_drbd_fail:Master order drbd_after inf: ms_drbd_fail:promote drbdd:start property $id="cib-bootstrap-options" \ dc-deadtime="60" \ cluster-delay="60" \ stonith-enabled="false" \ default-action-timeout="20" \ stonith-timeout="60" \ dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \ cluster-infrastructure="Heartbeat" \ last-lrm-refresh="1275383370" <---- end ---> ha.cf debugfile /var/log/ha-debug logfacility local0 mcast eth1 225.0.0.1 694 1 0 ping 10.14.33.118 ping 10.14.33.119 respawn hacluster /usr/lib/heartbeat/dopd apiauth dopd gid=haclient uid=hacluster node sm-storage-1a node sm-storage-1b crm yes <--- --> /etc/drbd.conf global { usage-count yes; } common { syncer { rate 512M; } } resource vm3 { protocol C; startup { wfc-timeout 0; degr-wfc-timeout 120; # become-primary-on both; } handlers { fence-peer "/usr/lib/heartbeat/drbd-peer-outdater -t 5"; } disk { on-io-error detach; fencing resource-only; } net { cram-hmac-alg sha1; # allow-two-primaries; # Wichtig für Primary/Primary Setup } on sm-storage-1a { device /dev/drbd0; disk /dev/sde1; address 10.255.255.203:7788; meta-disk internal; } on sm-storage-1b { device /dev/drbd0; disk /dev/sde1; address 10.255.255.204:7788; meta-disk internal; } } <--- sm-storage-1a = 10.12.33.118(eth0), 10.255.255.203(bond0), 10.13.0.118(bond1), 10.14.33.118(eth1) sm-storage-1b = 10.12.33.119(eth0), 10.255.255.204(bond0), 10.13.0.119(bond1), 10.14.33.119(eth1) Question: How can I divide shared storage network and drbd replication network? Is this a crm configuration or a ha.cf configuration. I'm confused. Thank you for your help. Ralf