Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
> On Thursday 02 September 2004 21:36, Christian Garling wrote:
>> Hello people,
>>
>> i have a problem with synchronisation. i have 2 servers with escalade
>> 7506-4P raid controller and three maxtor baracuda 160gb harddisks. they
>> are connected through two intel eepro 1000 gigabit ethernet cards
>> (bonding
>> mode 0). when i comment out the rate setting in drbd.conf the initial
>> sync
>> runs with the default value of 250kb/s. but when i use the rate setting
>> it
>> only runs with about 12kb/s. i tested the connection with iptraf monitor
>> and it seems everything to be ok. here is my actually configuration. its
>> very basicly at the moment.
>>
>> resource r0 {
>> protocol C;
>> incon-degr-cmd "halt -f";
>>
>> startup {
>> wfc-timeout 0;
>> degr-wfc-timeout 120;
>> }
>>
>> disk {
>> on-io-error panic;
>> }
>>
>> net {
>> timeout 60; # 6 seconds (unit = 0.1 seconds)
>> connect-int 10; # 10 seconds (unit = 1 second)
>> ping-int 10; # 10 seconds (unit = 1 second)
>> ko-count 5;
>> on-disconnect reconnect;
>> }
>>
>> syncer {
>> rate 10M;
>> }
>>
>> on node01 {
>> device /dev/drbd0;
>> disk /dev/sda1;
>> address 10.0.0.10:7788;
>> meta-disk internal;
>> }
>>
>> on node02 {
>> device /dev/drbd0;
>> disk /dev/sda1;
>> address 10.0.0.20:7788;
>> meta-disk internal;
>> }
>> }
>>
>
> I would really onjoy to see such a cluster in real live (I mean one
> that does only 12kb/s).
>
> I installed a csync2/DRBD/heartbeat cluster yesterday. The two boxes
> had 3Ware Escalades 9xxxx Controllers.
>
> We did some primary crash simulation test and had 20MB/sec resync
> simultaniously on two drbd-resources resyncing in parallel.
> (rate was set to 20M. Probabely these machines would do even more)
>
> Here the usual questions:
> Have you tested the bandwith of your network link ? How ? What numbers
> do you get ?
> Have you tested the bandwith of your disks ? How ? What numbers
> do you get ?
> Which kernel ? Which drbd release ? Hardware ?
> Have you tested without bonding ?
> Are you using jumbo-frames ? MTU ?
>
> -Philipp
> --
> : Dipl-Ing Philipp Reisner Tel +43-1-8178292-50 :
> : LINBIT Information Technologies GmbH Fax +43-1-8178292-82 :
> : Schönbrunnerstr 244, 1120 Vienna, Austria http://www.linbit.com :
> _______________________________________________
> drbd-user mailing list
> drbd-user at lists.linbit.com
> http://lists.linbit.com/mailman/listinfo/drbd-user
>
>
Hello,
Hardware: Tyan Thunder i7501 Pro, 512MB RAM, Intel Xeon 2400, 3Ware
Escalade 7506-4P RAID Controller, 3x Seagate Baracuda 160GB (RAID 5), 2x
Intel Etherexpress 1000MBit (no Jumbo-Frames, MTU on default value)
Software: Debian GNU/Linux Woody, Kernel 2.4.26, DRDB 0.7.3
Network link tested with iptraf. I sent a 190MB Package over the two
gigabit cards and measured a speed of 12MB/s. So I think the network cards
are alright. I tried with bonding mode 0 and 1. I can´t test it without
bonding, because the RAID on the second node rebuilds at the moment.
I dont tested the harddisks.
Greetings,
Christian Garling