Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hello,
we have two brand new Dell Server PE860 Quad-Core Xeon X3220 2.4GHz with
4GB memory, 2x250 GB SATA HDD and SAS 5iR RAID controller running RAID1
each. Both servers are connected with a gigabit crossover cable for DRBD
replication.
We're using Debian Linux with kernel 2.6.18-fza-5-amd64 (for use with
OpenVZ) drbd 8.2.4.
We noticed a heavy load when running some performance tests with bonnie++:
rod:/root# bonnie++ -d /vz/tmp -u 1000
Using uid:1000, gid:1000.
Writing with putc()...done
Writing intelligently...done
Rewriting...done
Reading with getc()...done
Reading intelligently...done
start 'em...done...done...done...
Create files in sequential order...done.
Stat files in sequential order...done.
Delete files in sequential order...done.
Create files in random order...done.
Stat files in random order...done.
Delete files in random order...done.
Version 1.03 ------Sequential Output------ --Sequential Input-
--Random-
-Per Chr- --Block-- -Rewrite- -Per Chr- --Block--
--Seeks--
Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP
/sec %CP
rod 8G 1790 2 4477 1 5257 1 65267 83 63655 7
151.5 0
------Sequential Create------ --------Random
Create--------
-Create-- --Read--- -Delete-- -Create-- --Read---
-Delete--
files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP
/sec %CP
16 +++++ +++ +++++ +++ +++++ +++ +++++ +++ +++++ +++
+++++ +++
rod,8G,1790,2,4477,1,5257,1,65267,83,63655,7,151.5,0,16,+++++,+++,+++++,+++,+++++,+++,+++++,+++,+++++,+++,+++++,+++
Our configuration (drbd.conf) on both servers:
global {
usage-count yes;
}
common {
syncer { rate 40M; }
}
resource r0 {
protocol C;
handlers {
pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f";
pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f";
local-io-error "echo o > /proc/sysrq-trigger ; halt -f";
outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater -t 5";
}
startup {
degr-wfc-timeout 120;
}
disk {
on-io-error detach;
}
net {
after-sb-0pri disconnect;
after-sb-1pri disconnect;
after-sb-2pri disconnect;
rr-conflict disconnect;
}
syncer {
rate 40M;
al-extents 257;
}
on rod {
device /dev/drbd0;
disk /dev/sda3;
address 192.168.1.2:7788;
meta-disk internal;
}
on todd {
device /dev/drbd0;
disk /dev/sda3;
address 192.168.1.1:7788;
meta-disk internal;
}
}
This is very slow. Anyone have an idea why it is so or what we can try
to boost up the performance?
Thanks,
Mario