Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hi Sam,
I tried the newest kernel 2.6.24 with the kernel module mptsas, without
any success.
Then I found the following solution:
http://lists.us.dell.com/pipermail/linux-precision/2007-June/001188.html
Now my performance is much better:
# dd if=/dev/zero bs=4096 count=10000 of=/tmp/testfile oflag=dsync
10000+0 records in
10000+0 records out
40960000 bytes (41 MB) copied, 7.02387 seconds, 5.8 MB/s
Mario
Sam schrieb:
> Mario,
>
> Your results are similar to ours. I have not yet found a solution to
> improve the 5iR controller's performance. I believe there is a newer
> version of the driver available, but I have not attempted to install
> it yet. I will keep you informed of any progress and ask you to do the
> same.
>
> Thanks,
>
> Sam
>
> On Sun, Mar 2, 2008 at 3:27 PM, Mario Peschel <mario at uni.de> wrote:
>> Hi Sam,
>>
>> this is the result of our dd:
>>
>> # dd if=/dev/zero bs=4096 count=10000 of=/tmp/testfile oflag=dsync
>> 10000+0 records in
>> 10000+0 records out
>> 40960000 bytes (41 MB) copied, 263.415 seconds, 155 kB/s
>>
>> Very slow. :-( But it shows that drbd is not the cause for this poor
>> performance.
>>
>> We also got two new Dell R200 servers with the newer SAS 6iR RAID
>> controller, this are the results:
>>
>> # dd if=/dev/zero bs=4096 count=10000 of=/tmp/testfile oflag=dsync
>> 10000+0 records in
>> 10000+0 records out
>> 40960000 bytes (41 MB) copied, 8.22052 seconds, 5.0 MB/s
>>
>> Looks like the newer controller is working better with Linux.
>>
>> But do you have any idea to get the 5iR controller working a bit faster?
>>
>> Mario
>>
>> Sam schrieb:
>>
>>
>>> Hi Mario,
>> >
>> > I think the problem might be with the SAS 5iR. I see similarly poor
>> > performance for small writes on a Dell SC1435 with the same
>> > controller.
>> >
>> > Try the following test and let me know your result:
>> >
>> > dd if=/dev/zero bs=4096 count=10000 of=/some/file/on/your/raid/disk oflag=dsync
>> >
>> > Sam
>> >
>> > On Fri, Feb 29, 2008 at 1:22 AM, Mario Peschel <mario at uni.de> wrote:
>> >> Hello,
>> >>
>> >> we have two brand new Dell Server PE860 Quad-Core Xeon X3220 2.4GHz with
>> >> 4GB memory, 2x250 GB SATA HDD and SAS 5iR RAID controller running RAID1
>> >> each. Both servers are connected with a gigabit crossover cable for DRBD
>> >> replication.
>> >>
>> >> We're using Debian Linux with kernel 2.6.18-fza-5-amd64 (for use with
>> >> OpenVZ) drbd 8.2.4.
>> >>
>> >> We noticed a heavy load when running some performance tests with bonnie++:
>> >>
>> >> rod:/root# bonnie++ -d /vz/tmp -u 1000
>> >> Using uid:1000, gid:1000.
>> >> Writing with putc()...done
>> >> Writing intelligently...done
>> >> Rewriting...done
>> >> Reading with getc()...done
>> >> Reading intelligently...done
>> >> start 'em...done...done...done...
>> >> Create files in sequential order...done.
>> >> Stat files in sequential order...done.
>> >> Delete files in sequential order...done.
>> >> Create files in random order...done.
>> >> Stat files in random order...done.
>> >> Delete files in random order...done.
>> >> Version 1.03 ------Sequential Output------ --Sequential Input-
>> >> --Random-
>> >> -Per Chr- --Block-- -Rewrite- -Per Chr- --Block--
>> >> --Seeks--
>> >> Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP
>> >> /sec %CP
>> >> rod 8G 1790 2 4477 1 5257 1 65267 83 63655 7
>> >> 151.5 0
>> >> ------Sequential Create------ --------Random
>> >> Create--------
>> >> -Create-- --Read--- -Delete-- -Create-- --Read---
>> >> -Delete--
>> >> files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP
>> >> /sec %CP
>> >> 16 +++++ +++ +++++ +++ +++++ +++ +++++ +++ +++++ +++
>> >> +++++ +++
>> >> rod,8G,1790,2,4477,1,5257,1,65267,83,63655,7,151.5,0,16,+++++,+++,+++++,+++,+++++,+++,+++++,+++,+++++,+++,+++++,+++
>> >>
>> >> Our configuration (drbd.conf) on both servers:
>> >>
>> >> global {
>> >> usage-count yes;
>> >> }
>> >> common {
>> >> syncer { rate 40M; }
>> >> }
>> >> resource r0 {
>> >> protocol C;
>> >> handlers {
>> >> pri-on-incon-degr "echo o > /proc/sysrq-trigger ; halt -f";
>> >> pri-lost-after-sb "echo o > /proc/sysrq-trigger ; halt -f";
>> >> local-io-error "echo o > /proc/sysrq-trigger ; halt -f";
>> >> outdate-peer "/usr/lib/heartbeat/drbd-peer-outdater -t 5";
>> >> }
>> >> startup {
>> >> degr-wfc-timeout 120;
>> >> }
>> >> disk {
>> >> on-io-error detach;
>> >> }
>> >> net {
>> >> after-sb-0pri disconnect;
>> >> after-sb-1pri disconnect;
>> >> after-sb-2pri disconnect;
>> >> rr-conflict disconnect;
>> >> }
>> >> syncer {
>> >> rate 40M;
>> >> al-extents 257;
>> >> }
>> >> on rod {
>> >> device /dev/drbd0;
>> >> disk /dev/sda3;
>> >> address 192.168.1.2:7788;
>> >> meta-disk internal;
>> >> }
>> >> on todd {
>> >> device /dev/drbd0;
>> >> disk /dev/sda3;
>> >> address 192.168.1.1:7788;
>> >> meta-disk internal;
>> >> }
>> >> }
>> >>
>> >> This is very slow. Anyone have an idea why it is so or what we can try
>> >> to boost up the performance?
>> >>
>> >> Thanks,
>> >>
>> >> Mario
>> >> _______________________________________________
>> >> drbd-user mailing list
>> >> drbd-user at lists.linbit.com
>> >> http://lists.linbit.com/mailman/listinfo/drbd-user
>> >>
>>
>>