[DRBD-user] IOwaits after upgrading

Olivier Le Cam Olivier.LeCam at crdp.ac-versailles.fr
Sun Mar 10 11:40:40 CET 2013

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.

Hi -

I have upgraded my DRBD storage from Debian/Lenny to Debian/Squeeze and 
iowait increased significantly. I have uploaded a graph for the last 
week CPU utilisation (with a long interrupt in the middle, snmp was not 
working as expected after upgrade) here:

The reason of this upgrade is mainly the heavy load I occasionally face 
and I'd like to give flashcache a try. Now I wonder if I should continue 
that way or downgrade back to Lenny due to the aggravated performance!... :)

I have been able to optimize things a bit (iowait were actually even 
worse the first hours after upgrade which you can see on the graph), 
thanks to good readings here and in the drbd manual. Here is the current 
common config:

common {
	protocol C;

	handlers {

	startup {
		degr-wfc-timeout 120;

	disk {
		on-io-error detach;

	net {
		max-buffers 10000;
		max-epoch-size 10000;
		sndbuf-size 0;
		unplug-watermark 10000;

	syncer {
		rate 90M;
		al-extents 3389;

- scheduler is deadline (tested as well cfq with same results)
- front_merges=0, read_expire=150, write_expire=1500, writes_starved=16
- current DRBD version deployed is 8.3.7

Servers are both identical, 48MB, 3ware RAID controler (9690SA-4I, 
512MB+BBU), direct ethernet bonding 2x1GB. DRBD runs over LVM over 3 x 5 
SAS 10k drives in RAID-5.

The servers always have 500~1000 write IOPS. Stopping DRBD replication 
reduces iowaits by about 50% which seems correct according to the C 

Is that some of you have seen similar behavior? Any 
pointers/comments/feedbacks will be greatly appreciated! Thanks in 


More information about the drbd-user mailing list