[DRBD-user] high io when diskless node added to the storage pool
Alex Kolesnik
drbd-user at abisoft.biz
Tue Sep 3 14:01:46 CEST 2019
Hi,
I have 3 proxmox nodes and just two of them have drbd storage backend:
LINSTOR ==> sp list
╭─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
┊ StoragePool ┊ Node ┊ Driver ┊ PoolName ┊ FreeCapacity ┊ TotalCapacity ┊ SupportsSnapshots ┊ State ┊
╞═════════════════════════════════════════════════════════════════════════════════════════════════════════════════════╡
┊ drbdpool ┊ vm-box-2 ┊ LVM_THIN ┊ vg_vm-box-2/drbdpool ┊ 297.99 GiB ┊ 300 GiB ┊ true ┊ Ok ┊
┊ drbdpool ┊ vm-box-4 ┊ LVM_THIN ┊ vg_vm-box-4/drbdpool ┊ 297.99 GiB ┊ 300 GiB ┊ true ┊ Ok ┊
╰─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
It used to work well moving VMs' virtual drives back and forth from local
storage to drbdpool.
After adding a diskless node to the pool:
LINSTOR ==> sp list
╭─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮
┊ StoragePool ┊ Node ┊ Driver ┊ PoolName ┊ FreeCapacity ┊ TotalCapacity ┊ SupportsSnapshots ┊ State ┊
╞═════════════════════════════════════════════════════════════════════════════════════════════════════════════════════╡
┊ drbdpool ┊ vm-box-2 ┊ LVM_THIN ┊ vg_vm-box-2/drbdpool ┊ 297.99 GiB ┊ 300 GiB ┊ true ┊ Ok ┊
┊ drbdpool ┊ vm-box-3 ┊ DISKLESS ┊ ┊ ┊ ┊ false ┊ Ok ┊
┊ drbdpool ┊ vm-box-4 ┊ LVM_THIN ┊ vg_vm-box-4/drbdpool ┊ 297.99 GiB ┊ 300 GiB ┊ true ┊ Ok ┊
╰─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯
moving a drive to drbdpool increases nodes' IO enormously while nothing seems to
be going on (well, the disk seems to be moving but VERY slow). The log displays
just this w/o any progress, so I had to stop the disk moving:
create full clone of drive scsi0 (LVM-Storage:126/vm-126-disk-0.qcow2)
trying to acquire cfs lock 'storage-drbdpool' ...
transferred: 0 bytes remaining: 10739277824 bytes total: 10739277824 bytes progression: 0.00 %
Did I miss something in the linstor configuration?
PS Am I correct that redundancy should be kept as 2 (in /etc/pve/storage.cfg)
when only two nodes have drbd storage backends?
--
Best regards,
Alex Kolesnik
More information about the drbd-user
mailing list