[DRBD-user] [DRBD-9.0.15-0rc1] Resource "stuck" during live migration

Roland Kammerer roland.kammerer at linbit.com
Thu Jul 26 09:25:50 CEST 2018


On Wed, Jul 25, 2018 at 08:49:02PM +0100, Yannis Milios wrote:
> Hello,
> 
> Currently testing 9.0.15-0rc1 on a 3 node PVE cluster.
> 
> Pkg versions:
> ------------------
> cat /proc/drbd
> version: 9.0.15-0rc1 (api:2/proto:86-114)
> GIT-hash: fc844fc366933c60f7303694ca1dea734dcb39bb build by root at pve1,
> 2018-07-23 18:47:08
> Transports (api:16): tcp (9.0.15-0rc1)
> ii  python-drbdmanage             0.99.18-1
> ii  drbdmanage-proxmox            2.2-1
> ii  drbd-utils                    9.5.0-1
> ---------------------
> Resource=vm-122-disk-1
> Replica count=3
> PVE nodes=pve1,pve2,pve3
> Resource is active on pve2 (Primary), the rest two nodes (pve1,pve2) are
> Secondary.
> 
> Tried to live migrate the VM from pve2 to pve3 and the process stuck just
> before starting. By inspecting dmesg on both nodes (pve2,pve3), I get the
> following crash..
> 
> 
> pve2 (Primary) node:
> https://privatebin.net/?fb5435a42b431af2#4xZpd9D5bYnB000+H3K0noZmkX20fTwGSziv5oO/Zlg=
> 
> pve3(Secondary)node:
> https://privatebin.net/?d3b1638fecb6728f#2StXbwDPT0JlFUKf686RJiR+4hl52jEmmij2UTtnSjs=
> 

We will look into it closer. For now I saw "zfs" in the second trace and
stopped. It is so freaking broken, it is not funny any more (it craps
out with all kinds of BS in our internal infrastructure as well). For
example we had to go back to a xenial kernel because the bionic ones zfs
is that broken :-/ </zfs rant, which I actually really like>

Regards, rck


More information about the drbd-user mailing list