<div dir="auto"><div><br><div class="gmail_extra"><br><div class="gmail_quote">On 10 Feb 2018 5:02 am, "Julien Escario" <<a href="mailto:escario@azylog.net">escario@azylog.net</a>> wrote:<br type="attribution"><blockquote class="quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">Hello,<br>
I'm just doing a lab about zpool as storage backend for DRBD (storing VM images<br>
with Proxmox).<br>
<br>
Right now, it's pretty good once tuned and I've been able to achieve 500MB/s<br>
write speed with just a little curiosity about concurrent write from both<br>
hypervisors cluster but that's not the point here.<br>
<br>
To complete resiliancy tests, I simplify unplugged a disk from a node. My toughs<br>
was DRBD was just going to detect ZFS failure and detach the ressources from<br>
failed device.<br></blockquote></div></div></div><div dir="auto"><br></div><div dir="auto"><div class="gmail_extra"><div class="gmail_quote"><blockquote class="quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"></blockquote></div></div></div><div dir="auto">Did you tell it to? <a href="https://docs.linbit.com/doc/users-guide-84/s-configure-io-error-behavior/">https://docs.linbit.com/doc/users-guide-84/s-configure-io-error-behavior/</a></div><div dir="auto"><br></div><div dir="auto"><div class="gmail_extra"><div class="gmail_quote"><blockquote class="quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
<br>
But ... nothing. I/O just hangs on VMs ran on the 'failed' node.<br>
<br>
My zpool status :<br>
<br>
NAME STATE READ WRITE CKSUM<br>
drbdpool UNAVAIL 0 0 0 insufficient replicas<br>
sda UNAVAIL 0 0 0<br>
<br>
but drbdadm show this for locally hosted VM (on the failed node) :<br>
vm-101-disk-1 role:Primary<br>
disk:UpToDate<br>
hyper-test-02 role:Secondary<br>
peer-disk:UpToDate<br>
<br>
and remote VM (on the 'sane' node from failed node point of view) :<br>
vm-104-disk-1 role:Secondary<br>
disk:Consistent<br>
hyper-test-02 connection:NetworkFailure<br>
<br>
<br>
So it seems that DRBD didn't detect the I/O failure.<br>
<br>
Is there a way to force automatic failover in this case ? I probably missed a<br>
detection mecanism.<br>
<br>
Best regards,<br>
Julien Escario<br>
<br>
______________________________<wbr>_________________<br>
drbd-user mailing list<br>
<a href="mailto:drbd-user@lists.linbit.com">drbd-user@lists.linbit.com</a><br>
<a href="http://lists.linbit.com/mailman/listinfo/drbd-user" rel="noreferrer" target="_blank">http://lists.linbit.com/<wbr>mailman/listinfo/drbd-user</a><br>
</blockquote></div><br></div></div></div>