[Bug 231117] I/O lockups inside bhyve vms
bugzilla-noreply at freebsd.org
bugzilla-noreply at freebsd.org
Tue May 7 08:03:02 UTC 2019
https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=231117
Mateusz Kwiatkowski <kwiat3k at panic.pl> changed:
What |Removed |Added
----------------------------------------------------------------------------
CC| |kwiat3k at panic.pl
--- Comment #19 from Mateusz Kwiatkowski <kwiat3k at panic.pl> ---
I have very similar problem as described in this issue: I/O in guests hangs.
I've experienced this with FreeBSD 11.2, 12.0 (both with ZFS inside) and Ubuntu
18.04 (ext4) guests.
This started happening after migrating guests from old hypervisor running 12.0
on Xeon, to the new on running CURRENT (r347183) on AMD Epyc. On old hypervisor
these VMs were running stable for couple of months.
On the new hypervisor there's plenty of free resources. Swap is disabled.
Mem: 3761M Active, 1636M Inact, 5802M Wired, 51G Free
ARC: 4000M Total, 487M MFU, 3322M MRU, 3456K Anon, 129M Header, 59M Other
2228M Compressed, 3202M Uncompressed, 1.44:1 Ratio
vfs.zfs.arc_min: 8215740416
vfs.zfs.arc_max: 52582796492
Procstat from bhyve provess:
root at utgard:~ # procstat -kk 95379
PID TID COMM TDNAME KSTACK
95379 101075 bhyve mevent mi_switch+0x174
sleepq_switch+0x110 sleepq_catch_signals+0x3e7 sleepq_wait_sig+0xf _sleep+0x2d0
kqueue_kevent+0xa94 kern_kevent_fp+0x95 kern_kevent+0x9f
kern_kevent_generic+0x70 sys_kevent+0x61 amd64_syscall+0x276
fast_syscall_common+0x101
95379 101258 bhyve blk-4:0-0 mi_switch+0x174
sleepq_switch+0x110 sleepq_catch_signals+0x3e7 sleepq_wait_sig+0xf _sleep+0x2d0
umtxq_sleep+0x153 do_wait+0x206 __umtx_op_wait_uint_private+0x7e
amd64_syscall+0x276 fast_syscall_common+0x101
95379 101259 bhyve blk-4:0-1 mi_switch+0x174
sleepq_switch+0x110 sleepq_catch_signals+0x3e7 sleepq_wait_sig+0xf _sleep+0x2d0
umtxq_sleep+0x153 do_wait+0x206 __umtx_op_wait_uint_private+0x7e
amd64_syscall+0x276 fast_syscall_common+0x101
95379 101260 bhyve blk-4:0-2 mi_switch+0x174
sleepq_switch+0x110 sleepq_catch_signals+0x3e7 sleepq_wait_sig+0xf _sleep+0x2d0
umtxq_sleep+0x153 do_wait+0x206 __umtx_op_wait_uint_private+0x7e
amd64_syscall+0x276 fast_syscall_common+0x101
95379 101261 bhyve blk-4:0-3 mi_switch+0x174
sleepq_switch+0x110 sleepq_catch_signals+0x3e7 sleepq_wait_sig+0xf _sleep+0x2d0
umtxq_sleep+0x153 do_wait+0x206 __umtx_op_wait_uint_private+0x7e
amd64_syscall+0x276 fast_syscall_common+0x101
95379 101262 bhyve blk-4:0-4 mi_switch+0x174
sleepq_switch+0x110 sleepq_catch_signals+0x3e7 sleepq_wait_sig+0xf _sleep+0x2d0
umtxq_sleep+0x153 do_wait+0x206 __umtx_op_wait_uint_private+0x7e
amd64_syscall+0x276 fast_syscall_common+0x101
95379 101263 bhyve blk-4:0-5 mi_switch+0x174
sleepq_switch+0x110 sleepq_catch_signals+0x3e7 sleepq_wait_sig+0xf _sleep+0x2d0
umtxq_sleep+0x153 do_wait+0x206 __umtx_op_wait_uint_private+0x7e
amd64_syscall+0x276 fast_syscall_common+0x101
95379 101264 bhyve blk-4:0-6 mi_switch+0x174
sleepq_switch+0x110 sleepq_catch_signals+0x3e7 sleepq_wait_sig+0xf _sleep+0x2d0
umtxq_sleep+0x153 do_wait+0x206 __umtx_op_wait_uint_private+0x7e
amd64_syscall+0x276 fast_syscall_common+0x101
95379 101265 bhyve blk-4:0-7 mi_switch+0x174
sleepq_switch+0x110 sleepq_catch_signals+0x3e7 sleepq_wait_sig+0xf _sleep+0x2d0
umtxq_sleep+0x153 do_wait+0x206 __umtx_op_wait_uint_private+0x7e
amd64_syscall+0x276 fast_syscall_common+0x101
95379 101266 bhyve vtnet-5:0 tx mi_switch+0x174
sleepq_switch+0x110 sleepq_catch_signals+0x3e7 sleepq_wait_sig+0xf _sleep+0x2d0
umtxq_sleep+0x153 do_wait+0x206 __umtx_op_wait_uint_private+0x7e
amd64_syscall+0x276 fast_syscall_common+0x101
95379 101267 bhyve vcpu 0 mi_switch+0x174
sleepq_switch+0x110 sleepq_timedwait+0x4f msleep_spin_sbt+0x144 vm_run+0x970
vmmdev_ioctl+0x7ea devfs_ioctl+0xca VOP_IOCTL_APV+0x63 vn_ioctl+0x124
devfs_ioctl_f+0x1f kern_ioctl+0x28a sys_ioctl+0x15d amd64_syscall+0x276
fast_syscall_common+0x101
95379 101268 bhyve vcpu 1 mi_switch+0x174
sleepq_switch+0x110 sleepq_timedwait+0x4f msleep_spin_sbt+0x144 vm_run+0x970
vmmdev_ioctl+0x7ea devfs_ioctl+0xca VOP_IOCTL_APV+0x63 vn_ioctl+0x124
devfs_ioctl_f+0x1f kern_ioctl+0x28a sys_ioctl+0x15d amd64_syscall+0x276
fast_syscall_common+0x101
I will be happy to provide more information to help solving this issue.
--
You are receiving this mail because:
You are the assignee for the bug.
More information about the freebsd-virtualization
mailing list