Storage blktests - nvme-tcp: WARNING: CPU: 6 PID: 1092 at kernel/workqueue.c:2628 check_flush_dependency+0x118/0x14c
[ 4562.677914] run blktests nvme/003 at 2022-05-08 17:29:17
[ 4562.730453] loop: module loaded
[ 4562.739501] nvmet: adding nsid 1 to subsystem blktests-subsystem-1
[ 4562.747503] nvmet_tcp: enabling port 0 (127.0.0.1:4420)
[ 4562.766946] nvme0: Admin Cmd(0x6), I/O Error (sct 0x0 / sc 0x2) MORE
[ 4562.776704] nvmet: creating discovery controller 1 for subsystem nqn.2014-08.org.nvmexpress.discovery for NQN nqn.2014-08.org.nvmexpress:uuid:00000000-0000-4000-8000-18c04d0db8bb.
[ 4562.793123] nvme nvme1: new ctrl: NQN "nqn.2014-08.org.nvmexpress.discovery", addr 127.0.0.1:4420
[ 4572.817382] nvme0: Admin Cmd(0x6), I/O Error (sct 0x0 / sc 0x2) MORE
[ 4572.824585] nvme nvme1: Removing ctrl: NQN "nqn.2014-08.org.nvmexpress.discovery"
[ 4572.832443] ------------[ cut here ]------------
[ 4572.837052] workqueue: WQ_MEM_RECLAIM nvmet-wq:nvmet_tcp_release_queue_work [nvmet_tcp] is flushing !WQ_MEM_RECLAIM nvmet_tcp_wq:nvmet_tcp_io_work [nvmet_tcp]
[ 4572.837065] WARNING: CPU: 6 PID: 1092 at kernel/workqueue.c:2628 check_flush_dependency+0x118/0x14c
[ 4572.860255] Modules linked in: loop nvme_tcp nvme_fabrics nvmet_tcp nvmet tls rfkill sunrpc vfat fat acpi_ipmi ipmi_ssif igb arm_cmn cppc_cpufreq arm_dsu_pmu acpi_tad fuse zram xfs ast i2c_algo_bit drm_vram_helper drm_kms_helper syscopyarea sysfillrect sysimgblt fb_sys_fops drm_ttm_helper ttm crct10dif_ce nvme ghash_ce sbsa_gwdt drm nvme_core xgene_hwmon ipmi_devintf ipmi_msghandler [last unloaded: scsi_debug]
[ 4572.896575] CPU: 6 PID: 1092 Comm: kworker/6:2 Not tainted 5.18.0-rc5 #1
[ 4572.903264] Hardware name: GIGABYTE R272-P30-JG/MP32-AR0-JG, BIOS F13 (SCP: 1.5.20210426) 05/21/2021
[ 4572.912383] Workqueue: nvmet-wq nvmet_tcp_release_queue_work [nvmet_tcp]
[ 4572.919074] pstate: 804000c9 (Nzcv daIF +PAN -UAO -TCO -DIT -SSBS BTYPE=--)
[ 4572.926023] pc : check_flush_dependency+0x118/0x14c
[ 4572.930890] lr : check_flush_dependency+0x118/0x14c
[ 4572.935756] sp : ffff80000c2bbc10
[ 4572.939059] x29: ffff80000c2bbc10 x28: 0000000000000000 x27: 0000000000000000
[ 4572.946183] x26: ffff07ff8cc77774 x25: fffffbffed066f15 x24: 0000000000000001
[ 4572.953307] x23: ffffc4e967f64560 x22: fffffbffed067000 x21: ffffc4e946393784
[ 4572.960430] x20: ffff07ff8cc77700 x19: ffff07ff9ab3f200 x18: ffffffffffffffff
[ 4572.967553] x17: 0000000000000000 x16: ffffc4e96576e6e0 x15: ffff80008c2bb8ef
[ 4572.974677] x14: 0000000000000001 x13: 204d49414c434552 x12: 5f4d454d5f515721
[ 4572.981801] x11: 00000000ffff7fff x10: 00000000ffff7fff x9 : ffffc4e9657d8020
[ 4572.988924] x8 : 00000000000bffe8 x7 : c0000000ffff7fff x6 : 00000000002bffa8
[ 4572.996048] x5 : 0000000000007fff x4 : ffff80000c2bba60 x3 : 0000000000000001
[ 4573.003171] x2 : 0000000000000000 x1 : 0000000000000027 x0 : 0000000000000092
[ 4573.010295] Call trace:
[ 4573.012729] check_flush_dependency+0x118/0x14c
[ 4573.017248] __flush_work.isra.0+0x200/0x2b0
[ 4573.021508] __cancel_work_timer+0x104/0x170
[ 4573.025767] cancel_work_sync+0x20/0x30
[ 4573.029592] nvmet_tcp_release_queue_work+0x124/0x344 [nvmet_tcp]
[ 4573.035674] process_one_work+0x1f0/0x450
[ 4573.039672] worker_thread+0x190/0x4c0
[ 4573.043411] kthread+0xd4/0xe0
[ 4573.046455] ret_from_fork+0x10/0x20
[ 4573.050020] ---[ end trace 0000000000000000 ]---
https://datawarehouse.cki-project.org/kcidb/tests/3433725
https://datawarehouse.cki-project.org/issue/1195