R10: 0000000000000000 R11: 0000000000000246 R12: 00007f0a2e59bf8c R13: 000000000000000b R14: 00007ffe07336010 R15: 00007ffe073360f8 sched: RT throttling activated BUG: workqueue lockup - pool cpus=1 node=0 flags=0x0 nice=0 stuck for 238s! BUG: workqueue lockup - pool cpus=1 node=0 flags=0x0 nice=-20 stuck for 236s! Showing busy workqueues and worker pools: workqueue events: flags=0x0 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=2/256 refcnt=3 pending: psi_avgs_work, kfree_rcu_monitor pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=3/256 refcnt=4 in-flight: 304:linkwatch_event pending: vmstat_shepherd, kfree_rcu_monitor workqueue events_unbound: flags=0x2 pwq 4: cpus=0-1 flags=0x4 nice=0 active=2/512 refcnt=3 pending: flush_memcg_stats_dwork, crng_reseed pwq 4: cpus=0-1 flags=0x4 nice=0 active=2/512 refcnt=3 pending: fsnotify_mark_destroy_workfn, toggle_allocation_gate workqueue events_power_efficient: flags=0x80 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=6/256 refcnt=7 pending: 2*neigh_managed_work, 2*neigh_periodic_work, gc_worker, reg_check_chans_work pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=2/256 refcnt=3 pending: wg_ratelimiter_gc_entries, check_lifetime workqueue mm_percpu_wq: flags=0x8 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: vmstat_update pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: vmstat_update workqueue writeback: flags=0x4a pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wb_workfn workqueue kblockd: flags=0x18 pwq 3: cpus=1 node=0 flags=0x0 nice=-20 active=1/256 refcnt=2 pending: blk_mq_timeout_work workqueue dm_bufio_cache: flags=0x8 pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: work_fn workqueue mld: flags=0x40008 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/1 refcnt=28 pending: mld_ifc_work inactive: 5*mld_ifc_work, 2*mld_dad_work, 9*mld_ifc_work, mld_dad_work, 5*mld_ifc_work, mld_dad_work, 3*mld_ifc_work pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/1 refcnt=99 pending: mld_dad_work inactive: 3*mld_ifc_work, mld_dad_work, 2*mld_ifc_work, 2*mld_dad_work, mld_ifc_work, mld_dad_work, 13*mld_ifc_work, 4*mld_dad_work, 7*mld_ifc_work, mld_dad_work, 2*mld_ifc_work, 3*mld_dad_work, 8*mld_ifc_work, mld_dad_work, mld_ifc_work, mld_dad_work, 3*mld_ifc_work, mld_dad_work, 16*mld_ifc_work, mld_dad_work, 6*mld_ifc_work, mld_dad_work, 14*mld_ifc_work, 2*mld_dad_work, 2*mld_ifc_work workqueue ipv6_addrconf: flags=0x40008 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/1 refcnt=90 in-flight: 25:addrconf_dad_work inactive: 88*addrconf_dad_work pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/1 refcnt=210 in-flight: 299:addrconf_dad_work inactive: 201*addrconf_dad_work, 7*addrconf_verify_work workqueue wg-kex-wg0: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=2/256 refcnt=3 pending: 2*wg_packet_handshake_send_worker workqueue wg-kex-wg1: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=2/256 refcnt=3 pending: 2*wg_packet_handshake_send_worker workqueue wg-kex-wg2: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=2/256 refcnt=3 pending: 2*wg_packet_handshake_send_worker workqueue wg-kex-wg0: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wg_packet_handshake_send_worker workqueue wg-crypt-wg0: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-kex-wg0: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wg_packet_handshake_send_worker workqueue wg-crypt-wg0: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-kex-wg1: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wg_packet_handshake_send_worker workqueue wg-crypt-wg1: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-crypt-wg1: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-kex-wg2: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wg_packet_handshake_send_worker pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wg_packet_handshake_send_worker workqueue wg-crypt-wg2: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-crypt-wg2: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-kex-wg0: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wg_packet_handshake_send_worker workqueue wg-crypt-wg0: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-kex-wg1: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wg_packet_handshake_send_worker workqueue wg-crypt-wg1: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-crypt-wg2: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pool 0: cpus=0 node=0 flags=0x0 nice=0 hung=239s workers=5 idle: 7 392 8 pool 2: cpus=1 node=0 flags=0x0 nice=0 hung=239s workers=4 idle: 37 303 56 Showing backtraces of running workers in stalled CPU-bound worker pools: pool 2: task:kworker/1:0 state:R running task stack:25800 pid:25 ppid:2 flags:0x00004000 Workqueue: ipv6_addrconf addrconf_dad_work Call Trace: context_switch kernel/sched/core.c:5382 [inline] __schedule+0xd6c/0x1570 kernel/sched/core.c:6695 schedule+0xc3/0x170 kernel/sched/core.c:6771 schedule_preempt_disabled+0x13/0x20 kernel/sched/core.c:6830 __mutex_lock_common kernel/locking/mutex.c:679 [inline] __mutex_lock+0x98c/0x14e0 kernel/locking/mutex.c:747 __mutex_lock_slowpath+0xe/0x10 kernel/locking/mutex.c:1035 mutex_lock+0xd6/0x110 kernel/locking/mutex.c:286 rtnl_lock+0x15/0x20 net/core/rtnetlink.c:78 addrconf_dad_work+0xc6/0x16b0 net/ipv6/addrconf.c:4125 process_one_work kernel/workqueue.c:2630 [inline] process_scheduled_works+0x720/0xeb0 kernel/workqueue.c:2703 worker_thread+0x968/0xe00 kernel/workqueue.c:2784 kthread+0x2b6/0x350 kernel/kthread.c:388 ret_from_fork+0x4a/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x11/0x20 arch/x86/entry/entry_64.S:304