INFO: task kworker/R-rcu_g:4 blocked for more than 142 seconds. Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:kworker/R-rcu_g state:D stack:28720 pid:4 tgid:4 ppid:2 flags:0x00004000 Call Trace: context_switch kernel/sched/core.c:5315 [inline] __schedule+0x1895/0x4b30 kernel/sched/core.c:6674 __schedule_loop kernel/sched/core.c:6751 [inline] schedule+0x14b/0x320 kernel/sched/core.c:6766 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6823 __mutex_lock_common kernel/locking/mutex.c:684 [inline] __mutex_lock+0x6a7/0xd70 kernel/locking/mutex.c:752 worker_detach_from_pool kernel/workqueue.c:2727 [inline] rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 kthread+0x2f0/0x390 kernel/kthread.c:389 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 INFO: task kworker/0:2:937 blocked for more than 143 seconds. Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:kworker/0:2 state:D stack:24120 pid:937 tgid:937 ppid:2 flags:0x00004000 Workqueue: 0x0 (mld) Call Trace: context_switch kernel/sched/core.c:5315 [inline] __schedule+0x1895/0x4b30 kernel/sched/core.c:6674 __schedule_loop kernel/sched/core.c:6751 [inline] schedule+0x14b/0x320 kernel/sched/core.c:6766 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6823 __mutex_lock_common kernel/locking/mutex.c:684 [inline] __mutex_lock+0x6a7/0xd70 kernel/locking/mutex.c:752 worker_attach_to_pool+0x31/0x390 kernel/workqueue.c:2669 create_worker+0x491/0x720 kernel/workqueue.c:2811 maybe_create_worker kernel/workqueue.c:3054 [inline] manage_workers kernel/workqueue.c:3106 [inline] worker_thread+0x318/0xd30 kernel/workqueue.c:3366 kthread+0x2f0/0x390 kernel/kthread.c:389 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 INFO: task kworker/R-wg-cr:5266 blocked for more than 143 seconds. Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:kworker/R-wg-cr state:D stack:26712 pid:5266 tgid:5266 ppid:2 flags:0x00004000 Workqueue: 0x0 (wg-crypt-wg1) Call Trace: context_switch kernel/sched/core.c:5315 [inline] __schedule+0x1895/0x4b30 kernel/sched/core.c:6674 __schedule_loop kernel/sched/core.c:6751 [inline] schedule+0x14b/0x320 kernel/sched/core.c:6766 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6823 __mutex_lock_common kernel/locking/mutex.c:684 [inline] __mutex_lock+0x6a7/0xd70 kernel/locking/mutex.c:752 worker_detach_from_pool kernel/workqueue.c:2727 [inline] rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 kthread+0x2f0/0x390 kernel/kthread.c:389 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 INFO: task kworker/R-wg-cr:5268 blocked for more than 144 seconds. Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:kworker/R-wg-cr state:D stack:28656 pid:5268 tgid:5268 ppid:2 flags:0x00004000 Call Trace: context_switch kernel/sched/core.c:5315 [inline] __schedule+0x1895/0x4b30 kernel/sched/core.c:6674 __schedule_loop kernel/sched/core.c:6751 [inline] schedule+0x14b/0x320 kernel/sched/core.c:6766 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6823 __mutex_lock_common kernel/locking/mutex.c:684 [inline] __mutex_lock+0x6a7/0xd70 kernel/locking/mutex.c:752 worker_detach_from_pool kernel/workqueue.c:2727 [inline] rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 kthread+0x2f0/0x390 kernel/kthread.c:389 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 INFO: task kworker/R-wg-cr:5271 blocked for more than 144 seconds. Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:kworker/R-wg-cr state:D stack:27248 pid:5271 tgid:5271 ppid:2 flags:0x00004000 Workqueue: 0x0 (wg-crypt-wg0) Call Trace: context_switch kernel/sched/core.c:5315 [inline] __schedule+0x1895/0x4b30 kernel/sched/core.c:6674 __schedule_loop kernel/sched/core.c:6751 [inline] schedule+0x14b/0x320 kernel/sched/core.c:6766 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6823 __mutex_lock_common kernel/locking/mutex.c:684 [inline] __mutex_lock+0x6a7/0xd70 kernel/locking/mutex.c:752 worker_detach_from_pool kernel/workqueue.c:2727 [inline] rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 kthread+0x2f0/0x390 kernel/kthread.c:389 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 INFO: task kworker/R-wg-cr:5277 blocked for more than 144 seconds. Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:kworker/R-wg-cr state:D stack:26912 pid:5277 tgid:5277 ppid:2 flags:0x00004000 Workqueue: 0x0 (wg-crypt-wg1) Call Trace: context_switch kernel/sched/core.c:5315 [inline] __schedule+0x1895/0x4b30 kernel/sched/core.c:6674 __schedule_loop kernel/sched/core.c:6751 [inline] schedule+0x14b/0x320 kernel/sched/core.c:6766 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6823 __mutex_lock_common kernel/locking/mutex.c:684 [inline] __mutex_lock+0x6a7/0xd70 kernel/locking/mutex.c:752 worker_detach_from_pool kernel/workqueue.c:2727 [inline] rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 kthread+0x2f0/0x390 kernel/kthread.c:389 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 INFO: task kworker/R-wg-cr:5283 blocked for more than 145 seconds. Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:kworker/R-wg-cr state:D stack:26912 pid:5283 tgid:5283 ppid:2 flags:0x00004000 Workqueue: 0x0 (wg-crypt-wg1) Call Trace: context_switch kernel/sched/core.c:5315 [inline] __schedule+0x1895/0x4b30 kernel/sched/core.c:6674 __schedule_loop kernel/sched/core.c:6751 [inline] schedule+0x14b/0x320 kernel/sched/core.c:6766 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6823 __mutex_lock_common kernel/locking/mutex.c:684 [inline] __mutex_lock+0x6a7/0xd70 kernel/locking/mutex.c:752 worker_detach_from_pool kernel/workqueue.c:2727 [inline] rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 kthread+0x2f0/0x390 kernel/kthread.c:389 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 INFO: task kworker/R-wg-cr:5284 blocked for more than 145 seconds. Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:kworker/R-wg-cr state:D stack:26912 pid:5284 tgid:5284 ppid:2 flags:0x00004000 Workqueue: 0x0 (wg-crypt-wg2) Call Trace: context_switch kernel/sched/core.c:5315 [inline] __schedule+0x1895/0x4b30 kernel/sched/core.c:6674 __schedule_loop kernel/sched/core.c:6751 [inline] schedule+0x14b/0x320 kernel/sched/core.c:6766 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6823 __mutex_lock_common kernel/locking/mutex.c:684 [inline] __mutex_lock+0x6a7/0xd70 kernel/locking/mutex.c:752 worker_detach_from_pool kernel/workqueue.c:2727 [inline] rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 kthread+0x2f0/0x390 kernel/kthread.c:389 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 INFO: task kworker/1:6:5291 blocked for more than 145 seconds. Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:kworker/1:6 state:D stack:28016 pid:5291 tgid:5291 ppid:2 flags:0x00004000 Workqueue: 0x0 (wg-crypt-wg2) Call Trace: context_switch kernel/sched/core.c:5315 [inline] __schedule+0x1895/0x4b30 kernel/sched/core.c:6674 __schedule_loop kernel/sched/core.c:6751 [inline] schedule+0x14b/0x320 kernel/sched/core.c:6766 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6823 __mutex_lock_common kernel/locking/mutex.c:684 [inline] __mutex_lock+0x6a7/0xd70 kernel/locking/mutex.c:752 worker_attach_to_pool+0x31/0x390 kernel/workqueue.c:2669 create_worker+0x491/0x720 kernel/workqueue.c:2811 maybe_create_worker kernel/workqueue.c:3054 [inline] manage_workers kernel/workqueue.c:3106 [inline] worker_thread+0x318/0xd30 kernel/workqueue.c:3366 kthread+0x2f0/0x390 kernel/kthread.c:389 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 INFO: task kworker/0:8:5402 blocked for more than 146 seconds. Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:kworker/0:8 state:D stack:30096 pid:5402 tgid:5402 ppid:2 flags:0x00004000 Call Trace: context_switch kernel/sched/core.c:5315 [inline] __schedule+0x1895/0x4b30 kernel/sched/core.c:6674 __schedule_loop kernel/sched/core.c:6751 [inline] schedule+0x14b/0x320 kernel/sched/core.c:6766 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6823 kthread+0x23b/0x390 kernel/kthread.c:382 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 Future hung task reports are suppressed, see sysctl kernel.hung_task_warnings INFO: task kworker/1:7:5403 blocked for more than 146 seconds. Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:kworker/1:7 state:D stack:30096 pid:5403 tgid:5403 ppid:2 flags:0x00004000 Call Trace: context_switch kernel/sched/core.c:5315 [inline] __schedule+0x1895/0x4b30 kernel/sched/core.c:6674 __schedule_loop kernel/sched/core.c:6751 [inline] schedule+0x14b/0x320 kernel/sched/core.c:6766 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6823 kthread+0x23b/0x390 kernel/kthread.c:382 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 Future hung task reports are suppressed, see sysctl kernel.hung_task_warnings Showing all locks held in the system: 1 lock held by kworker/R-rcu_g/4: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline] #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 5 locks held by kworker/0:1/9: 6 locks held by kworker/u8:0/11: #0: ffff88801baed948 ((wq_completion)netns){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3204 [inline] #0: ffff88801baed948 ((wq_completion)netns){+.+.}-{0:0}, at: process_scheduled_works+0x93b/0x1850 kernel/workqueue.c:3310 #1: ffffc90000107d00 (net_cleanup_work){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3205 [inline] #1: ffffc90000107d00 (net_cleanup_work){+.+.}-{0:0}, at: process_scheduled_works+0x976/0x1850 kernel/workqueue.c:3310 #2: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: cleanup_net+0x16a/0xcc0 net/core/net_namespace.c:580 #3: ffff88805e6ac0e8 (&dev->mutex){....}-{3:3}, at: device_lock include/linux/device.h:1014 [inline] #3: ffff88805e6ac0e8 (&dev->mutex){....}-{3:3}, at: devl_dev_lock net/devlink/devl_internal.h:108 [inline] #3: ffff88805e6ac0e8 (&dev->mutex){....}-{3:3}, at: devlink_pernet_pre_exit+0x13b/0x440 net/devlink/core.c:506 #4: ffff88805e6ad250 (&devlink->lock_key#2){+.+.}-{3:3}, at: devl_lock net/devlink/core.c:276 [inline] #4: ffff88805e6ad250 (&devlink->lock_key#2){+.+.}-{3:3}, at: devl_dev_lock net/devlink/devl_internal.h:109 [inline] #4: ffff88805e6ad250 (&devlink->lock_key#2){+.+.}-{3:3}, at: devlink_pernet_pre_exit+0x14d/0x440 net/devlink/core.c:506 #5: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: nsim_destroy+0x71/0x5c0 drivers/net/netdevsim/netdev.c:773 1 lock held by khungtaskd/30: #0: ffffffff8e937ee0 (rcu_read_lock){....}-{1:2}, at: rcu_lock_acquire include/linux/rcupdate.h:337 [inline] #0: ffffffff8e937ee0 (rcu_read_lock){....}-{1:2}, at: rcu_read_lock include/linux/rcupdate.h:849 [inline] #0: ffffffff8e937ee0 (rcu_read_lock){....}-{1:2}, at: debug_show_all_locks+0x55/0x2a0 kernel/locking/lockdep.c:6701 1 lock held by kworker/R-write/32: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_attach_to_pool+0x31/0x390 kernel/workqueue.c:2669 1 lock held by kworker/0:2/937: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_attach_to_pool+0x31/0x390 kernel/workqueue.c:2669 3 locks held by kworker/u8:6/1127: #0: ffff88801ac89148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3204 [inline] #0: ffff88801ac89148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_scheduled_works+0x93b/0x1850 kernel/workqueue.c:3310 #1: ffffc900043b7d00 ((linkwatch_work).work){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3205 [inline] #1: ffffc900043b7d00 ((linkwatch_work).work){+.+.}-{0:0}, at: process_scheduled_works+0x976/0x1850 kernel/workqueue.c:3310 #2: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: linkwatch_event+0xe/0x60 net/core/link_watch.c:276 1 lock held by kworker/u8:7/2466: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_attach_to_pool+0x31/0x390 kernel/workqueue.c:2669 1 lock held by kworker/R-mld/2726: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_attach_to_pool+0x31/0x390 kernel/workqueue.c:2669 1 lock held by kworker/R-bat_e/2861: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_attach_to_pool+0x31/0x390 kernel/workqueue.c:2669 2 locks held by getty/4992: #0: ffff8880323020a0 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x25/0x70 drivers/tty/tty_ldisc.c:243 #1: ffffc900031332f0 (&ldata->atomic_read_lock){+.+.}-{3:3}, at: n_tty_read+0x6a6/0x1e00 drivers/tty/n_tty.c:2211 7 locks held by kworker/0:3/5234: 1 lock held by kworker/R-wg-cr/5268: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline] #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 1 lock held by kworker/R-wg-cr/5271: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline] #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 1 lock held by kworker/R-wg-cr/5272: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline] #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 1 lock held by kworker/R-wg-cr/5275: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline] #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 1 lock held by kworker/R-wg-cr/5277: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline] #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 1 lock held by kworker/R-wg-cr/5281: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline] #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 1 lock held by kworker/R-wg-cr/5283: 1 lock held by kworker/R-wg-cr/5284: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_detach_from_pool kernel/workqueue.c:2727 [inline] #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: rescuer_thread+0xaf5/0x10a0 kernel/workqueue.c:3526 3 locks held by kworker/1:4/5286: #0: ffff88801ac81948 ((wq_completion)events_power_efficient){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3204 [inline] #0: ffff88801ac81948 ((wq_completion)events_power_efficient){+.+.}-{0:0}, at: process_scheduled_works+0x93b/0x1850 kernel/workqueue.c:3310 #1: ffffc90003f67d00 ((crda_timeout).work){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3205 [inline] #1: ffffc90003f67d00 ((crda_timeout).work){+.+.}-{0:0}, at: process_scheduled_works+0x976/0x1850 kernel/workqueue.c:3310 #2: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: crda_timeout_work+0x15/0x50 net/wireless/reg.c:540 1 lock held by kworker/1:6/5291: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: worker_attach_to_pool+0x31/0x390 kernel/workqueue.c:2669 1 lock held by kworker/R-bond0/5422: #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: set_pf_worker kernel/workqueue.c:3316 [inline] #0: ffffffff8e7e24e8 (wq_pool_attach_mutex){+.+.}-{3:3}, at: rescuer_thread+0xd0/0x10a0 kernel/workqueue.c:3443 3 locks held by syz-executor/5423: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: default_device_exit_batch+0xe9/0xaa0 net/core/dev.c:11930 #2: ffffffff8e7d1dd0 (cpu_hotplug_lock){++++}-{0:0}, at: flush_all_backlogs net/core/dev.c:6021 [inline] #2: ffffffff8e7d1dd0 (cpu_hotplug_lock){++++}-{0:0}, at: unregister_netdevice_many_notify+0x5ea/0x1da0 net/core/dev.c:11380 2 locks held by syz-executor/5426: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: tc_action_net_exit include/net/act_api.h:173 [inline] #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: gate_exit_net+0x30/0x100 net/sched/act_gate.c:659 2 locks held by syz-executor/5431: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: setup_net+0x602/0x9e0 net/core/net_namespace.c:378 2 locks held by syz-executor/5432: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: setup_net+0x602/0x9e0 net/core/net_namespace.c:378 2 locks held by syz-executor/5437: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: ip_tunnel_init_net+0x20e/0x720 net/ipv4/ip_tunnel.c:1159 2 locks held by syz-executor/5446: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: wg_netns_pre_exit+0x1f/0x1e0 drivers/net/wireguard/device.c:414 2 locks held by syz-executor/5448: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: ip_tunnel_init_net+0x20e/0x720 net/ipv4/ip_tunnel.c:1159 2 locks held by syz-executor/5454: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: setup_net+0x602/0x9e0 net/core/net_namespace.c:378 2 locks held by syz-executor/5456: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: setup_net+0x602/0x9e0 net/core/net_namespace.c:378 2 locks held by syz-executor/5460: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: ip_tunnel_init_net+0x20e/0x720 net/ipv4/ip_tunnel.c:1159 2 locks held by syz-executor/5471: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: ip_tunnel_init_net+0x20e/0x720 net/ipv4/ip_tunnel.c:1159 2 locks held by syz-executor/5475: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: ip_tunnel_init_net+0x20e/0x720 net/ipv4/ip_tunnel.c:1159 2 locks held by syz-executor/5479: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: ip_tunnel_init_net+0x20e/0x720 net/ipv4/ip_tunnel.c:1159 2 locks held by syz-executor/5481: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: ip_tunnel_init_net+0x20e/0x720 net/ipv4/ip_tunnel.c:1159 2 locks held by syz-executor/5490: #0: ffffffff8fcbf1d0 (pernet_ops_rwsem){++++}-{3:3}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:490 #1: ffffffff8fccbcc8 (rtnl_mutex){+.+.}-{3:3}, at: ip_tunnel_init_net+0x20e/0x720 net/ipv4/ip_tunnel.c:1159 ============================================= NMI backtrace for cpu 1 CPU: 1 UID: 0 PID: 30 Comm: khungtaskd Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 08/06/2024 Call Trace: __dump_stack lib/dump_stack.c:94 [inline] dump_stack_lvl+0x241/0x360 lib/dump_stack.c:120 nmi_cpu_backtrace+0x49c/0x4d0 lib/nmi_backtrace.c:113 nmi_trigger_cpumask_backtrace+0x198/0x320 lib/nmi_backtrace.c:62 trigger_all_cpu_backtrace include/linux/nmi.h:162 [inline] check_hung_uninterruptible_tasks kernel/hung_task.c:223 [inline] watchdog+0xff4/0x1040 kernel/hung_task.c:379 kthread+0x2f0/0x390 kernel/kthread.c:389 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244 Sending NMI from CPU 1 to CPUs 0: NMI backtrace for cpu 0 CPU: 0 UID: 0 PID: 5234 Comm: kworker/0:3 Not tainted 6.11.0-syzkaller-08833-gde5cb0dcb74c #0 Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 08/06/2024 Workqueue: wg-crypt-wg2 wg_packet_encrypt_worker RIP: 0010:__dev_queue_xmit+0x8f/0x3e80 net/core/dev.c:4333 Code: 89 48 8d 94 24 e0 00 00 00 48 c1 ea 03 48 b8 f1 f1 f1 f1 00 f2 f2 f2 48 89 04 1a 48 b9 f2 f2 f2 f2 00 00 04 f2 48 89 4c 1a 0e <48> 05 01 01 01 01 48 89 44 1a 16 66 c7 44 1a 1e 04 f2 48 b8 f2 f2 RSP: 0018:ffffc90000006c00 EFLAGS: 00000a06 RAX: f2f2f200f1f1f1f1 RBX: dffffc0000000000 RCX: f2040000f2f2f2f2 RDX: 1ffff92000000d9c RSI: 0000000000000000 RDI: ffff8880a3fe2640 RBP: ffffc90000006ef0 R08: ffffffff8a1f5d34 R09: 00080aaaaaaaaaaa R10: 0aaaaaaaaaaa0000 R11: 00080aaaaaaaaaaa R12: ffff8880a3fe2718 R13: 000000000000000e R14: ffff8880a3fe2640 R15: 1ffff110147fc4e3 FS: 0000000000000000(0000) GS:ffff8880b8600000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 00007f86267ded58 CR3: 000000000e734000 CR4: 00000000003506f0 DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400 Call Trace: dev_queue_xmit include/linux/netdevice.h:3094 [inline] neigh_hh_output include/net/neighbour.h:526 [inline] neigh_output include/net/neighbour.h:540 [inline] ip_finish_output2+0xd41/0x1390 net/ipv4/ip_output.c:236 synproxy_send_client_synack+0x8b8/0xf30 net/netfilter/nf_synproxy_core.c:484 nft_synproxy_eval_v4+0x3ca/0x610 net/netfilter/nft_synproxy.c:59 nft_synproxy_do_eval+0x362/0xa60 net/netfilter/nft_synproxy.c:141 expr_call_ops_eval net/netfilter/nf_tables_core.c:240 [inline] nft_do_chain+0x4ad/0x1da0 net/netfilter/nf_tables_core.c:288 nft_do_chain_inet+0x418/0x6b0 net/netfilter/nft_chain_filter.c:161 nf_hook_entry_hookfn include/linux/netfilter.h:154 [inline] nf_hook_slow+0xc3/0x220 net/netfilter/core.c:626 nf_hook include/linux/netfilter.h:269 [inline] NF_HOOK+0x29e/0x450 include/linux/netfilter.h:312 NF_HOOK+0x3a4/0x450 include/linux/netfilter.h:314 __netif_receive_skb_one_core net/core/dev.c:5662 [inline] __netif_receive_skb+0x2bf/0x650 net/core/dev.c:5775 process_backlog+0x662/0x15b0 net/core/dev.c:6107 __napi_poll+0xcb/0x490 net/core/dev.c:6771 napi_poll net/core/dev.c:6840 [inline] net_rx_action+0x89b/0x1240 net/core/dev.c:6962 handle_softirqs+0x2c5/0x980 kernel/softirq.c:554 do_softirq+0x11b/0x1e0 kernel/softirq.c:455 __local_bh_enable_ip+0x1bb/0x200 kernel/softirq.c:382 spin_unlock_bh include/linux/spinlock.h:396 [inline] ptr_ring_consume_bh include/linux/ptr_ring.h:367 [inline] wg_packet_encrypt_worker+0x2e8/0x1610 drivers/net/wireguard/send.c:293 process_one_work kernel/workqueue.c:3229 [inline] process_scheduled_works+0xa63/0x1850 kernel/workqueue.c:3310 worker_thread+0x870/0xd30 kernel/workqueue.c:3391 kthread+0x2f0/0x390 kernel/kthread.c:389 ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244