INFO: task kworker/u4:0:9 blocked for more than 143 seconds.
Not tainted syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/u4:0 state:D stack:21304 pid:9 ppid:2 flags:0x00004000
Workqueue: events_unbound linkwatch_event
Call Trace:
<TASK>
context_switch kernel/sched/core.c:5245 [inline]
__schedule+0x11d1/0x40e0 kernel/sched/core.c:6562
schedule+0xb9/0x180 kernel/sched/core.c:6638
schedule_preempt_disabled+0xf/0x20 kernel/sched/core.c:6697
__mutex_lock_common kernel/locking/mutex.c:679 [inline]
__mutex_lock+0x562/0xaf0 kernel/locking/mutex.c:747
linkwatch_event+0xa/0x50 net/core/link_watch.c:263
process_one_work+0x8a2/0x1160 kernel/workqueue.c:2292
worker_thread+0xaa2/0x1270 kernel/workqueue.c:2439
kthread+0x29d/0x330 kernel/kthread.c:376
ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:295
</TASK>
INFO: task kworker/1:11:26665 blocked for more than 143 seconds.
Not tainted syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/1:11 state:D stack:24976 pid:26665 ppid:2 flags:0x00004000
Workqueue: ipv6_addrconf addrconf_verify_work
Call Trace:
<TASK>
context_switch kernel/sched/core.c:5245 [inline]
__schedule+0x11d1/0x40e0 kernel/sched/core.c:6562
schedule+0xb9/0x180 kernel/sched/core.c:6638
schedule_preempt_disabled+0xf/0x20 kernel/sched/core.c:6697
__mutex_lock_common kernel/locking/mutex.c:679 [inline]
__mutex_lock+0x562/0xaf0 kernel/locking/mutex.c:747
addrconf_verify_work+0x15/0x30 net/ipv6/addrconf.c:4672
process_one_work+0x8a2/0x1160 kernel/workqueue.c:2292
worker_thread+0xaa2/0x1270 kernel/workqueue.c:2439
kthread+0x29d/0x330 kernel/kthread.c:376
ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:295
</TASK>
INFO: task kworker/1:36:26719 blocked for more than 144 seconds.
Not tainted syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/1:36 state:D stack:21872 pid:26719 ppid:2 flags:0x00004000
Workqueue: events switchdev_deferred_process_work
Call Trace:
<TASK>
context_switch kernel/sched/core.c:5245 [inline]
__schedule+0x11d1/0x40e0 kernel/sched/core.c:6562
schedule+0xb9/0x180 kernel/sched/core.c:6638
schedule_preempt_disabled+0xf/0x20 kernel/sched/core.c:6697
__mutex_lock_common kernel/locking/mutex.c:679 [inline]
__mutex_lock+0x562/0xaf0 kernel/locking/mutex.c:747
switchdev_deferred_process_work+0xa/0x20 net/switchdev/switchdev.c:104
process_one_work+0x8a2/0x1160 kernel/workqueue.c:2292
worker_thread+0xaa2/0x1270 kernel/workqueue.c:2439
kthread+0x29d/0x330 kernel/kthread.c:376
ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:295
</TASK>
Showing all locks held in the system:
3 locks held by kworker/u4:0/9:
#0: ffff888017479138 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
#1: ffffc900000e7d00 ((linkwatch_work).work){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
#2: ffffffff8dd47de8 (rtnl_mutex){+.+.}-{3:3}, at: linkwatch_event+0xa/0x50 net/core/link_watch.c:263
5 locks held by kworker/u4:1/11:
#0: ffff888017616938 ((wq_completion)netns){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
#1: ffffc90000107d00 (net_cleanup_work){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
#2: ffffffff8dd3af10 (pernet_ops_rwsem){++++}-{3:3}, at: cleanup_net+0x148/0xba0 net/core/net_namespace.c:594
#3: ffffffff8dd47de8 (rtnl_mutex){+.+.}-{3:3}, at: default_device_exit_batch+0xee/0xa80 net/core/dev.c:11535
#4: ffffffff8cb33278 (rcu_state.exp_mutex){+.+.}-{3:3}, at: exp_funnel_lock kernel/rcu/tree_exp.h:291 [inline]
#4: ffffffff8cb33278 (rcu_state.exp_mutex){+.+.}-{3:3}, at: synchronize_rcu_expedited+0x2ec/0x890 kernel/rcu/tree_exp.h:962
1 lock held by rcu_tasks_kthre/12:
#0: ffffffff8cb2df30 (rcu_tasks.tasks_gp_mutex){+.+.}-{3:3}, at: rcu_tasks_one_gp+0x33/0xf00 kernel/rcu/tasks.h:517
1 lock held by rcu_tasks_trace/13:
#0: ffffffff8cb2e750 (rcu_tasks_trace.tasks_gp_mutex){+.+.}-{3:3}, at: rcu_tasks_one_gp+0x33/0xf00 kernel/rcu/tasks.h:517
4 locks held by ksoftirqd/1/21:
1 lock held by khungtaskd/28:
#0: ffffffff8cb2d5a0 (rcu_read_lock){....}-{1:2}, at: rcu_lock_acquire include/linux/rcupdate.h:350 [inline]
#0: ffffffff8cb2d5a0 (rcu_read_lock){....}-{1:2}, at: rcu_read_lock include/linux/rcupdate.h:791 [inline]
#0: ffffffff8cb2d5a0 (rcu_read_lock){....}-{1:2}, at: debug_show_all_locks+0x51/0x290 kernel/locking/lockdep.c:6513
2 locks held by getty/4027:
#0: ffff888030519098 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x21/0x70 drivers/tty/tty_ldisc.c:244
#1: ffffc9000327b2f0 (&ldata->atomic_read_lock){+.+.}-{3:3}, at: n_tty_read+0x429/0x1390 drivers/tty/n_tty.c:2198
2 locks held by kworker/u4:9/11349:
3 locks held by kworker/0:20/21964:
#0: ffff88814cd5f938 ((wq_completion)ipv6_addrconf){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
#1: ffffc9000507fd00 ((work_completion)(&(&net->ipv6.addr_chk_work)->work)){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
#2: ffffffff8dd47de8 (rtnl_mutex){+.+.}-{3:3}, at: addrconf_verify_work+0x15/0x30 net/ipv6/addrconf.c:4672
2 locks held by kworker/0:23/21968:
#0: ffff888017472138 ((wq_completion)rcu_gp){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
#1: ffffc90003447d00 ((work_completion)(&rew->rew_work)){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
2 locks held by kworker/u4:11/25736:
3 locks held by kworker/1:11/26665:
#0: ffff88814cd5f938 ((wq_completion)ipv6_addrconf){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
#1: ffffc90004f3fd00 ((work_completion)(&(&net->ipv6.addr_chk_work)->work)){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
#2: ffffffff8dd47de8 (rtnl_mutex){+.+.}-{3:3}, at: addrconf_verify_work+0x15/0x30 net/ipv6/addrconf.c:4672
3 locks held by kworker/1:36/26719:
#0: ffff888017470938 ((wq_completion)events){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
#1: ffffc90004ce7d00 (deferred_process_work){+.+.}-{0:0}, at: process_one_work+0x7b0/0x1160 kernel/workqueue.c:2267
#2: ffffffff8dd47de8 (rtnl_mutex){+.+.}-{3:3}, at: switchdev_deferred_process_work+0xa/0x20 net/switchdev/switchdev.c:104
=============================================
NMI backtrace for cpu 1
CPU: 1 PID: 28 Comm: khungtaskd Not tainted syzkaller #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 03/18/2026
Call Trace:
<TASK>
dump_stack_lvl+0x188/0x24e lib/dump_stack.c:106
nmi_cpu_backtrace+0x3e6/0x460 lib/nmi_backtrace.c:111
nmi_trigger_cpumask_backtrace+0x1d4/0x450 lib/nmi_backtrace.c:62
trigger_all_cpu_backtrace include/linux/nmi.h:148 [inline]
check_hung_uninterruptible_tasks kernel/hung_task.c:220 [inline]
watchdog+0xeee/0xf30 kernel/hung_task.c:377
kthread+0x29d/0x330 kernel/kthread.c:376
ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:295
</TASK>
Sending NMI from CPU 1 to CPUs 0:
NMI backtrace for cpu 0
CPU: 0 PID: 16481 Comm: kworker/u4:10 Not tainted syzkaller #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 03/18/2026
Workqueue: bat_events batadv_nc_worker
RIP: 0010:do_csum lib/checksum.c:78 [inline]
RIP: 0010:csum_partial+0x145/0x2c0 lib/checksum.c:128
Code: c7 04 31 db 4c 89 e8 48 c1 e8 03 0f b6 04 28 84 c0 75 2b 41 8b 7f fc 41 01 dc 31 db 41 01 fc 0f 92 c3 44 89 e6 e8 cb 9e 4c fd <4d> 39 f7 73 2d e8 61 9d 4c fd 49 83 c5 04 49 83 c7 04 eb c6 44 89
RSP: 0018:ffffc90000007218 EFLAGS: 00000246
RAX: ffffffff8435d7e5 RBX: 0000000000000000 RCX: 0000000000000100
RDX: ffff88802a5a1dc0 RSI: 0000000046746b29 RDI: 000000001baaaaff
RBP: dffffc0000000000 R08: ffffffff8b89d760 R09: 0000000000000002
R10: 0000000000000002 R11: 0000000000000100 R12: 0000000046746b29
R13: ffff888054645058 R14: ffff888054645070 R15: ffff88805464505c
FS: 0000000000000000(0000) GS:ffff8880b8e00000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 0000556be669f168 CR3: 000000000c88e000 CR4: 00000000003506f0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
Call Trace:
<IRQ>
__skb_checksum+0xf3/0x7a0 net/core/skbuff.c:3020
skb_checksum net/core/skbuff.c:3100 [inline]
__skb_checksum_complete+0x4e/0x3f0 net/core/skbuff.c:3222
nf_conntrack_icmpv6_error+0x1ca/0x5d0 net/netfilter/nf_conntrack_proto_icmpv6.c:200
nf_conntrack_handle_icmp+0xd2/0x17a net/netfilter/nf_conntrack_core.c:1925
nf_conntrack_in+0x1347/0x17b0 net/netfilter/nf_conntrack_core.c:2020
nf_hook_entry_hookfn include/linux/netfilter.h:142 [inline]
nf_hook_slow+0xb9/0x200 net/netfilter/core.c:614
nf_hook include/linux/netfilter.h:257 [inline]
NF_HOOK include/linux/netfilter.h:300 [inline]
br_nf_pre_routing_ipv6+0x55f/0x6b0 net/bridge/br_netfilter_ipv6.c:243
nf_hook_entry_hookfn include/linux/netfilter.h:142 [inline]
nf_hook_bridge_pre net/bridge/br_input.c:260 [inline]
br_handle_frame+0x1167/0x13c0 net/bridge/br_input.c:406
__netif_receive_skb_core+0x1004/0x38f0 net/core/dev.c:5513
__netif_receive_skb_one_core net/core/dev.c:5617 [inline]
__netif_receive_skb+0x74/0x290 net/core/dev.c:5733
process_backlog+0x38d/0x6f0 net/core/dev.c:6061
__napi_poll+0xc0/0x460 net/core/dev.c:6628
napi_poll net/core/dev.c:6695 [inline]
net_rx_action+0x5dd/0xb20 net/core/dev.c:6809
handle_softirqs+0x2a1/0x930 kernel/softirq.c:596
do_softirq+0x142/0x210 kernel/softirq.c:497
</IRQ>
<TASK>
__local_bh_enable_ip+0x180/0x1c0 kernel/softirq.c:421
spin_unlock_bh include/linux/spinlock.h:396 [inline]
batadv_nc_purge_paths+0x316/0x3b0 net/batman-adv/network-coding.c:471
batadv_nc_worker+0x324/0x600 net/batman-adv/network-coding.c:720
process_one_work+0x8a2/0x1160 kernel/workqueue.c:2292
worker_thread+0xaa2/0x1270 kernel/workqueue.c:2439
kthread+0x29d/0x330 kernel/kthread.c:376
ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:295
</TASK>
net_ratelimit: 11414 callbacks suppressed
bridge0: received packet on veth1_to_bridge with own address as source address (addr:ea:73:15:8f:20:c8, vlan:0)
bridge0: received packet on veth1_to_bridge with own address as source address (addr:ea:73:15:8f:20:c8, vlan:0)
bridge0: received packet on veth1_to_bridge with own address as source address (addr:aa:aa:aa:aa:aa:0c, vlan:0)
bridge0: received packet on veth1_to_bridge with own address as source address (addr:ea:73:15:8f:20:c8, vlan:0)
bridge0: received packet on veth1_to_bridge with own address as source address (addr:aa:aa:aa:aa:aa:0c, vlan:0)
bridge0: received packet on veth1_to_bridge with own address as source address (addr:ea:73:15:8f:20:c8, vlan:0)
bridge0: received packet on veth1_to_bridge with own address as source address (addr:aa:aa:aa:aa:aa:0c, vlan:0)
bridge0: received packet on veth1_to_bridge with own address as source address (addr:ea:73:15:8f:20:c8, vlan:0)
bridge0: received packet on bridge_slave_1 with own address as source address (addr:aa:aa:aa:aa:aa:0c, vlan:0)
bridge0: received packet on veth1_to_bridge with own address as source address (addr:ea:73:15:8f:20:c8, vlan:0)