INFO: task syz.0.628:7859 blocked for more than 144 seconds.
Not tainted 6.1.129-syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:syz.0.628 state:D stack:26560 pid:7859 ppid:6076 flags:0x00004004
Call Trace:
context_switch kernel/sched/core.c:5243 [inline]
__schedule+0x143f/0x4570 kernel/sched/core.c:6560
schedule+0xbf/0x180 kernel/sched/core.c:6636
io_schedule+0x88/0x100 kernel/sched/core.c:8788
bit_wait_io+0xe/0xc0 kernel/sched/wait_bit.c:209
__wait_on_bit+0xa8/0x2e0 kernel/sched/wait_bit.c:49
out_of_line_wait_on_bit+0x1d0/0x250 kernel/sched/wait_bit.c:64
wait_on_bit_io include/linux/wait_bit.h:101 [inline]
__wait_on_buffer fs/buffer.c:122 [inline]
wait_on_buffer include/linux/buffer_head.h:385 [inline]
__bread_slow fs/buffer.c:1178 [inline]
__bread_gfp+0x2a8/0x370 fs/buffer.c:1370
sb_bread include/linux/buffer_head.h:338 [inline]
fat_fill_super+0x1c9a/0x4d60 fs/fat/inode.c:1651
mount_bdev+0x2c9/0x3f0 fs/super.c:1443
legacy_get_tree+0xeb/0x180 fs/fs_context.c:632
vfs_get_tree+0x88/0x270 fs/super.c:1573
do_new_mount+0x2ba/0xb40 fs/namespace.c:3056
do_mount fs/namespace.c:3399 [inline]
__do_sys_mount fs/namespace.c:3607 [inline]
__se_sys_mount+0x2d5/0x3c0 fs/namespace.c:3584
do_syscall_x64 arch/x86/entry/common.c:51 [inline]
do_syscall_64+0x3b/0xb0 arch/x86/entry/common.c:81
entry_SYSCALL_64_after_hwframe+0x68/0xd2
RIP: 0033:0x7fad5898d169
RSP: 002b:00007fad5975e038 EFLAGS: 00000246 ORIG_RAX: 00000000000000a5
RAX: ffffffffffffffda RBX: 00007fad58ba6080 RCX: 00007fad5898d169
RDX: 0000400000000040 RSI: 00004000000004c0 RDI: 0000400000000480
RBP: 00007fad58a0e2a0 R08: 0000000000000000 R09: 0000000000000000
R10: 0000000000008205 R11: 0000000000000246 R12: 0000000000000000
R13: 0000000000000000 R14: 00007fad58ba6080 R15: 00007ffef631ab78
Showing all locks held in the system:
1 lock held by rcu_tasks_kthre/12:
#0: ffffffff8d32e850 (rcu_tasks.tasks_gp_mutex){+.+.}-{3:3}, at: rcu_tasks_one_gp+0x29/0xe30 kernel/rcu/tasks.h:517
1 lock held by rcu_tasks_trace/13:
#0: ffffffff8d32f050 (rcu_tasks_trace.tasks_gp_mutex){+.+.}-{3:3}, at: rcu_tasks_one_gp+0x29/0xe30 kernel/rcu/tasks.h:517
1 lock held by khungtaskd/28:
#0: ffffffff8d32e680 (rcu_read_lock){....}-{1:2}, at: rcu_lock_acquire include/linux/rcupdate.h:350 [inline]
#0: ffffffff8d32e680 (rcu_read_lock){....}-{1:2}, at: rcu_read_lock include/linux/rcupdate.h:791 [inline]
#0: ffffffff8d32e680 (rcu_read_lock){....}-{1:2}, at: debug_show_all_locks+0x51/0x290 kernel/locking/lockdep.c:6510
5 locks held by kworker/u4:2/41:
#0: ffff888017e16938 ((wq_completion)netns){+.+.}-{0:0}, at: process_one_work+0x806/0x1260 kernel/workqueue.c:2267
#1: ffffc90000b27d20 (net_cleanup_work){+.+.}-{0:0}, at: process_one_work+0x806/0x1260 kernel/workqueue.c:2267
#2: ffffffff8e542f90 (pernet_ops_rwsem){++++}-{3:3}, at: cleanup_net+0x166/0xd20 net/core/net_namespace.c:594
#3: ffffffff8e54faa8 (rtnl_mutex){+.+.}-{3:3}, at: default_device_exit_batch+0xe5/0xa20 net/core/dev.c:11446
#4: ffffffff8d333cb8 (rcu_state.exp_mutex){+.+.}-{3:3}, at: exp_funnel_lock kernel/rcu/tree_exp.h:291 [inline]
#4: ffffffff8d333cb8 (rcu_state.exp_mutex){+.+.}-{3:3}, at: synchronize_rcu_expedited+0x4f0/0x930 kernel/rcu/tree_exp.h:962
1 lock held by dhcpcd/3917:
#0: ffffffff8e54faa8 (rtnl_mutex){+.+.}-{3:3}, at: devinet_ioctl+0x29a/0x1ae0 net/ipv4/devinet.c:1081
2 locks held by getty/4006:
#0: ffff888030c3e098 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x21/0x70 drivers/tty/tty_ldisc.c:244
#1: ffffc9000325e2f0 (&ldata->atomic_read_lock){+.+.}-{3:3}, at: n_tty_read+0x54a/0x1620 drivers/tty/n_tty.c:2198
2 locks held by kworker/1:17/4539:
#0: ffff888017c72138 ((wq_completion)rcu_gp){+.+.}-{0:0}, at: process_one_work+0x806/0x1260 kernel/workqueue.c:2267
#1: ffffc90003557d20 ((work_completion)(&rew->rew_work)){+.+.}-{0:0}, at: process_one_work+0x806/0x1260 kernel/workqueue.c:2267
3 locks held by kworker/u4:13/4925:
#0: ffff888017c79138 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_one_work+0x806/0x1260 kernel/workqueue.c:2267
#1: ffffc90003507d20 ((linkwatch_work).work){+.+.}-{0:0}, at: process_one_work+0x806/0x1260 kernel/workqueue.c:2267
#2: ffffffff8e54faa8 (rtnl_mutex){+.+.}-{3:3}, at: linkwatch_event+0xa/0x50 net/core/link_watch.c:263
3 locks held by kworker/0:19/5930:
#0: ffff888017c70938 ((wq_completion)events){+.+.}-{0:0}, at: process_one_work+0x806/0x1260 kernel/workqueue.c:2267
#1: ffffc90004a3fd20 (deferred_process_work){+.+.}-{0:0}, at: process_one_work+0x806/0x1260 kernel/workqueue.c:2267
#2: ffffffff8e54faa8 (rtnl_mutex){+.+.}-{3:3}, at: switchdev_deferred_process_work+0xa/0x20 net/switchdev/switchdev.c:104
1 lock held by syz.0.628/7859:
#0: ffff8880283580e0 (&type->s_umount_key#87/1){+.+.}-{3:3}, at: alloc_super+0x217/0x930 fs/super.c:228
1 lock held by syz-executor/9569:
#0: ffffffff8e54faa8 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock net/core/rtnetlink.c:74 [inline]
#0: ffffffff8e54faa8 (rtnl_mutex){+.+.}-{3:3}, at: rtnetlink_rcv_msg+0x7cb/0x1000 net/core/rtnetlink.c:6147
3 locks held by syz.9.954/9592:
#0: ffffffff8e5b09f0 (cb_lock){++++}-{3:3}, at: genl_rcv+0x15/0x40 net/netlink/genetlink.c:860
#1: ffffffff8e5b08a8 (genl_mutex){+.+.}-{3:3}, at: genl_lock net/netlink/genetlink.c:33 [inline]
#1: ffffffff8e5b08a8 (genl_mutex){+.+.}-{3:3}, at: genl_rcv_msg+0x125/0xf50 net/netlink/genetlink.c:848
#2: ffffffff8e54faa8 (rtnl_mutex){+.+.}-{3:3}, at: __tipc_nl_compat_doit net/tipc/netlink_compat.c:362 [inline]
#2: ffffffff8e54faa8 (rtnl_mutex){+.+.}-{3:3}, at: tipc_nl_compat_doit+0x215/0x600 net/tipc/netlink_compat.c:397
=============================================
NMI backtrace for cpu 1
CPU: 1 PID: 28 Comm: khungtaskd Not tainted 6.1.129-syzkaller #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 02/12/2025
Call Trace:
__dump_stack lib/dump_stack.c:88 [inline]
dump_stack_lvl+0x1e3/0x2cb lib/dump_stack.c:106
nmi_cpu_backtrace+0x4e1/0x560 lib/nmi_backtrace.c:111
nmi_trigger_cpumask_backtrace+0x1ca/0x430 lib/nmi_backtrace.c:62
trigger_all_cpu_backtrace include/linux/nmi.h:148 [inline]
check_hung_uninterruptible_tasks kernel/hung_task.c:220 [inline]
watchdog+0xf88/0xfd0 kernel/hung_task.c:377
kthread+0x28d/0x320 kernel/kthread.c:376
ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:295
Sending NMI from CPU 1 to CPUs 0:
NMI backtrace for cpu 0 skipped: idling at native_safe_halt arch/x86/include/asm/irqflags.h:51 [inline]
NMI backtrace for cpu 0 skipped: idling at arch_safe_halt arch/x86/include/asm/irqflags.h:89 [inline]
NMI backtrace for cpu 0 skipped: idling at acpi_safe_halt drivers/acpi/processor_idle.c:111 [inline]
NMI backtrace for cpu 0 skipped: idling at acpi_idle_do_entry+0x10f/0x340 drivers/acpi/processor_idle.c:567