INFO: task syz.0.2181:15467 blocked for more than 143 seconds.
Not tainted 6.15.0-syzkaller-11173-g546b1c9e93c2 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:syz.0.2181 state:D stack:28888 pid:15467 tgid:15466 ppid:5812 task_flags:0x400140 flags:0x00004004
Call Trace:
context_switch kernel/sched/core.c:5396 [inline]
__schedule+0x116a/0x5de0 kernel/sched/core.c:6785
__schedule_loop kernel/sched/core.c:6863 [inline]
schedule+0xe7/0x3a0 kernel/sched/core.c:6878
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6935
__mutex_lock_common kernel/locking/mutex.c:679 [inline]
__mutex_lock+0x6c7/0xb90 kernel/locking/mutex.c:747
bdev_open+0x41a/0xe40 block/bdev.c:945
blkdev_open+0x280/0x3f0 block/fops.c:676
do_dentry_open+0x741/0x1c10 fs/open.c:964
vfs_open+0x82/0x3f0 fs/open.c:1094
do_open fs/namei.c:3887 [inline]
path_openat+0x1de4/0x2cb0 fs/namei.c:4046
do_filp_open+0x20b/0x470 fs/namei.c:4073
do_sys_openat2+0x11b/0x1d0 fs/open.c:1437
do_sys_open fs/open.c:1452 [inline]
__do_sys_openat fs/open.c:1468 [inline]
__se_sys_openat fs/open.c:1463 [inline]
__x64_sys_openat+0x174/0x210 fs/open.c:1463
do_syscall_x64 arch/x86/entry/syscall_64.c:63 [inline]
do_syscall_64+0xcd/0x4c0 arch/x86/entry/syscall_64.c:94
entry_SYSCALL_64_after_hwframe+0x77/0x7f
RIP: 0033:0x7fabcfd8d2d0
RSP: 002b:00007fabd0c58b70 EFLAGS: 00000293 ORIG_RAX: 0000000000000101
RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fabcfd8d2d0
RDX: 0000000000000000 RSI: 00007fabd0c58c10 RDI: 00000000ffffff9c
RBP: 00007fabd0c58c10 R08: 0000000000000000 R09: 002364626e2f7665
R10: 0000000000000000 R11: 0000000000000293 R12: 0000000000000000
R13: 0000000000000000 R14: 00007fabcffb5fa0 R15: 00007ffd7d4bc678
Showing all locks held in the system:
3 locks held by kworker/u8:0/12:
#0: ffff88801b881148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_one_work+0x12a2/0x1b70 kernel/workqueue.c:3213
#1: ffffc90000117d10 ((linkwatch_work).work){+.+.}-{0:0}, at: process_one_work+0x929/0x1b70 kernel/workqueue.c:3214
#2: ffffffff90347d68 (rtnl_mutex){+.+.}-{4:4}, at: linkwatch_event+0x51/0xc0 net/core/link_watch.c:303
3 locks held by kworker/u8:1/13:
#0: ffff8880b843bc98 (&rq->__lock){-.-.}-{2:2}, at: raw_spin_rq_lock_nested kernel/sched/core.c:614 [inline]
#0: ffff8880b843bc98 (&rq->__lock){-.-.}-{2:2}, at: raw_spin_rq_lock_nested+0x7e/0x130 kernel/sched/core.c:599
#1: ffff8880b8524088 (&per_cpu_ptr(group->pcpu, cpu)->seq){-.-.}-{0:0}, at: psi_task_switch+0x2c1/0x8e0 kernel/sched/psi.c:987
#2: ffff8880b8525b18 (&base->lock){-.-.}-{2:2}, at: lock_timer_base+0x127/0x1d0 kernel/time/timer.c:1004
2 locks held by ksoftirqd/0/15:
1 lock held by khungtaskd/31:
#0: ffffffff8e5c4e40 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire include/linux/rcupdate.h:331 [inline]
#0: ffffffff8e5c4e40 (rcu_read_lock){....}-{1:3}, at: rcu_read_lock include/linux/rcupdate.h:841 [inline]
#0: ffffffff8e5c4e40 (rcu_read_lock){....}-{1:3}, at: debug_show_all_locks+0x36/0x1c0 kernel/locking/lockdep.c:6770
2 locks held by dhcpcd/5483:
#0: ffffffff9031c228 (vlan_ioctl_mutex){+.+.}-{4:4}, at: sock_ioctl+0x49c/0x6b0 net/socket.c:1273
#1: ffffffff90347d68 (rtnl_mutex){+.+.}-{4:4}, at: vlan_ioctl_handler+0xce/0xa70 net/8021q/vlan.c:553
2 locks held by getty/5582:
#0: ffff8880374220a0 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x24/0x80 drivers/tty/tty_ldisc.c:243
#1: ffffc9000332b2f0 (&ldata->atomic_read_lock){+.+.}-{4:4}, at: n_tty_read+0x41b/0x14f0 drivers/tty/n_tty.c:2222
3 locks held by kworker/u8:9/5916:
#0: ffff8880323bd148 ((wq_completion)ipv6_addrconf){+.+.}-{0:0}, at: process_one_work+0x12a2/0x1b70 kernel/workqueue.c:3213
#1: ffffc9000490fd10 ((work_completion)(&(&net->ipv6.addr_chk_work)->work)){+.+.}-{0:0}, at: process_one_work+0x929/0x1b70 kernel/workqueue.c:3214
#2: ffffffff90347d68 (rtnl_mutex){+.+.}-{4:4}, at: rtnl_net_lock include/linux/rtnetlink.h:130 [inline]
#2: ffffffff90347d68 (rtnl_mutex){+.+.}-{4:4}, at: addrconf_verify_work+0x12/0x30 net/ipv6/addrconf.c:4738
3 locks held by kworker/u8:13/7364:
#0: ffff8880b843bc98 (&rq->__lock){-.-.}-{2:2}, at: raw_spin_rq_lock_nested kernel/sched/core.c:614 [inline]
#0: ffff8880b843bc98 (&rq->__lock){-.-.}-{2:2}, at: raw_spin_rq_lock_nested+0x7e/0x130 kernel/sched/core.c:599
#1: ffff8880b8424088 (&per_cpu_ptr(group->pcpu, cpu)->seq){-.-.}-{0:0}, at: psi_task_switch+0x2c1/0x8e0 kernel/sched/psi.c:987
#2: ffff8880b8425b18 (&base->lock){-.-.}-{2:2}, at: lock_timer_base+0x127/0x1d0 kernel/time/timer.c:1004
1 lock held by udevd/12365:
#0: ffff888142705358 (&disk->open_mutex){+.+.}-{4:4}, at: bdev_open+0x41a/0xe40 block/bdev.c:945
1 lock held by syz.0.2181/15467:
#0: ffff888142705358 (&disk->open_mutex){+.+.}-{4:4}, at: bdev_open+0x41a/0xe40 block/bdev.c:945
8 locks held by syz-executor/17321:
#0: ffff8880377d2428 (sb_writers#8){.+.+}-{0:0}
, at: ksys_write+0x12a/0x250 fs/read_write.c:738
#1: ffff8880331a0488 (&of->mutex){+.+.}-{4:4}, at: kernfs_fop_write_iter+0x28f/0x510 fs/kernfs/file.c:325
#2: ffff888142735b48 (kn->active#52){.+.+}-{0:0}, at: kernfs_fop_write_iter+0x2b2/0x510 fs/kernfs/file.c:326
#3: ffffffff8f8e7208 (nsim_bus_dev_list_lock){+.+.}-{4:4}, at: del_device_store+0xd1/0x4a0 drivers/net/netdevsim/bus.c:216
#4: ffff8880571b50e8 (&dev->mutex){....}-{4:4}, at: device_lock include/linux/device.h:884 [inline]
#4: ffff8880571b50e8 (&dev->mutex){....}-{4:4}, at: __device_driver_lock drivers/base/dd.c:1094 [inline]
#4: ffff8880571b50e8 (&dev->mutex){....}-{4:4}, at: device_release_driver_internal+0xa4/0x620 drivers/base/dd.c:1292
#5: ffff888057768250 (&devlink->lock_key#22){+.+.}-{4:4}, at: nsim_drv_remove+0x4a/0x1d0 drivers/net/netdevsim/dev.c:1675
#6: ffffffff90347d68 (rtnl_mutex){+.+.}-{4:4}, at: nsim_destroy+0xfc/0x5d0 drivers/net/netdevsim/netdev.c:1063
#7: ffffffff8e5d03f8 (rcu_state.exp_mutex){+.+.}-{4:4}, at: exp_funnel_lock+0x1a3/0x3c0 kernel/rcu/tree_exp.h:336
2 locks held by syz.1.2494/17522:
#0: ffffffff90331d10 (pernet_ops_rwsem){++++}-{4:4}, at: copy_net_ns+0x286/0x5f0 net/core/net_namespace.c:570
#1: ffffffff90347d68 (rtnl_mutex){+.+.}-{4:4}, at: rtnl_net_lock include/linux/rtnetlink.h:130 [inline]
#1: ffffffff90347d68 (rtnl_mutex){+.+.}-{4:4}, at: register_netdevice_notifier_net+0x23/0xb0 net/core/dev.c:2083
2 locks held by vhost-17522/17523:
#0: ffff8880259750e8 (&vtsk->exit_mutex){+.+.}-{4:4}, at: vhost_task_fn+0x2e9/0x460 kernel/vhost_task.c:54
#1: ffff88805ab50f78 (&worker->mutex){+.+.}-{4:4}, at: vhost_worker_killed+0x55/0x3d0 drivers/vhost/vhost.c:423
3 locks held by syz.5.2495/17527:
#0: ffffffff903ec510 (cb_lock){++++}-{4:4}, at: genl_rcv+0x19/0x40 net/netlink/genetlink.c:1218
#1: ffff8880379f96d0 (nlk_cb_mutex-GENERIC){+.+.}-{4:4}, at: __netlink_dump_start+0x150/0x990 net/netlink/af_netlink.c:2388
#2: ffffffff90347d68 (rtnl_mutex){+.+.}-{4:4}, at: ethnl_default_dump_one+0x178/0xa20 net/ethtool/netlink.c:576
=============================================
NMI backtrace for cpu 0
CPU: 0 UID: 0 PID: 31 Comm: khungtaskd Not tainted 6.15.0-syzkaller-11173-g546b1c9e93c2 #0 PREEMPT(full)
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 05/07/2025
Call Trace:
__dump_stack lib/dump_stack.c:94 [inline]
dump_stack_lvl+0x116/0x1f0 lib/dump_stack.c:120
nmi_cpu_backtrace+0x27b/0x390 lib/nmi_backtrace.c:113
nmi_trigger_cpumask_backtrace+0x29c/0x300 lib/nmi_backtrace.c:62
trigger_all_cpu_backtrace include/linux/nmi.h:158 [inline]
check_hung_uninterruptible_tasks kernel/hung_task.c:307 [inline]
watchdog+0xf70/0x12c0 kernel/hung_task.c:470
kthread+0x3c5/0x780 kernel/kthread.c:464
ret_from_fork+0x5d4/0x6f0 arch/x86/kernel/process.c:148
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:245
Sending NMI from CPU 0 to CPUs 1:
NMI backtrace for cpu 1
CPU: 1 UID: 0 PID: 24 Comm: kworker/1:0 Not tainted 6.15.0-syzkaller-11173-g546b1c9e93c2 #0 PREEMPT(full)
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 05/07/2025
Workqueue: events drain_vmap_area_work
RIP: 0010:iterate_chain_key kernel/locking/lockdep.c:456 [inline]
RIP: 0010:__lock_acquire+0x689/0x1c90 kernel/locking/lockdep.c:5228
Code: 44 29 f1 41 01 c6 c1 c2 08 31 d1 89 ca 29 c8 44 01 f1 c1 c2 10 89 cf 31 d0 89 c6 41 29 c6 01 c8 c1 ce 0d 44 31 f6 89 f2 29 f7 <01> f0 c1 c2 04 89 c0 31 fa 48 c1 e2 20 48 09 c2 41 8b 45 20 49 89
RSP: 0018:ffffc900001e79a8 EFLAGS: 00000003
RAX: 00000000111c7254 RBX: 0000000000000003 RCX: 00000000199044fb
RDX: 00000000b879bb7b RSI: 00000000b879bb7b RDI: 0000000061168980
RBP: ffff88801e6d8000 R08: 0000000000000001 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000001 R12: ffff88801e6d8af0
R13: ffff88801e6d8b68 R14: 00000000d2b6071a R15: 0000000000000003
FS: 0000000000000000(0000) GS:ffff888124865000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 00007f5c7f689a10 CR3: 000000000e382000 CR4: 00000000003526f0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000009 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Call Trace:
lock_acquire kernel/locking/lockdep.c:5871 [inline]
lock_acquire+0x179/0x350 kernel/locking/lockdep.c:5828
__raw_spin_lock include/linux/spinlock_api_smp.h:133 [inline]
_raw_spin_lock+0x2e/0x40 kernel/locking/spinlock.c:154
spin_lock include/linux/spinlock.h:351 [inline]
node_pool_add_va mm/vmalloc.c:1909 [inline]
purge_vmap_node+0x5c0/0xa30 mm/vmalloc.c:2272
__purge_vmap_area_lazy+0xa06/0xc60 mm/vmalloc.c:2348
drain_vmap_area_work+0x27/0x40 mm/vmalloc.c:2382
process_one_work+0x9cf/0x1b70 kernel/workqueue.c:3238
process_scheduled_works kernel/workqueue.c:3321 [inline]
worker_thread+0x6c8/0xf10 kernel/workqueue.c:3402
kthread+0x3c5/0x780 kernel/kthread.c:464
ret_from_fork+0x5d4/0x6f0 arch/x86/kernel/process.c:148
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:245
vkms_vblank_simulate: vblank timer overrun