INFO: task udevd:7640 blocked for more than 143 seconds.
Not tainted 6.13.0-rc3-next-20241220-syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:udevd state:D stack:21408 pid:7640 tgid:7640 ppid:5206 flags:0x00004002
Call Trace:
context_switch kernel/sched/core.c:5371 [inline]
__schedule+0x189f/0x4c80 kernel/sched/core.c:6758
__schedule_loop kernel/sched/core.c:6835 [inline]
schedule+0x14b/0x320 kernel/sched/core.c:6850
schedule_timeout+0x15a/0x290 kernel/time/sleep_timeout.c:99
wait_for_reconnect drivers/block/nbd.c:1043 [inline]
nbd_handle_cmd drivers/block/nbd.c:1085 [inline]
nbd_queue_rq+0x7d9/0x2ef0 drivers/block/nbd.c:1143
blk_mq_dispatch_rq_list+0xad3/0x19d0 block/blk-mq.c:2120
__blk_mq_do_dispatch_sched block/blk-mq-sched.c:170 [inline]
blk_mq_do_dispatch_sched block/blk-mq-sched.c:184 [inline]
__blk_mq_sched_dispatch_requests+0xb8a/0x1840 block/blk-mq-sched.c:309
blk_mq_sched_dispatch_requests+0xd6/0x190 block/blk-mq-sched.c:331
blk_mq_run_hw_queue+0x354/0x500 block/blk-mq.c:2354
blk_mq_flush_plug_list+0x118e/0x1870 block/blk-mq.c:2915
__blk_flush_plug+0x420/0x500 block/blk-core.c:1213
blk_finish_plug block/blk-core.c:1240 [inline]
__submit_bio+0x46a/0x560 block/blk-core.c:637
__submit_bio_noacct_mq block/blk-core.c:710 [inline]
submit_bio_noacct_nocheck+0x4d3/0xe30 block/blk-core.c:739
submit_bh fs/buffer.c:2819 [inline]
block_read_full_folio+0x9b3/0xae0 fs/buffer.c:2446
filemap_read_folio+0x148/0x3b0 mm/filemap.c:2348
do_read_cache_folio+0x373/0x5b0 mm/filemap.c:3893
read_mapping_folio include/linux/pagemap.h:1032 [inline]
read_part_sector+0xb3/0x260 block/partitions/core.c:722
adfspart_check_ICS+0xd9/0x9a0 block/partitions/acorn.c:360
check_partition block/partitions/core.c:141 [inline]
blk_add_partitions block/partitions/core.c:589 [inline]
bdev_disk_changed+0x72c/0x13f0 block/partitions/core.c:693
blkdev_get_whole+0x2d2/0x450 block/bdev.c:707
bdev_open+0x2d4/0xc50 block/bdev.c:916
blkdev_open+0x389/0x4f0 block/fops.c:652
do_dentry_open+0xdec/0x1960 fs/open.c:955
vfs_open+0x3b/0x370 fs/open.c:1085
do_open fs/namei.c:3828 [inline]
path_openat+0x2c74/0x3580 fs/namei.c:3987
do_filp_open+0x27f/0x4e0 fs/namei.c:4014
do_sys_openat2+0x13e/0x1d0 fs/open.c:1427
do_sys_open fs/open.c:1442 [inline]
__do_sys_openat fs/open.c:1458 [inline]
__se_sys_openat fs/open.c:1453 [inline]
__x64_sys_openat+0x247/0x2a0 fs/open.c:1453
do_syscall_x64 arch/x86/entry/common.c:52 [inline]
do_syscall_64+0xf3/0x230 arch/x86/entry/common.c:83
entry_SYSCALL_64_after_hwframe+0x77/0x7f
RIP: 0033:0x7fd43ad169a4
RSP: 002b:00007fff676629e0 EFLAGS: 00000246 ORIG_RAX: 0000000000000101
RAX: ffffffffffffffda RBX: 000055be8f2d1a40 RCX: 00007fd43ad169a4
RDX: 00000000000a0800 RSI: 000055be8f2cc330 RDI: 00000000ffffff9c
RBP: 000055be8f2cc330 R08: 0000000000000001 R09: 7fffffffffffffff
R10: 0000000000000000 R11: 0000000000000246 R12: 00000000000a0800
R13: 000055be8f2c2a60 R14: 0000000000000001 R15: 000055be8f2b1910
Showing all locks held in the system:
2 locks held by kworker/u8:0/11:
3 locks held by kworker/u8:1/12:
#0: ffff88801ac89148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3204 [inline]
#0: ffff88801ac89148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_scheduled_works+0x93b/0x1840 kernel/workqueue.c:3310
#1: ffffc90000117c60 ((linkwatch_work).work){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3205 [inline]
#1: ffffc90000117c60 ((linkwatch_work).work){+.+.}-{0:0}, at: process_scheduled_works+0x976/0x1840 kernel/workqueue.c:3310
#2: ffffffff8fcbab48 (rtnl_mutex){+.+.}-{4:4}, at: linkwatch_event+0xe/0x60 net/core/link_watch.c:281
1 lock held by khungtaskd/30:
#0: ffffffff8e937da0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire include/linux/rcupdate.h:337 [inline]
#0: ffffffff8e937da0 (rcu_read_lock){....}-{1:3}, at: rcu_read_lock include/linux/rcupdate.h:849 [inline]
#0: ffffffff8e937da0 (rcu_read_lock){....}-{1:3}, at: debug_show_all_locks+0x55/0x2a0 kernel/locking/lockdep.c:6744
3 locks held by kworker/1:2/1163:
#0: ffff88801ac80d48 ((wq_completion)events){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3204 [inline]
#0: ffff88801ac80d48 ((wq_completion)events){+.+.}-{0:0}, at: process_scheduled_works+0x93b/0x1840 kernel/workqueue.c:3310
#1: ffffc9000445fc60 ((work_completion)(&data->fib_event_work)){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3205 [inline]
#1: ffffc9000445fc60 ((work_completion)(&data->fib_event_work)){+.+.}-{0:0}, at: process_scheduled_works+0x976/0x1840 kernel/workqueue.c:3310
#2: ffff88806ac0c240 (&data->fib_lock){+.+.}-{4:4}, at: nsim_fib_event_work+0x2d1/0x4130 drivers/net/netdevsim/fib.c:1490
2 locks held by getty/5591:
#0: ffff8880319140a0 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x25/0x70 drivers/tty/tty_ldisc.c:243
#1: ffffc90002fde2f0 (&ldata->atomic_read_lock){+.+.}-{4:4}, at: n_tty_read+0x6a6/0x1e00 drivers/tty/n_tty.c:2211
2 locks held by kworker/u8:8/6103:
#0: ffff8880b863e998 (&rq->__lock){-.-.}-{2:2}, at: raw_spin_rq_lock_nested+0x2a/0x140 kernel/sched/core.c:598
#1: ffff8880b8628948 (&per_cpu_ptr(group->pcpu, cpu)->seq){-.-.}-{0:0}, at: psi_task_switch+0x387/0x7a0 kernel/sched/psi.c:975
6 locks held by kworker/u8:9/6104:
#0: ffff88801baf5948 ((wq_completion)netns){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3204 [inline]
#0: ffff88801baf5948 ((wq_completion)netns){+.+.}-{0:0}, at: process_scheduled_works+0x93b/0x1840 kernel/workqueue.c:3310
#1: ffffc9000498fc60 (net_cleanup_work){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3205 [inline]
#1: ffffc9000498fc60 (net_cleanup_work){+.+.}-{0:0}, at: process_scheduled_works+0x976/0x1840 kernel/workqueue.c:3310
#2: ffffffff8fcae690 (pernet_ops_rwsem){++++}-{4:4}, at: cleanup_net+0x16a/0xd50 net/core/net_namespace.c:602
#3: ffff8880684c10e8 (&dev->mutex){....}-{4:4}, at: device_lock include/linux/device.h:1014 [inline]
#3: ffff8880684c10e8 (&dev->mutex){....}-{4:4}, at: devl_dev_lock net/devlink/devl_internal.h:108 [inline]
#3: ffff8880684c10e8 (&dev->mutex){....}-{4:4}, at: devlink_pernet_pre_exit+0x13b/0x440 net/devlink/core.c:506
#4: ffff88806a808250 (&devlink->lock_key#58){+.+.}-{4:4}, at: devl_lock net/devlink/core.c:276 [inline]
#4: ffff88806a808250 (&devlink->lock_key#58){+.+.}-{4:4}, at: devl_dev_lock net/devlink/devl_internal.h:109 [inline]
#4: ffff88806a808250 (&devlink->lock_key#58){+.+.}-{4:4}, at: devlink_pernet_pre_exit+0x14d/0x440 net/devlink/core.c:506
#5: ffffffff8fcbab48 (rtnl_mutex){+.+.}-{4:4}, at: register_netdevice_notifier_net+0x1a/0xa0 net/core/dev.c:1889
3 locks held by kworker/u8:10/6396:
#0: ffff88814d255948 ((wq_completion)ipv6_addrconf){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3204 [inline]
#0: ffff88814d255948 ((wq_completion)ipv6_addrconf){+.+.}-{0:0}, at: process_scheduled_works+0x93b/0x1840 kernel/workqueue.c:3310
#1: ffffc90003fafc60 ((work_completion)(&(&ifa->dad_work)->work)){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3205 [inline]
#1: ffffc90003fafc60 ((work_completion)(&(&ifa->dad_work)->work)){+.+.}-{0:0}, at: process_scheduled_works+0x976/0x1840 kernel/workqueue.c:3310
#2: ffffffff8fcbab48 (rtnl_mutex){+.+.}-{4:4}, at: addrconf_dad_work+0xd0/0x16f0 net/ipv6/addrconf.c:4215
3 locks held by udevd/7640:
#0: ffff8881443114c8 (&disk->open_mutex){+.+.}-{4:4}, at: bdev_open+0xf0/0xc50 block/bdev.c:904
#1: ffff888141b52a10 (set->srcu){.+.+}-{0:0}, at: srcu_lock_acquire include/linux/srcu.h:164 [inline]
#1: ffff888141b52a10 (set->srcu){.+.+}-{0:0}, at: srcu_read_lock include/linux/srcu.h:256 [inline]
#1: ffff888141b52a10 (set->srcu){.+.+}-{0:0}, at: blk_mq_run_hw_queue+0x32b/0x500 block/blk-mq.c:2354
#2: ffff888025e6e178 (&cmd->lock){+.+.}-{4:4}, at: nbd_queue_rq+0xfe/0x2ef0 drivers/block/nbd.c:1135
1 lock held by syz-executor/13542:
#0: ffffffff8e93d278 (rcu_state.exp_mutex){+.+.}-{4:4}, at: exp_funnel_lock kernel/rcu/tree_exp.h:302 [inline]
#0: ffffffff8e93d278 (rcu_state.exp_mutex){+.+.}-{4:4}, at: synchronize_rcu_expedited+0x381/0x830 kernel/rcu/tree_exp.h:996
1 lock held by syz-executor/13545:
#0: ffffffff8e93d278 (rcu_state.exp_mutex){+.+.}-{4:4}, at: exp_funnel_lock kernel/rcu/tree_exp.h:334 [inline]
#0: ffffffff8e93d278 (rcu_state.exp_mutex){+.+.}-{4:4}, at: synchronize_rcu_expedited+0x451/0x830 kernel/rcu/tree_exp.h:996
2 locks held by syz-executor/13566:
#0: ffffffff8fcae690 (pernet_ops_rwsem){++++}-{4:4}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:512
#1: ffffffff8fcbab48 (rtnl_mutex){+.+.}-{4:4}, at: wg_netns_pre_exit+0x1f/0x1e0 drivers/net/wireguard/device.c:415
2 locks held by syz-executor/13569:
#0: ffffffff8fcae690 (pernet_ops_rwsem){++++}-{4:4}, at: copy_net_ns+0x328/0x570 net/core/net_namespace.c:512
#1: ffffffff8fcbab48 (rtnl_mutex){+.+.}-{4:4}, at: ip_tunnel_init_net+0x20e/0x720 net/ipv4/ip_tunnel.c:1159
=============================================
NMI backtrace for cpu 1
CPU: 1 UID: 0 PID: 30 Comm: khungtaskd Not tainted 6.13.0-rc3-next-20241220-syzkaller #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 09/13/2024
Call Trace:
__dump_stack lib/dump_stack.c:94 [inline]
dump_stack_lvl+0x241/0x360 lib/dump_stack.c:120
nmi_cpu_backtrace+0x49c/0x4d0 lib/nmi_backtrace.c:113
nmi_trigger_cpumask_backtrace+0x198/0x320 lib/nmi_backtrace.c:62
trigger_all_cpu_backtrace include/linux/nmi.h:162 [inline]
check_hung_uninterruptible_tasks kernel/hung_task.c:234 [inline]
watchdog+0xff6/0x1040 kernel/hung_task.c:397
kthread+0x7a9/0x920 kernel/kthread.c:464
ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:148
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
Sending NMI from CPU 1 to CPUs 0:
NMI backtrace for cpu 0
CPU: 0 UID: 0 PID: 941 Comm: kworker/0:2 Not tainted 6.13.0-rc3-next-20241220-syzkaller #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 09/13/2024
Workqueue: events drain_vmap_area_work
RIP: 0010:preempt_count_sub+0x6/0x170 kernel/sched/core.c:5872
Code: c7 c7 94 e4 1a 90 e8 59 31 98 00 e9 66 ff ff ff 0f 1f 40 00 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 90 f3 0f 1e fa 41 56 <53> 89 fb 48 c7 c0 80 e9 3d 9a 48 c1 e8 03 49 be 00 00 00 00 00 fc
RSP: 0018:ffffc90003c77300 EFLAGS: 00000202
RAX: ffffc90003c78000 RBX: ffffc90003c773f0 RCX: ffffc90003c77e48
RDX: dffffc0000000000 RSI: ffffc90003c70000 RDI: 0000000000000001
RBP: ffffc90003c773e0 R08: ffffc90003c77e30 R09: 0000000000000000
R10: ffffc90003c77430 R11: fffff5200078ee88 R12: dffffc0000000000
R13: ffffc90003c773e0 R14: ffffc90003c70000 R15: ffffc90003c77e40
FS: 0000000000000000(0000) GS:ffff8880b8600000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 000055ae6f32b088 CR3: 000000000e736000 CR4: 00000000003526f0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
Call Trace:
unwind_next_frame+0x18e6/0x22d0 arch/x86/kernel/unwind_orc.c:672
arch_stack_walk+0x11c/0x150 arch/x86/kernel/stacktrace.c:25
stack_trace_save+0x118/0x1d0 kernel/stacktrace.c:122
save_stack+0xfb/0x1f0 mm/page_owner.c:156
__reset_page_owner+0x76/0x430 mm/page_owner.c:297
reset_page_owner include/linux/page_owner.h:25 [inline]
free_pages_prepare mm/page_alloc.c:1127 [inline]
free_frozen_pages+0xe0d/0x10e0 mm/page_alloc.c:2660
kasan_depopulate_vmalloc_pte+0x74/0x90 mm/kasan/shadow.c:408
apply_to_pte_range mm/memory.c:2901 [inline]
apply_to_pmd_range mm/memory.c:2945 [inline]
apply_to_pud_range mm/memory.c:2981 [inline]
apply_to_p4d_range mm/memory.c:3017 [inline]
__apply_to_page_range+0x806/0xde0 mm/memory.c:3051
kasan_release_vmalloc+0xa5/0xd0 mm/kasan/shadow.c:529
kasan_release_vmalloc_node mm/vmalloc.c:2196 [inline]
purge_vmap_node+0x22f/0x8d0 mm/vmalloc.c:2213
__purge_vmap_area_lazy+0x708/0xae0 mm/vmalloc.c:2304
drain_vmap_area_work+0x27/0x40 mm/vmalloc.c:2338
process_one_work kernel/workqueue.c:3229 [inline]
process_scheduled_works+0xa66/0x1840 kernel/workqueue.c:3310
worker_thread+0x870/0xd30 kernel/workqueue.c:3391
kthread+0x7a9/0x920 kernel/kthread.c:464
ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:148
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244