INFO: task kworker/u8:8:2829 blocked for more than 144 seconds.
Not tainted 6.9.0-syzkaller-10323-g8f6a15f095a6 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/u8:8 state:D stack:19096 pid:2829 tgid:2829 ppid:2 flags:0x00004000
Workqueue: loop3 loop_workfn
Call Trace:
<TASK>
context_switch kernel/sched/core.c:5408 [inline]
__schedule+0xf15/0x5d00 kernel/sched/core.c:6745
__schedule_loop kernel/sched/core.c:6822 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6837
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6894
rwsem_down_write_slowpath kernel/locking/rwsem.c:1178 [inline]
__down_write_common+0x950/0x13f0 kernel/locking/rwsem.c:1306
inode_lock include/linux/fs.h:791 [inline]
shmem_file_write_iter+0x8c/0x140 mm/shmem.c:2909
call_write_iter include/linux/fs.h:2114 [inline]
do_iter_readv_writev+0x504/0x780 fs/read_write.c:741
vfs_iter_write+0x1eb/0x9c0 fs/read_write.c:895
lo_write_bvec drivers/block/loop.c:246 [inline]
lo_write_simple drivers/block/loop.c:267 [inline]
do_req_filebacked drivers/block/loop.c:491 [inline]
loop_handle_cmd drivers/block/loop.c:1907 [inline]
loop_process_work+0x1577/0x20c0 drivers/block/loop.c:1942
process_one_work+0x9fb/0x1b60 kernel/workqueue.c:3231
process_scheduled_works kernel/workqueue.c:3312 [inline]
worker_thread+0x6c8/0xf70 kernel/workqueue.c:3393
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
</TASK>
INFO: task syz-executor.3:9484 blocked for more than 145 seconds.
Not tainted 6.9.0-syzkaller-10323-g8f6a15f095a6 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:syz-executor.3 state:D stack:19944 pid:9484 tgid:9483 ppid:8817 flags:0x00004002
Call Trace:
<TASK>
context_switch kernel/sched/core.c:5408 [inline]
__schedule+0xf15/0x5d00 kernel/sched/core.c:6745
__schedule_loop kernel/sched/core.c:6822 [inline]
schedule+0xe7/0x350 kernel/sched/core.c:6837
io_schedule+0xbf/0x130 kernel/sched/core.c:9043
folio_wait_bit_common+0x3d8/0x9b0 mm/filemap.c:1307
folio_wait_writeback+0x3d/0x2a0 mm/page-writeback.c:3140
__filemap_fdatawait_range+0x11e/0x230 mm/filemap.c:532
filemap_write_and_wait_range mm/filemap.c:694 [inline]
filemap_write_and_wait_range+0xeb/0x130 mm/filemap.c:676
filemap_write_and_wait include/linux/pagemap.h:66 [inline]
sync_blockdev block/bdev.c:207 [inline]
sync_blockdev+0x69/0x90 block/bdev.c:203
sync_filesystem+0x20a/0x290 fs/sync.c:70
generic_shutdown_super+0x7e/0x3d0 fs/super.c:621
kill_block_super+0x3b/0x90 fs/super.c:1676
ntfs3_kill_sb+0x3f/0xf0 fs/ntfs3/super.c:1798
deactivate_locked_super+0xbe/0x1a0 fs/super.c:473
deactivate_super+0xde/0x100 fs/super.c:506
cleanup_mnt+0x222/0x450 fs/namespace.c:1267
task_work_run+0x14e/0x250 kernel/task_work.c:180
exit_task_work include/linux/task_work.h:38 [inline]
do_exit+0xa7d/0x2c10 kernel/exit.c:877
do_group_exit+0xd3/0x2a0 kernel/exit.c:1026
get_signal+0x2616/0x2710 kernel/signal.c:2911
arch_do_signal_or_restart+0x90/0x7e0 arch/x86/kernel/signal.c:310
exit_to_user_mode_loop kernel/entry/common.c:111 [inline]
exit_to_user_mode_prepare include/linux/entry-common.h:328 [inline]
__syscall_exit_to_user_mode_work kernel/entry/common.c:207 [inline]
syscall_exit_to_user_mode+0x14a/0x2a0 kernel/entry/common.c:218
do_syscall_64+0xdc/0x260 arch/x86/entry/common.c:89
entry_SYSCALL_64_after_hwframe+0x77/0x7f
RIP: 0033:0x7fa7f827cee9
RSP: 002b:00007fa7f8f7f0c8 EFLAGS: 00000246 ORIG_RAX: 0000000000000101
RAX: fffffffffffffff4 RBX: 00007fa7f83abf80 RCX: 00007fa7f827cee9
RDX: 00000000000026e1 RSI: 0000000020000180 RDI: ffffffffffffff9c
RBP: 00007fa7f82c949e R08: 0000000000000000 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000
R13: 000000000000000b R14: 00007fa7f83abf80 R15: 00007fffc2fd1bd8
</TASK>
Showing all locks held in the system:
4 locks held by kworker/u8:0/11:
1 lock held by khungtaskd/30:
#0: ffffffff8dbb1760 (rcu_read_lock){....}-{1:2}, at: rcu_lock_acquire include/linux/rcupdate.h:329 [inline]
#0: ffffffff8dbb1760 (rcu_read_lock){....}-{1:2}, at: rcu_read_lock include/linux/rcupdate.h:781 [inline]
#0: ffffffff8dbb1760 (rcu_read_lock){....}-{1:2}, at: debug_show_all_locks+0x75/0x340 kernel/locking/lockdep.c:6614
2 locks held by kswapd0/88:
4 locks held by kswapd1/89:
3 locks held by kworker/0:2/784:
2 locks held by kworker/u8:7/2813:
4 locks held by kworker/u8:8/2829:
#0: ffff88805a4ed948 ((wq_completion)loop3){+.+.}-{0:0}, at: process_one_work+0x12bf/0x1b60 kernel/workqueue.c:3206
#1: ffffc9000b36fd80 ((work_completion)(&worker->work)){+.+.}-{0:0}, at: process_one_work+0x957/0x1b60 kernel/workqueue.c:3207
#2: ffff8880162f2420 (sb_writers#6){.+.+}-{0:0}, at: lo_write_bvec drivers/block/loop.c:246 [inline]
#2: ffff8880162f2420 (sb_writers#6){.+.+}-{0:0}, at: lo_write_simple drivers/block/loop.c:267 [inline]
#2: ffff8880162f2420 (sb_writers#6){.+.+}-{0:0}, at: do_req_filebacked drivers/block/loop.c:491 [inline]
#2: ffff8880162f2420 (sb_writers#6){.+.+}-{0:0}, at: loop_handle_cmd drivers/block/loop.c:1907 [inline]
#2: ffff8880162f2420 (sb_writers#6){.+.+}-{0:0}, at: loop_process_work+0x1577/0x20c0 drivers/block/loop.c:1942
#3: ffff88805ca49d10 (&sb->s_type->i_mutex_key#12){+.+.}-{3:3}, at: inode_lock include/linux/fs.h:791 [inline]
#3: ffff88805ca49d10 (&sb->s_type->i_mutex_key#12){+.+.}-{3:3}, at: shmem_file_write_iter+0x8c/0x140 mm/shmem.c:2909
2 locks held by kworker/u8:9/2838:
#0: ffff888015489148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_one_work+0x12bf/0x1b60 kernel/workqueue.c:3206
#1: ffffc9000b44fd80 ((work_completion)(&sub_info->work)){+.+.}-{0:0}, at: process_one_work+0x957/0x1b60 kernel/workqueue.c:3207
6 locks held by kworker/u8:10/3827:
1 lock held by jbd2/sda1-8/4498:
2 locks held by getty/4839:
#0: ffff88802b3650a0 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x24/0x80 drivers/tty/tty_ldisc.c:243
#1: ffffc90002f0e2f0 (&ldata->atomic_read_lock){+.+.}-{3:3}, at: n_tty_read+0xfc8/0x1490 drivers/tty/n_tty.c:2201
2 locks held by syz-fuzzer/5076:
#0: ffff88807e30d658 (&vma->vm_lock->lock){++++}-{3:3}, at: vma_start_read include/linux/mm.h:678 [inline]
#0: ffff88807e30d658 (&vma->vm_lock->lock){++++}-{3:3}, at: lock_vma_under_rcu+0x1e2/0x8f0 mm/memory.c:5844
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __perform_reclaim mm/page_alloc.c:3856 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_direct_reclaim mm/page_alloc.c:3881 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_slowpath mm/page_alloc.c:4287 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_noprof+0xaae/0x2460 mm/page_alloc.c:4673
2 locks held by syz-fuzzer/5091:
2 locks held by syz-fuzzer/5093:
#0: ffff88807e30d658 (&vma->vm_lock->lock){++++}-{3:3}, at: vma_start_read include/linux/mm.h:678 [inline]
#0: ffff88807e30d658 (&vma->vm_lock->lock){++++}-{3:3}, at: lock_vma_under_rcu+0x1e2/0x8f0 mm/memory.c:5844
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __perform_reclaim mm/page_alloc.c:3856 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_direct_reclaim mm/page_alloc.c:3881 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_slowpath mm/page_alloc.c:4287 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_noprof+0xaae/0x2460 mm/page_alloc.c:4673
4 locks held by kworker/u9:4/5115:
#0: ffff88802cd97148 ((wq_completion)hci4){+.+.}-{0:0}, at: process_one_work+0x12bf/0x1b60 kernel/workqueue.c:3206
#1: ffffc9000359fd80 ((work_completion)(&hdev->cmd_sync_work)){+.+.}-{0:0}, at: process_one_work+0x957/0x1b60 kernel/workqueue.c:3207
#2: ffff88801f0e8d88 (&hdev->req_lock){+.+.}-{3:3}, at: hci_cmd_sync_work+0x170/0x410 net/bluetooth/hci_sync.c:309
#3: ffff88801f0e8078 (&hdev->lock){+.+.}-{3:3}, at: hci_abort_conn_sync+0x150/0xb50 net/bluetooth/hci_sync.c:5431
3 locks held by kworker/1:5/5158:
2 locks held by syz-executor.1/8299:
#0: ffff88807bdbf398 (&mm->mmap_lock){++++}-{3:3}, at: mmap_read_trylock include/linux/mmap_lock.h:163 [inline]
#0: ffff88807bdbf398 (&mm->mmap_lock){++++}-{3:3}, at: get_mmap_lock_carefully mm/memory.c:5715 [inline]
#0: ffff88807bdbf398 (&mm->mmap_lock){++++}-{3:3}, at: lock_mm_and_find_vma+0x35/0x6a0 mm/memory.c:5775
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __perform_reclaim mm/page_alloc.c:3856 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_direct_reclaim mm/page_alloc.c:3881 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_slowpath mm/page_alloc.c:4287 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_noprof+0xaae/0x2460 mm/page_alloc.c:4673
3 locks held by syz-executor.4/9320:
#0: ffff88807ceca420 (sb_writers#22){.+.+}-{0:0}, at: open_last_lookups fs/namei.c:3563 [inline]
#0: ffff88807ceca420 (sb_writers#22){.+.+}-{0:0}, at: path_openat+0x19e6/0x29f0 fs/namei.c:3804
#1: ffff88805b3d83a0 (&type->i_mutex_dir_key#14){++++}-{3:3}, at: inode_lock include/linux/fs.h:791 [inline]
#1: ffff88805b3d83a0 (&type->i_mutex_dir_key#14){++++}-{3:3}, at: open_last_lookups fs/namei.c:3571 [inline]
#1: ffff88805b3d83a0 (&type->i_mutex_dir_key#14){++++}-{3:3}, at: path_openat+0x8c7/0x29f0 fs/namei.c:3804
#2: ffff88805b3d8100 (&ni->ni_lock#2/5){+.+.}-{3:3}, at: ni_lock_dir fs/ntfs3/ntfs_fs.h:1128 [inline]
#2: ffff88805b3d8100 (&ni->ni_lock#2/5){+.+.}-{3:3}, at: ntfs_atomic_open+0x38a/0x650 fs/ntfs3/namei.c:404
4 locks held by kworker/u8:2/9386:
2 locks held by syz-executor.4/9481:
1 lock held by syz-executor.3/9484:
#0: ffff8880523d20e0 (&type->s_umount_key#92){++++}-{3:3}, at: __super_lock fs/super.c:56 [inline]
#0: ffff8880523d20e0 (&type->s_umount_key#92){++++}-{3:3}, at: __super_lock_excl fs/super.c:71 [inline]
#0: ffff8880523d20e0 (&type->s_umount_key#92){++++}-{3:3}, at: deactivate_super+0xd6/0x100 fs/super.c:505
2 locks held by syz-executor.3/9681:
#0: ffff888079aab7a0 (mapping.invalidate_lock){++++}-{3:3}, at: filemap_invalidate_lock_shared include/linux/fs.h:846 [inline]
#0: ffff888079aab7a0 (mapping.invalidate_lock){++++}-{3:3}, at: filemap_fault+0x5c2/0x2610 mm/filemap.c:3320
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __perform_reclaim mm/page_alloc.c:3856 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_direct_reclaim mm/page_alloc.c:3881 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_slowpath mm/page_alloc.c:4287 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_noprof+0xaae/0x2460 mm/page_alloc.c:4673
3 locks held by syz-executor.0/9719:
#0: ffff88822fd34420 (sb_writers#22){.+.+}-{0:0}, at: open_last_lookups fs/namei.c:3563 [inline]
#0: ffff88822fd34420 (sb_writers#22){.+.+}-{0:0}, at: path_openat+0x19e6/0x29f0 fs/namei.c:3804
#1: ffff888076b23740 (&type->i_mutex_dir_key#14){++++}-{3:3}, at: inode_lock include/linux/fs.h:791 [inline]
#1: ffff888076b23740 (&type->i_mutex_dir_key#14){++++}-{3:3}, at: open_last_lookups fs/namei.c:3571 [inline]
#1: ffff888076b23740 (&type->i_mutex_dir_key#14){++++}-{3:3}, at: path_openat+0x8c7/0x29f0 fs/namei.c:3804
#2: ffff888076b234a0 (&ni->ni_lock#2/5){+.+.}-{3:3}, at: ni_lock_dir fs/ntfs3/ntfs_fs.h:1128 [inline]
#2: ffff888076b234a0 (&ni->ni_lock#2/5){+.+.}-{3:3}, at: ntfs_atomic_open+0x38a/0x650 fs/ntfs3/namei.c:404
2 locks held by modprobe/9750:
#0: ffff888028325fa0 (mapping.invalidate_lock){++++}-{3:3}, at: filemap_invalidate_lock_shared include/linux/fs.h:846 [inline]
#0: ffff888028325fa0 (mapping.invalidate_lock){++++}-{3:3}, at: filemap_fault+0x5c2/0x2610 mm/filemap.c:3320
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __perform_reclaim mm/page_alloc.c:3856 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_direct_reclaim mm/page_alloc.c:3881 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_slowpath mm/page_alloc.c:4287 [inline]
#1: ffffffff8dd33860 (fs_reclaim){+.+.}-{0:0}, at: __alloc_pages_noprof+0xaae/0x2460 mm/page_alloc.c:4673
=============================================
NMI backtrace for cpu 1
CPU: 1 PID: 30 Comm: khungtaskd Not tainted 6.9.0-syzkaller-10323-g8f6a15f095a6 #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 04/02/2024
Call Trace:
<TASK>
__dump_stack lib/dump_stack.c:88 [inline]
dump_stack_lvl+0x116/0x1f0 lib/dump_stack.c:114
nmi_cpu_backtrace+0x27b/0x390 lib/nmi_backtrace.c:113
nmi_trigger_cpumask_backtrace+0x29c/0x300 lib/nmi_backtrace.c:62
trigger_all_cpu_backtrace include/linux/nmi.h:162 [inline]
check_hung_uninterruptible_tasks kernel/hung_task.c:223 [inline]
watchdog+0xf86/0x1240 kernel/hung_task.c:379
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
</TASK>
Sending NMI from CPU 1 to CPUs 0:
NMI backtrace for cpu 0
CPU: 0 PID: 784 Comm: kworker/0:2 Not tainted 6.9.0-syzkaller-10323-g8f6a15f095a6 #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 04/02/2024
Workqueue: events refill_work
RIP: 0010:check_region_inline mm/kasan/generic.c:169 [inline]
RIP: 0010:kasan_check_range+0x150/0x1a0 mm/kasan/generic.c:189
Code: c0 01 49 89 d3 48 39 d0 74 11 80 38 00 74 ef 4d 8d 1c 2c 48 85 c0 48 89 c2 75 93 48 89 da 4c 89 d8 4c 29 da e9 2c ff ff ff 5b <b8> 01 00 00 00 5d 41 5c c3 cc cc cc cc b8 01 00 00 00 c3 cc cc cc
RSP: 0000:ffffc90004277290 EFLAGS: 00000246
RAX: ffffed1005e4a7c5 RBX: 0000000000000001 RCX: ffffffff81c9142f
RDX: ffffed1005e4a7c5 RSI: 0000000000000004 RDI: ffff88802f253e20
RBP: ffffed1005e4a7c4 R08: 0000000000000001 R09: ffffed1005e4a7c4
R10: ffff88802f253e23 R11: 0000000000000003 R12: ffff88802f253e00
R13: dffffc0000000000 R14: 0000000000000002 R15: ffffc90004277350
FS: 0000000000000000(0000) GS:ffff8880b9200000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 00007f4ae0a7b920 CR3: 0000000214d16000 CR4: 00000000003506f0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
Call Trace:
<NMI>
</NMI>
<TASK>
instrument_atomic_read_write include/linux/instrumented.h:96 [inline]
atomic_try_cmpxchg_relaxed include/linux/atomic/atomic-instrumented.h:1345 [inline]
__refcount_add_not_zero include/linux/refcount.h:147 [inline]
__refcount_inc_not_zero include/linux/refcount.h:218 [inline]
refcount_inc_not_zero include/linux/refcount.h:236 [inline]
shrinker_try_get include/linux/shrinker.h:141 [inline]
shrink_slab+0x32f/0x1310 mm/shrinker.c:657
shrink_one+0x493/0x7c0 mm/vmscan.c:4790
shrink_many mm/vmscan.c:4851 [inline]
lru_gen_shrink_node mm/vmscan.c:4951 [inline]
shrink_node+0x22bd/0x39b0 mm/vmscan.c:5910
shrink_zones mm/vmscan.c:6168 [inline]
do_try_to_free_pages+0x35f/0x1940 mm/vmscan.c:6230
try_to_free_pages+0x2b6/0x720 mm/vmscan.c:6465
__perform_reclaim mm/page_alloc.c:3859 [inline]
__alloc_pages_direct_reclaim mm/page_alloc.c:3881 [inline]
__alloc_pages_slowpath mm/page_alloc.c:4287 [inline]
__alloc_pages_noprof+0xb38/0x2460 mm/page_alloc.c:4673
alloc_pages_mpol_noprof+0x275/0x610 mm/mempolicy.c:2265
skb_page_frag_refill+0x25a/0x350 net/core/sock.c:2929
virtnet_rq_alloc+0x33/0x8e0 drivers/net/virtio_net.c:882
add_recvbuf_mergeable drivers/net/virtio_net.c:2110 [inline]
try_fill_recv+0x783/0x1840 drivers/net/virtio_net.c:2155
refill_work+0x135/0x200 drivers/net/virtio_net.c:2233
process_one_work+0x9fb/0x1b60 kernel/workqueue.c:3231
process_scheduled_works kernel/workqueue.c:3312 [inline]
worker_thread+0x6c8/0xf70 kernel/workqueue.c:3393
kthread+0x2c1/0x3a0 kernel/kthread.c:389
ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
</TASK>