INFO: task bch-reclaim/loo:6066 blocked for more than 143 seconds. Not tainted 6.16.0-rc2-syzkaller-00087-g24770983ccfe #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:bch-reclaim/loo state:D stack:18952 pid:6066 tgid:6066 ppid:2 task_flags:0x200040 flags:0x00004000 Call Trace: context_switch kernel/sched/core.c:5396 [inline] __schedule+0x16f5/0x4d00 kernel/sched/core.c:6785 __schedule_loop kernel/sched/core.c:6863 [inline] schedule+0x165/0x360 kernel/sched/core.c:6878 schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6935 __mutex_lock_common kernel/locking/mutex.c:679 [inline] __mutex_lock+0x724/0xe80 kernel/locking/mutex.c:747 bch2_journal_reclaim_thread+0x16b/0x4f0 fs/bcachefs/journal_reclaim.c:767 kthread+0x711/0x8a0 kernel/kthread.c:464 ret_from_fork+0x3fc/0x770 arch/x86/kernel/process.c:148 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:245 Showing all locks held in the system: 1 lock held by pool_workqueue_/3: #0: ffffffff8e1448b8 (rcu_state.exp_mutex){+.+.}-{4:4}, at: exp_funnel_lock kernel/rcu/tree_exp.h:336 [inline] #0: ffffffff8e1448b8 (rcu_state.exp_mutex){+.+.}-{4:4}, at: synchronize_rcu_expedited+0x3b9/0x730 kernel/rcu/tree_exp.h:998 1 lock held by khungtaskd/31: #0: ffffffff8e13eda0 (rcu_read_lock){....}-{1:3}, at: rcu_lock_acquire include/linux/rcupdate.h:331 [inline] #0: ffffffff8e13eda0 (rcu_read_lock){....}-{1:3}, at: rcu_read_lock include/linux/rcupdate.h:841 [inline] #0: ffffffff8e13eda0 (rcu_read_lock){....}-{1:3}, at: debug_show_all_locks+0x2e/0x180 kernel/locking/lockdep.c:6770 4 locks held by kworker/u8:3/50: #0: ffff888026e57948 ((wq_completion)btree_update#67){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3213 [inline] #0: ffff888026e57948 ((wq_completion)btree_update#67){+.+.}-{0:0}, at: process_scheduled_works+0x9b4/0x17b0 kernel/workqueue.c:3321 #1: ffffc90000bb7bc0 ((work_completion)(&c->btree_interior_update_work)){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3214 [inline] #1: ffffc90000bb7bc0 ((work_completion)(&c->btree_interior_update_work)){+.+.}-{0:0}, at: process_scheduled_works+0x9ef/0x17b0 kernel/workqueue.c:3321 #2: ffff888083c84390 (&c->btree_trans_barrier){.+.+}-{0:0}, at: srcu_lock_acquire include/linux/srcu.h:161 [inline] #2: ffff888083c84390 (&c->btree_trans_barrier){.+.+}-{0:0}, at: srcu_read_lock include/linux/srcu.h:253 [inline] #2: ffff888083c84390 (&c->btree_trans_barrier){.+.+}-{0:0}, at: bch2_trans_srcu_lock+0xaf/0x220 fs/bcachefs/btree_iter.c:3255 #3: ffff888083ca6710 (&c->gc_lock){.+.+}-{4:4}, at: bch2_btree_update_start+0x542/0x1dc0 fs/bcachefs/btree_update_interior.c:1211 3 locks held by kworker/u8:4/60: 1 lock held by udevd/5199: #0: ffff8880271c2948 (vm_lock){++++}-{0:0}, at: do_user_addr_fault+0x2d9/0x1390 arch/x86/mm/fault.c:1327 2 locks held by getty/5591: #0: ffff88814cbbf0a0 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x25/0x70 drivers/tty/tty_ldisc.c:243 #1: ffffc9000332b2f0 (&ldata->atomic_read_lock){+.+.}-{4:4}, at: n_tty_read+0x43e/0x1400 drivers/tty/n_tty.c:2222 6 locks held by syz-executor332/5840: #0: ffff888048c5c0e0 (&type->s_umount_key#50){+.+.}-{4:4}, at: __super_lock fs/super.c:57 [inline] #0: ffff888048c5c0e0 (&type->s_umount_key#50){+.+.}-{4:4}, at: __super_lock_excl fs/super.c:72 [inline] #0: ffff888048c5c0e0 (&type->s_umount_key#50){+.+.}-{4:4}, at: deactivate_super+0xa9/0xe0 fs/super.c:506 #1: ffff888083c80278 (&c->state_lock){+.+.}-{4:4}, at: __bch2_fs_stop+0xf8/0x900 fs/bcachefs/super.c:677 #2: ffff888083ccaf68 (&j->reclaim_lock){+.+.}-{4:4}, at: journal_flush_done+0x86/0x810 fs/bcachefs/journal_reclaim.c:878 #3: ffff888083c84390 (&c->btree_trans_barrier){.+.+}-{0:0}, at: srcu_lock_acquire include/linux/srcu.h:161 [inline] #3: ffff888083c84390 (&c->btree_trans_barrier){.+.+}-{0:0}, at: srcu_read_lock include/linux/srcu.h:253 [inline] #3: ffff888083c84390 (&c->btree_trans_barrier){.+.+}-{0:0}, at: __bch2_trans_get+0x7f4/0xd80 fs/bcachefs/btree_iter.c:3440 #4: ffff888083c84738 (&wb->flushing.lock){+.+.}-{4:4}, at: btree_write_buffer_flush_seq+0x1829/0x19a0 fs/bcachefs/btree_write_buffer.c:572 #5: ffff888083ca6710 (&c->gc_lock){.+.+}-{4:4}, at: bch2_btree_update_start+0x542/0x1dc0 fs/bcachefs/btree_update_interior.c:1211 6 locks held by syz-executor332/5843: #0: ffff88807e2660e0 (&type->s_umount_key#50){+.+.}-{4:4}, at: __super_lock fs/super.c:57 [inline] #0: ffff88807e2660e0 (&type->s_umount_key#50){+.+.}-{4:4}, at: __super_lock_excl fs/super.c:72 [inline] #0: ffff88807e2660e0 (&type->s_umount_key#50){+.+.}-{4:4}, at: deactivate_super+0xa9/0xe0 fs/super.c:506 #1: ffff888052800278 (&c->state_lock){+.+.}-{4:4}, at: __bch2_fs_stop+0xf8/0x900 fs/bcachefs/super.c:677 #2: ffff88805284af68 (&j->reclaim_lock){+.+.}-{4:4}, at: journal_flush_done+0x86/0x810 fs/bcachefs/journal_reclaim.c:878 #3: ffff888052804390 (&c->btree_trans_barrier){.+.+}-{0:0}, at: srcu_lock_acquire include/linux/srcu.h:161 [inline] #3: ffff888052804390 (&c->btree_trans_barrier){.+.+}-{0:0}, at: srcu_read_lock include/linux/srcu.h:253 [inline] #3: ffff888052804390 (&c->btree_trans_barrier){.+.+}-{0:0}, at: __bch2_trans_get+0x7f4/0xd80 fs/bcachefs/btree_iter.c:3440 #4: ffff888052804738 (&wb->flushing.lock){+.+.}-{4:4}, at: btree_write_buffer_flush_seq+0x1829/0x19a0 fs/bcachefs/btree_write_buffer.c:572 #5: ffff888052826710 (&c->gc_lock){.+.+}-{4:4}, at: bch2_btree_update_start+0x542/0x1dc0 fs/bcachefs/btree_update_interior.c:1211 2 locks held by kworker/0:5/5930: 1 lock held by udevd/5969: 1 lock held by bch-reclaim/loo/6066: #0: ffff88805284af68 (&j->reclaim_lock){+.+.}-{4:4}, at: bch2_journal_reclaim_thread+0x16b/0x4f0 fs/bcachefs/journal_reclaim.c:767 1 lock held by udevd/6141: #0: ffff888028f5d588 (vm_lock){++++}-{0:0}, at: do_user_addr_fault+0x2d9/0x1390 arch/x86/mm/fault.c:1327 1 lock held by bch-reclaim/loo/6914: #0: ffff888083ccaf68 (&j->reclaim_lock){+.+.}-{4:4}, at: bch2_journal_reclaim_thread+0x16b/0x4f0 fs/bcachefs/journal_reclaim.c:767 3 locks held by syz-executor332/8057: 2 locks held by syz-executor332/8065: #0: ffff888037c00278 (&c->state_lock){+.+.}-{4:4}, at: bch2_fs_read_write_early+0x1e/0x40 fs/bcachefs/super.c:575 #1: ffffffff8dfe57a8 (wq_pool_mutex){+.+.}-{4:4}, at: apply_wqattrs_lock kernel/workqueue.c:5181 [inline] #1: ffffffff8dfe57a8 (wq_pool_mutex){+.+.}-{4:4}, at: __alloc_workqueue+0x9eb/0x1b70 kernel/workqueue.c:5736 1 lock held by syz-executor332/8072: #0: ffffffff8dfe57a8 (wq_pool_mutex){+.+.}-{4:4}, at: apply_wqattrs_lock kernel/workqueue.c:5181 [inline] #0: ffffffff8dfe57a8 (wq_pool_mutex){+.+.}-{4:4}, at: __alloc_workqueue+0x9eb/0x1b70 kernel/workqueue.c:5736 ============================================= NMI backtrace for cpu 1 CPU: 1 UID: 0 PID: 31 Comm: khungtaskd Not tainted 6.16.0-rc2-syzkaller-00087-g24770983ccfe #0 PREEMPT(full) Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 05/07/2025 Call Trace: dump_stack_lvl+0x189/0x250 lib/dump_stack.c:120 nmi_cpu_backtrace+0x39e/0x3d0 lib/nmi_backtrace.c:113 nmi_trigger_cpumask_backtrace+0x17a/0x300 lib/nmi_backtrace.c:62 trigger_all_cpu_backtrace include/linux/nmi.h:158 [inline] check_hung_uninterruptible_tasks kernel/hung_task.c:307 [inline] watchdog+0xfee/0x1030 kernel/hung_task.c:470 kthread+0x711/0x8a0 kernel/kthread.c:464 ret_from_fork+0x3fc/0x770 arch/x86/kernel/process.c:148 ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:245 Sending NMI from CPU 1 to CPUs 0: NMI backtrace for cpu 0 CPU: 0 UID: 0 PID: 8057 Comm: syz-executor332 Not tainted 6.16.0-rc2-syzkaller-00087-g24770983ccfe #0 PREEMPT(full) Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 05/07/2025 RIP: 0010:__list_add include/linux/list.h:150 [inline] RIP: 0010:list_add include/linux/list.h:169 [inline] RIP: 0010:lru_gen_add_folio+0x610/0xc70 include/linux/mm_inline.h:278 Code: 00 00 fc ff df 43 80 3c 26 00 74 08 4c 89 ef e8 76 b2 27 00 4d 8b 7d 00 48 8b 7c 24 08 4c 89 ee 4c 89 fa e8 12 73 cb 02 84 c0 <0f> 84 f9 01 00 00 e8 35 34 c4 ff 4c 89 e1 4d 8d 67 08 4c 89 e0 48 RSP: 0018:ffffc90003ee75c0 EFLAGS: 00000002 RAX: 0000000000000001 RBX: ffffea0001b9d540 RCX: 0000000000000001 RDX: ffffea0001b9d508 RSI: ffff88801cac3320 RDI: ffffea0001b9d548 RBP: ffffc90003ee7690 R08: ffffffff8fa10df7 R09: 1ffffffff1f421be R10: dffffc0000000000 R11: fffffbfff1f421bf R12: dffffc0000000000 R13: ffff88801cac3320 R14: 1ffff11003958664 R15: ffffea0001b9d508 FS: 00007f21816ea6c0(0000) GS:ffff888125c51000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000200001000000 CR3: 000000009b711000 CR4: 0000000000350ef0 Call Trace: lruvec_add_folio include/linux/mm_inline.h:344 [inline] lru_add+0x6ad/0xd80 mm/swap.c:154 folio_batch_move_lru+0x21e/0x3a0 mm/swap.c:168 __folio_batch_add_and_move+0x5ad/0xd20 mm/swap.c:196 filemap_add_folio+0x1af/0x270 mm/filemap.c:985 __filemap_get_folio+0x4f6/0xaf0 mm/filemap.c:1990 bch2_filemap_get_contig_folios_d+0x123/0x480 fs/bcachefs/fs-io-pagecache.c:31 __bch2_buffered_write fs/bcachefs/fs-io-buffered.c:836 [inline] bch2_buffered_write fs/bcachefs/fs-io-buffered.c:1026 [inline] bch2_write_iter+0x6f4/0x2b90 fs/bcachefs/fs-io-buffered.c:1082 new_sync_write fs/read_write.c:593 [inline] vfs_write+0x54b/0xa90 fs/read_write.c:686 ksys_write+0x145/0x250 fs/read_write.c:738 do_syscall_x64 arch/x86/entry/syscall_64.c:63 [inline] do_syscall_64+0xfa/0x3b0 arch/x86/entry/syscall_64.c:94 entry_SYSCALL_64_after_hwframe+0x77/0x7f RIP: 0033:0x7f2181f46169 Code: 28 00 00 00 75 05 48 83 c4 28 c3 e8 c1 1f 00 00 90 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 c7 c1 b0 ff ff ff f7 d8 64 89 01 48 RSP: 002b:00007f21816ea218 EFLAGS: 00000246 ORIG_RAX: 0000000000000001 RAX: ffffffffffffffda RBX: 00007f2181fd34a8 RCX: 00007f2181f46169 RDX: 000000000208e24b RSI: 0000200000000000 RDI: 0000000000000004 RBP: 00007f2181fd34a0 R08: 0000000000000000 R09: 0000000000000000 R10: 0000000000000000 R11: 0000000000000246 R12: 00007f2181f98374 R13: 00002000000015c0 R14: 00002000000015c8 R15: 0000200000000080