INFO: task syz.1.2:5349 blocked for more than 143 seconds.
Not tainted 6.11.0-syzkaller-11653-g3630400697a3 #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:syz.1.2 state:D stack:27256 pid:5349 tgid:5319 ppid:5218 flags:0x00000004
Call Trace:
context_switch kernel/sched/core.c:5315 [inline]
__schedule+0x1843/0x4ae0 kernel/sched/core.c:6675
__schedule_loop kernel/sched/core.c:6752 [inline]
schedule+0x14b/0x320 kernel/sched/core.c:6767
schedule_preempt_disabled+0x13/0x30 kernel/sched/core.c:6824
rwsem_down_write_slowpath+0xeee/0x13b0 kernel/locking/rwsem.c:1178
__down_write_common kernel/locking/rwsem.c:1306 [inline]
__down_write kernel/locking/rwsem.c:1315 [inline]
down_write+0x1d7/0x220 kernel/locking/rwsem.c:1580
inode_lock include/linux/fs.h:815 [inline]
bch2_fallocate_dispatch+0x1e2/0x540 fs/bcachefs/fs-io.c:762
vfs_fallocate+0x569/0x6e0 fs/open.c:333
do_vfs_ioctl+0x2583/0x2e40 fs/ioctl.c:886
__do_sys_ioctl fs/ioctl.c:905 [inline]
__se_sys_ioctl+0x81/0x170 fs/ioctl.c:893
do_syscall_x64 arch/x86/entry/common.c:52 [inline]
do_syscall_64+0xf3/0x230 arch/x86/entry/common.c:83
entry_SYSCALL_64_after_hwframe+0x77/0x7f
RIP: 0033:0x7f6f06f7dff9
RSP: 002b:00007f6f07e2b038 EFLAGS: 00000246 ORIG_RAX: 0000000000000010
RAX: ffffffffffffffda RBX: 00007f6f07136058 RCX: 00007f6f06f7dff9
RDX: 00000000200000c0 RSI: 0000000040305829 RDI: 0000000000000009
RBP: 00007f6f06ff0296 R08: 0000000000000000 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000
R13: 0000000000000000 R14: 00007f6f07136058 R15: 00007ffd9d8d5198
Showing all locks held in the system:
2 locks held by kworker/u8:0/11:
#0: ffff88801ac89148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3204 [inline]
#0: ffff88801ac89148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_scheduled_works+0x93b/0x1850 kernel/workqueue.c:3310
#1: ffffc90000107d00 ((reaper_work).work){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3205 [inline]
#1: ffffc90000107d00 ((reaper_work).work){+.+.}-{0:0}, at: process_scheduled_works+0x976/0x1850 kernel/workqueue.c:3310
2 locks held by kworker/u8:1/12:
#0: ffff88801ac89148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3204 [inline]
#0: ffff88801ac89148 ((wq_completion)events_unbound){+.+.}-{0:0}, at: process_scheduled_works+0x93b/0x1850 kernel/workqueue.c:3310
#1: ffffc90000117d00 (connector_reaper_work){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3205 [inline]
#1: ffffc90000117d00 (connector_reaper_work){+.+.}-{0:0}, at: process_scheduled_works+0x976/0x1850 kernel/workqueue.c:3310
1 lock held by khungtaskd/30:
#0: ffffffff8e937da0 (rcu_read_lock){....}-{1:2}, at: rcu_lock_acquire include/linux/rcupdate.h:337 [inline]
#0: ffffffff8e937da0 (rcu_read_lock){....}-{1:2}, at: rcu_read_lock include/linux/rcupdate.h:849 [inline]
#0: ffffffff8e937da0 (rcu_read_lock){....}-{1:2}, at: debug_show_all_locks+0x55/0x2a0 kernel/locking/lockdep.c:6701
3 locks held by kworker/u8:2/35:
#0: ffff8880202ba148 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3204 [inline]
#0: ffff8880202ba148 ((wq_completion)writeback){+.+.}-{0:0}, at: process_scheduled_works+0x93b/0x1850 kernel/workqueue.c:3310
#1: ffffc90000ab7d00 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3205 [inline]
#1: ffffc90000ab7d00 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_scheduled_works+0x976/0x1850 kernel/workqueue.c:3310
#2: ffff888027c2e0e0 (&type->s_umount_key#59){++++}-{3:3}, at: super_trylock_shared+0x22/0xf0 fs/super.c:562
2 locks held by kworker/u8:4/62:
#0: ffff888146ecd948 ((wq_completion)iou_exit){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3204 [inline]
#0: ffff888146ecd948 ((wq_completion)iou_exit){+.+.}-{0:0}, at: process_scheduled_works+0x93b/0x1850 kernel/workqueue.c:3310
#1: ffffc900015d7d00 ((work_completion)(&ctx->exit_work)){+.+.}-{0:0}, at: process_one_work kernel/workqueue.c:3205 [inline]
#1: ffffc900015d7d00 ((work_completion)(&ctx->exit_work)){+.+.}-{0:0}, at: process_scheduled_works+0x976/0x1850 kernel/workqueue.c:3310
2 locks held by kworker/u8:7/1306:
2 locks held by getty/4973:
#0: ffff88802e7eb0a0 (&tty->ldisc_sem){++++}-{0:0}, at: tty_ldisc_ref_wait+0x25/0x70 drivers/tty/tty_ldisc.c:243
#1: ffffc90002efe2f0 (&ldata->atomic_read_lock){+.+.}-{3:3}, at: n_tty_read+0x6a6/0x1e00 drivers/tty/n_tty.c:2211
5 locks held by syz.1.2/5320:
#0: ffff888027c2e420 (sb_writers#16){.+.+}-{0:0}, at: mnt_want_write+0x3f/0x90 fs/namespace.c:515
#1: ffff88805cd7a6a0 (&sb->s_type->i_mutex_key#25){++++}-{3:3}, at: inode_lock include/linux/fs.h:815 [inline]
#1: ffff88805cd7a6a0 (&sb->s_type->i_mutex_key#25){++++}-{3:3}, at: do_truncate+0x20c/0x310 fs/open.c:63
#2: ffff88805df00a38 (&c->snapshot_create_lock){.+.+}-{3:3}, at: bch2_truncate+0x16d/0x2c0 fs/bcachefs/io_misc.c:290
#3: ffff88805df04398 (&c->btree_trans_barrier){.+.+}-{0:0}, at: srcu_lock_acquire include/linux/srcu.h:151 [inline]
#3: ffff88805df04398 (&c->btree_trans_barrier){.+.+}-{0:0}, at: srcu_read_lock include/linux/srcu.h:250 [inline]
#3: ffff88805df04398 (&c->btree_trans_barrier){.+.+}-{0:0}, at: __bch2_trans_get+0x7de/0xd20 fs/bcachefs/btree_iter.c:3215
#4: ffff88805df266d0 (&c->gc_lock){.+.+}-{3:3}, at: bch2_btree_update_start+0x682/0x14e0 fs/bcachefs/btree_update_interior.c:1202
2 locks held by syz.1.2/5349:
#0: ffff888027c2e420 (sb_writers#16){.+.+}-{0:0}, at: file_start_write include/linux/fs.h:2931 [inline]
#0: ffff888027c2e420 (sb_writers#16){.+.+}-{0:0}, at: vfs_fallocate+0x4fe/0x6e0 fs/open.c:332
#1: ffff88805cd7a6a0 (&sb->s_type->i_mutex_key#25){++++}-{3:3}, at: inode_lock include/linux/fs.h:815 [inline]
#1: ffff88805cd7a6a0 (&sb->s_type->i_mutex_key#25){++++}-{3:3}, at: bch2_fallocate_dispatch+0x1e2/0x540 fs/bcachefs/fs-io.c:762
2 locks held by bch-reclaim/loo/5343:
1 lock held by syz-executor/5371:
#0: ffffffff8e93d338 (rcu_state.exp_mutex){+.+.}-{3:3}, at: exp_funnel_lock kernel/rcu/tree_exp.h:297 [inline]
#0: ffffffff8e93d338 (rcu_state.exp_mutex){+.+.}-{3:3}, at: synchronize_rcu_expedited+0x381/0x830 kernel/rcu/tree_exp.h:976
1 lock held by syz.1.109/6071:
#0: ffff888027c2e0e0 (&type->s_umount_key#59){++++}-{3:3}, at: __super_lock fs/super.c:56 [inline]
#0: ffff888027c2e0e0 (&type->s_umount_key#59){++++}-{3:3}, at: super_lock+0x196/0x400 fs/super.c:120
1 lock held by syz.1.121/6537:
#0: ffff888027c2e0e0 (&type->s_umount_key#59){++++}-{3:3}, at: __super_lock fs/super.c:58 [inline]
#0: ffff888027c2e0e0 (&type->s_umount_key#59){++++}-{3:3}, at: super_lock+0x27c/0x400 fs/super.c:120
1 lock held by syz.1.198/6789:
#0: ffff888027c2e0e0 (&type->s_umount_key#59){++++}-{3:3}, at: __super_lock fs/super.c:58 [inline]
#0: ffff888027c2e0e0 (&type->s_umount_key#59){++++}-{3:3}, at: super_lock+0x27c/0x400 fs/super.c:120
2 locks held by syz.2.202/6807:
1 lock held by syz.3.203/6809:
=============================================
NMI backtrace for cpu 1
CPU: 1 UID: 0 PID: 30 Comm: khungtaskd Not tainted 6.11.0-syzkaller-11653-g3630400697a3 #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 09/13/2024
Call Trace:
__dump_stack lib/dump_stack.c:94 [inline]
dump_stack_lvl+0x241/0x360 lib/dump_stack.c:120
nmi_cpu_backtrace+0x49c/0x4d0 lib/nmi_backtrace.c:113
nmi_trigger_cpumask_backtrace+0x198/0x320 lib/nmi_backtrace.c:62
trigger_all_cpu_backtrace include/linux/nmi.h:162 [inline]
check_hung_uninterruptible_tasks kernel/hung_task.c:223 [inline]
watchdog+0xff4/0x1040 kernel/hung_task.c:379
kthread+0x2f0/0x390 kernel/kthread.c:389
ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244
Sending NMI from CPU 1 to CPUs 0:
NMI backtrace for cpu 0
CPU: 0 UID: 0 PID: 5343 Comm: bch-reclaim/loo Not tainted 6.11.0-syzkaller-11653-g3630400697a3 #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 09/13/2024
RIP: 0010:check_wait_context kernel/locking/lockdep.c:4849 [inline]
RIP: 0010:__lock_acquire+0x92c/0x2050 kernel/locking/lockdep.c:5149
Code: 0e 00 75 23 90 48 c7 c7 60 c8 0a 8c 48 c7 c6 00 cb 0a 8c e8 46 83 e5 ff 48 ba 00 00 00 00 00 fc ff df 90 0f 0b 90 90 90 31 db <48> 8d ab c4 00 00 00 48 89 e8 48 c1 e8 03 0f b6 04 10 84 c0 0f 85
RSP: 0018:ffffc90004617010 EFLAGS: 00000006
RAX: 000000000006a018 RBX: ffffffff93c716d8 RCX: ffffffff81702360
RDX: dffffc0000000000 RSI: 0000000000000008 RDI: ffffffff9423f8c8
RBP: 00000000ffffffff R08: ffffffff9423f8cf R09: 1ffffffff2847f19
R10: dffffc0000000000 R11: fffffbfff2847f1a R12: 0000000000000000
R13: ffff88802d424700 R14: 0000000000000000 R15: ffff88802d424700
FS: 0000000000000000(0000) GS:ffff8880b8600000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 00007f03143ff000 CR3: 000000000e734000 CR4: 00000000003506f0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
Call Trace:
lock_acquire+0x1ed/0x550 kernel/locking/lockdep.c:5822
__raw_spin_lock_irqsave include/linux/spinlock_api_smp.h:110 [inline]
_raw_spin_lock_irqsave+0xd5/0x120 kernel/locking/spinlock.c:162
__bch2_time_stats_update+0x1b4/0x370 fs/bcachefs/time_stats.c:127
bch2_trans_begin+0x785/0x1c00 fs/bcachefs/btree_iter.c:3094
bch2_btree_write_buffer_flush_locked+0x15e/0x59f0 fs/bcachefs/btree_write_buffer.c:281
btree_write_buffer_flush_seq+0x1a43/0x1bc0 fs/bcachefs/btree_write_buffer.c:510
bch2_btree_write_buffer_journal_flush+0x4e/0x80 fs/bcachefs/btree_write_buffer.c:525
journal_flush_pins+0x5f7/0xb20 fs/bcachefs/journal_reclaim.c:565
__bch2_journal_reclaim+0x789/0xdc0 fs/bcachefs/journal_reclaim.c:698
bch2_journal_reclaim_thread+0x174/0x560 fs/bcachefs/journal_reclaim.c:740
kthread+0x2f0/0x390 kernel/kthread.c:389
ret_from_fork+0x4b/0x80 arch/x86/kernel/process.c:147
ret_from_fork_asm+0x1a/0x30 arch/x86/entry/entry_64.S:244