INFO: task syz-executor:6419 blocked for more than 143 seconds. Not tainted 6.11.0-rc7-syzkaller-g5f5673607153 #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:syz-executor state:D stack:0 pid:6419 tgid:6419 ppid:1 flags:0x0000000d Call trace: __switch_to+0x420/0x6dc arch/arm64/kernel/process.c:603 context_switch kernel/sched/core.c:5188 [inline] __schedule+0x13d4/0x2418 kernel/sched/core.c:6529 __schedule_loop kernel/sched/core.c:6606 [inline] schedule+0xbc/0x238 kernel/sched/core.c:6621 bch2_fs_read_only+0x87c/0xf14 fs/bcachefs/super.c:346 __bch2_fs_stop+0x100/0x510 fs/bcachefs/super.c:618 bch2_put_super+0x40/0x50 fs/bcachefs/fs.c:1826 generic_shutdown_super+0x128/0x2b8 fs/super.c:642 bch2_kill_sb+0x40/0x58 fs/bcachefs/fs.c:2054 deactivate_locked_super+0xc4/0x12c fs/super.c:473 deactivate_super+0xe0/0x100 fs/super.c:506 cleanup_mnt+0x34c/0x3dc fs/namespace.c:1373 __cleanup_mnt+0x20/0x30 fs/namespace.c:1380 task_work_run+0x230/0x2e0 kernel/task_work.c:228 resume_user_mode_work include/linux/resume_user_mode.h:50 [inline] do_notify_resume+0x178/0x1f4 arch/arm64/kernel/entry-common.c:151 exit_to_user_mode_prepare arch/arm64/kernel/entry-common.c:169 [inline] exit_to_user_mode arch/arm64/kernel/entry-common.c:178 [inline] el0_svc+0xac/0x168 arch/arm64/kernel/entry-common.c:713 el0t_64_sync_handler+0x84/0xfc arch/arm64/kernel/entry-common.c:730 el0t_64_sync+0x190/0x194 arch/arm64/kernel/entry.S:598 INFO: task syz.1.485:7670 blocked for more than 143 seconds. Not tainted 6.11.0-rc7-syzkaller-g5f5673607153 #0 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. task:syz.1.485 state:D stack:0 pid:7670 tgid:7668 ppid:6418 flags:0x00000009 Call trace: __switch_to+0x420/0x6dc arch/arm64/kernel/process.c:603 context_switch kernel/sched/core.c:5188 [inline] __schedule+0x13d4/0x2418 kernel/sched/core.c:6529 __schedule_loop kernel/sched/core.c:6606 [inline] schedule+0xbc/0x238 kernel/sched/core.c:6621 schedule_preempt_disabled+0x18/0x2c kernel/sched/core.c:6678 rwsem_down_read_slowpath+0x534/0x858 kernel/locking/rwsem.c:1086 __down_read_common kernel/locking/rwsem.c:1250 [inline] __down_read kernel/locking/rwsem.c:1263 [inline] down_read+0xa0/0x2fc kernel/locking/rwsem.c:1528 __super_lock fs/super.c:58 [inline] super_lock+0x244/0x33c fs/super.c:120 super_lock_shared fs/super.c:139 [inline] iterate_supers+0xa4/0x4dc fs/super.c:931 quota_sync_all fs/quota/quota.c:69 [inline] __do_sys_quotactl fs/quota/quota.c:938 [inline] __se_sys_quotactl fs/quota/quota.c:917 [inline] __arm64_sys_quotactl+0x388/0xc9c fs/quota/quota.c:917 __invoke_syscall arch/arm64/kernel/syscall.c:35 [inline] invoke_syscall+0x98/0x2b8 arch/arm64/kernel/syscall.c:49 el0_svc_common+0x130/0x23c arch/arm64/kernel/syscall.c:132 do_el0_svc+0x48/0x58 arch/arm64/kernel/syscall.c:151 el0_svc+0x54/0x168 arch/arm64/kernel/entry-common.c:712 el0t_64_sync_handler+0x84/0xfc arch/arm64/kernel/entry-common.c:730 el0t_64_sync+0x190/0x194 arch/arm64/kernel/entry.S:598 Showing all locks held in the system: 4 locks held by kworker/u8:1/13: #0: ffff0000d8970148 ((wq_completion)btree_node_rewrite){+.+.}-{0:0}, at: process_one_work+0x624/0x15b8 kernel/workqueue.c:3205 #1: ffff8000975d7c20 ((work_completion)(&a->work)){+.+.}-{0:0}, at: process_one_work+0x6a0/0x15b8 kernel/workqueue.c:3205 #2: ffff0000f1e042c0 (&c->btree_trans_barrier){.+.+}-{0:0}, at: srcu_lock_acquire+0x18/0x54 include/linux/srcu.h:150 #3: ffff0000f1e26690 (&c->gc_lock){++++}-{3:3}, at: bch2_btree_update_start+0x588/0x15ac fs/bcachefs/btree_update_interior.c:1197 1 lock held by khungtaskd/31: #0: ffff80008f74dfa0 (rcu_read_lock){....}-{1:2}, at: rcu_lock_acquire+0xc/0x44 include/linux/rcupdate.h:325 5 locks held by kworker/u8:7/2036: #0: ffff0000c1869948 ((wq_completion)netns){+.+.}-{0:0}, at: process_one_work+0x624/0x15b8 kernel/workqueue.c:3205 #1: ffff80009f7e7c20 (net_cleanup_work){+.+.}-{0:0}, at: process_one_work+0x6a0/0x15b8 kernel/workqueue.c:3205 #2: ffff80009242b950 (pernet_ops_rwsem){++++}-{3:3}, at: cleanup_net+0x13c/0x9b4 net/core/net_namespace.c:594 #3: ffff000141691428 (&wg->device_update_lock){+.+.}-{3:3}, at: wg_destruct+0x11c/0x2c0 drivers/net/wireguard/device.c:249 #4: ffff80008f7531c0 (rcu_state.barrier_mutex){+.+.}-{3:3}, at: rcu_barrier+0x58/0x534 kernel/rcu/tree.c:4486 1 lock held by syslogd/6012: #0: ffff0001b36ae698 (&rq->__lock){-.-.}-{2:2}, at: raw_spin_rq_lock_nested kernel/sched/core.c:560 [inline] #0: ffff0001b36ae698 (&rq->__lock){-.-.}-{2:2}, at: raw_spin_rq_lock kernel/sched/sched.h:1415 [inline] #0: ffff0001b36ae698 (&rq->__lock){-.-.}-{2:2}, at: rq_lock kernel/sched/sched.h:1714 [inline] #0: ffff0001b36ae698 (&rq->__lock){-.-.}-{2:2}, at: __schedule+0x2d8/0x2418 kernel/sched/core.c:6436 1 lock held by udevd/6027: 2 locks held by getty/6175: #0: ffff0000d6fc20a0 (&tty->ldisc_sem){++++}-{0:0}, at: ldsem_down_read+0x3c/0x4c drivers/tty/tty_ldsem.c:340 #1: ffff80009b50e2f0 (&ldata->atomic_read_lock){+.+.}-{3:3}, at: n_tty_read+0x41c/0x1228 drivers/tty/n_tty.c:2211 3 locks held by kworker/1:3/6407: #0: ffff0000c0028948 ((wq_completion)events){+.+.}-{0:0}, at: process_one_work+0x624/0x15b8 kernel/workqueue.c:3205 #1: ffff80009c997c20 ((work_completion)(&data->fib_event_work)){+.+.}-{0:0}, at: process_one_work+0x6a0/0x15b8 kernel/workqueue.c:3205 #2: ffff000153dfe240 (&data->fib_lock){+.+.}-{3:3}, at: nsim_fib_event_work+0x270/0x32bc drivers/net/netdevsim/fib.c:1489 2 locks held by syz-executor/6419: #0: ffff0000ee8a80e0 (&type->s_umount_key#71){++++}-{3:3}, at: __super_lock fs/super.c:56 [inline] #0: ffff0000ee8a80e0 (&type->s_umount_key#71){++++}-{3:3}, at: __super_lock_excl fs/super.c:71 [inline] #0: ffff0000ee8a80e0 (&type->s_umount_key#71){++++}-{3:3}, at: deactivate_super+0xd8/0x100 fs/super.c:505 #1: ffff0000f1e00278 (&c->state_lock){+.+.}-{3:3}, at: __bch2_fs_stop+0xf8/0x510 fs/bcachefs/super.c:617 3 locks held by kworker/0:4/6473: #0: ffff0000c0028948 ((wq_completion)events){+.+.}-{0:0}, at: process_one_work+0x624/0x15b8 kernel/workqueue.c:3205 #1: ffff80009c507c20 ((work_completion)(&data->fib_event_work)){+.+.}-{0:0}, at: process_one_work+0x6a0/0x15b8 kernel/workqueue.c:3205 #2: ffff0000ec999240 (&data->fib_lock){+.+.}-{3:3}, at: nsim_fib_event_work+0x270/0x32bc drivers/net/netdevsim/fib.c:1489 1 lock held by syz.1.485/7670: #0: ffff0000ee8a80e0 (&type->s_umount_key#71){++++}-{3:3}, at: __super_lock fs/super.c:58 [inline] #0: ffff0000ee8a80e0 (&type->s_umount_key#71){++++}-{3:3}, at: super_lock+0x244/0x33c fs/super.c:120 2 locks held by kworker/u8:10/8753: #0: ffff0000c257d148 ((wq_completion)iou_exit){+.+.}-{0:0}, at: process_one_work+0x624/0x15b8 kernel/workqueue.c:3205 #1: ffff80009f427c20 ((work_completion)(&ctx->exit_work)){+.+.}-{0:0}, at: process_one_work+0x6a0/0x15b8 kernel/workqueue.c:3205 2 locks held by syz-executor/8772: #0: ffff0001b36ae698 (&rq->__lock){-.-.}-{2:2}, at: raw_spin_rq_lock_nested kernel/sched/core.c:560 [inline] #0: ffff0001b36ae698 (&rq->__lock){-.-.}-{2:2}, at: raw_spin_rq_lock kernel/sched/sched.h:1415 [inline] #0: ffff0001b36ae698 (&rq->__lock){-.-.}-{2:2}, at: rq_lock kernel/sched/sched.h:1714 [inline] #0: ffff0001b36ae698 (&rq->__lock){-.-.}-{2:2}, at: __schedule+0x2d8/0x2418 kernel/sched/core.c:6436 #1: ffff0001b3698f48 (&per_cpu_ptr(group->pcpu, cpu)->seq){-.-.}-{0:0}, at: psi_task_switch+0x3c0/0x618 kernel/sched/psi.c:989 1 lock held by syz.4.1160/9615: #0: ffff0000ce0b00e0 (&type->s_umount_key#66/1){+.+.}-{3:3}, at: alloc_super+0x1b0/0x83c fs/super.c:344 2 locks held by syz.0.1168/9635: 1 lock held by dhcpcd/9648: #0: ffff0000dd1ce208 (&sb->s_type->i_mutex_key#10){+.+.}-{3:3}, at: inode_lock include/linux/fs.h:800 [inline] #0: ffff0000dd1ce208 (&sb->s_type->i_mutex_key#10){+.+.}-{3:3}, at: __sock_release net/socket.c:658 [inline] #0: ffff0000dd1ce208 (&sb->s_type->i_mutex_key#10){+.+.}-{3:3}, at: sock_close+0x80/0x1e8 net/socket.c:1421 1 lock held by dhcpcd/9654: #0: ffff0000dc7ca258 (sk_lock-AF_PACKET){+.+.}-{0:0}, at: lock_sock include/net/sock.h:1607 [inline] #0: ffff0000dc7ca258 (sk_lock-AF_PACKET){+.+.}-{0:0}, at: packet_do_bind+0x4c/0xb2c net/packet/af_packet.c:3266 3 locks held by syz.3.1174/9658: =============================================