====================================================== WARNING: possible circular locking dependency detected 5.15.181-syzkaller #0 Not tainted ------------------------------------------------------ syz.4.501/6352 is trying to acquire lock: ffff0000ee81c488 (&HFSPLUS_I(inode)->extents_lock){+.+.}-{3:3}, at: hfsplus_get_block+0x2cc/0x11c8 fs/hfsplus/extents.c:260 but task is already holding lock: ffff0000d5fba8f8 (&sbi->alloc_mutex){+.+.}-{3:3}, at: hfsplus_block_allocate+0x88/0x7bc fs/hfsplus/bitmap.c:35 which lock already depends on the new lock. the existing dependency chain (in reverse order) is: -> #1 (&sbi->alloc_mutex){+.+.}-{3:3}: __mutex_lock_common+0x194/0x1edc kernel/locking/mutex.c:596 __mutex_lock kernel/locking/mutex.c:729 [inline] mutex_lock_nested+0xac/0x11c kernel/locking/mutex.c:743 hfsplus_block_allocate+0x88/0x7bc fs/hfsplus/bitmap.c:35 hfsplus_file_extend+0x738/0x12f4 fs/hfsplus/extents.c:469 hfsplus_bmap_reserve+0xe8/0x494 fs/hfsplus/btree.c:357 hfsplus_create_cat+0x170/0xdd4 fs/hfsplus/catalog.c:272 hfsplus_fill_super+0xf8c/0x16dc fs/hfsplus/super.c:560 mount_bdev+0x264/0x358 fs/super.c:1400 hfsplus_mount+0x44/0x58 fs/hfsplus/super.c:641 legacy_get_tree+0xd4/0x16c fs/fs_context.c:611 vfs_get_tree+0x90/0x274 fs/super.c:1530 do_new_mount+0x228/0x810 fs/namespace.c:3013 path_mount+0x5b4/0x1000 fs/namespace.c:3343 do_mount fs/namespace.c:3356 [inline] __do_sys_mount fs/namespace.c:3564 [inline] __se_sys_mount fs/namespace.c:3541 [inline] __arm64_sys_mount+0x514/0x5e4 fs/namespace.c:3541 __invoke_syscall arch/arm64/kernel/syscall.c:38 [inline] invoke_syscall+0x98/0x2b8 arch/arm64/kernel/syscall.c:52 el0_svc_common+0x138/0x258 arch/arm64/kernel/syscall.c:142 do_el0_svc+0x58/0x14c arch/arm64/kernel/syscall.c:181 el0_svc+0x78/0x1e0 arch/arm64/kernel/entry-common.c:608 el0t_64_sync_handler+0xcc/0xe4 arch/arm64/kernel/entry-common.c:626 el0t_64_sync+0x1a0/0x1a4 arch/arm64/kernel/entry.S:584 -> #0 (&HFSPLUS_I(inode)->extents_lock){+.+.}-{3:3}: check_prev_add kernel/locking/lockdep.c:3053 [inline] check_prevs_add kernel/locking/lockdep.c:3172 [inline] validate_chain kernel/locking/lockdep.c:3788 [inline] __lock_acquire+0x2928/0x651c kernel/locking/lockdep.c:5012 lock_acquire+0x1f4/0x620 kernel/locking/lockdep.c:5623 __mutex_lock_common+0x194/0x1edc kernel/locking/mutex.c:596 __mutex_lock kernel/locking/mutex.c:729 [inline] mutex_lock_nested+0xac/0x11c kernel/locking/mutex.c:743 hfsplus_get_block+0x2cc/0x11c8 fs/hfsplus/extents.c:260 block_read_full_page+0x298/0xc40 fs/buffer.c:2290 hfsplus_readpage+0x28/0x38 fs/hfsplus/inode.c:28 do_read_cache_page+0x5f4/0x8f8 mm/filemap.c:-1 read_cache_page+0x68/0x88 mm/filemap.c:3574 read_mapping_page include/linux/pagemap.h:515 [inline] hfsplus_block_allocate+0xd8/0x7bc fs/hfsplus/bitmap.c:37 hfsplus_file_extend+0x738/0x12f4 fs/hfsplus/extents.c:469 hfsplus_get_block+0x324/0x11c8 fs/hfsplus/extents.c:245 __block_write_begin_int+0x3e4/0x1588 fs/buffer.c:2012 __block_write_begin fs/buffer.c:2062 [inline] block_write_begin fs/buffer.c:2122 [inline] cont_write_begin+0x4e4/0x6dc fs/buffer.c:2471 hfsplus_write_begin+0xa8/0xf8 fs/hfsplus/inode.c:53 pagecache_write_begin+0xa0/0xc0 mm/filemap.c:3608 __page_symlink+0x114/0x23c fs/namei.c:5191 page_symlink+0x88/0xac fs/namei.c:5214 hfsplus_symlink+0xb8/0x214 fs/hfsplus/dir.c:449 vfs_symlink+0x238/0x3b0 fs/namei.c:4429 do_symlinkat+0x184/0x5a8 fs/namei.c:4458 __do_sys_symlinkat fs/namei.c:4475 [inline] __se_sys_symlinkat fs/namei.c:4472 [inline] __arm64_sys_symlinkat+0xa4/0xbc fs/namei.c:4472 __invoke_syscall arch/arm64/kernel/syscall.c:38 [inline] invoke_syscall+0x98/0x2b8 arch/arm64/kernel/syscall.c:52 el0_svc_common+0x138/0x258 arch/arm64/kernel/syscall.c:142 do_el0_svc+0x58/0x14c arch/arm64/kernel/syscall.c:181 el0_svc+0x78/0x1e0 arch/arm64/kernel/entry-common.c:608 el0t_64_sync_handler+0xcc/0xe4 arch/arm64/kernel/entry-common.c:626 el0t_64_sync+0x1a0/0x1a4 arch/arm64/kernel/entry.S:584 other info that might help us debug this: Possible unsafe locking scenario: CPU0 CPU1 ---- ---- lock(&sbi->alloc_mutex); lock(&HFSPLUS_I(inode)->extents_lock); lock(&sbi->alloc_mutex); lock(&HFSPLUS_I(inode)->extents_lock); *** DEADLOCK *** 5 locks held by syz.4.501/6352: #0: ffff0000e203c460 (sb_writers#26){.+.+}-{0:0}, at: mnt_want_write+0x44/0x9c fs/namespace.c:377 #1: ffff0000ee81cd40 (&type->i_mutex_dir_key#13/1){+.+.}-{3:3}, at: inode_lock_nested include/linux/fs.h:822 [inline] #1: ffff0000ee81cd40 (&type->i_mutex_dir_key#13/1){+.+.}-{3:3}, at: filename_create+0x1ac/0x39c fs/namei.c:3835 #2: ffff0000d5fba998 (&sbi->vh_mutex){+.+.}-{3:3}, at: hfsplus_symlink+0x78/0x214 fs/hfsplus/dir.c:444 #3: ffff0000ee81d8c8 (&hip->extents_lock){+.+.}-{3:3}, at: hfsplus_file_extend+0x180/0x12f4 fs/hfsplus/extents.c:458 #4: ffff0000d5fba8f8 (&sbi->alloc_mutex){+.+.}-{3:3}, at: hfsplus_block_allocate+0x88/0x7bc fs/hfsplus/bitmap.c:35 stack backtrace: CPU: 0 PID: 6352 Comm: syz.4.501 Not tainted 5.15.181-syzkaller #0 Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 04/19/2025 Call trace: dump_backtrace+0x0/0x43c arch/arm64/kernel/stacktrace.c:152 show_stack+0x2c/0x3c arch/arm64/kernel/stacktrace.c:216 __dump_stack+0x30/0x40 lib/dump_stack.c:88 dump_stack_lvl+0xf8/0x160 lib/dump_stack.c:106 dump_stack+0x1c/0x5c lib/dump_stack.c:113 print_circular_bug+0x148/0x1b0 kernel/locking/lockdep.c:2011 check_noncircular+0x240/0x2d4 kernel/locking/lockdep.c:2133 check_prev_add kernel/locking/lockdep.c:3053 [inline] check_prevs_add kernel/locking/lockdep.c:3172 [inline] validate_chain kernel/locking/lockdep.c:3788 [inline] __lock_acquire+0x2928/0x651c kernel/locking/lockdep.c:5012 lock_acquire+0x1f4/0x620 kernel/locking/lockdep.c:5623 __mutex_lock_common+0x194/0x1edc kernel/locking/mutex.c:596 __mutex_lock kernel/locking/mutex.c:729 [inline] mutex_lock_nested+0xac/0x11c kernel/locking/mutex.c:743 hfsplus_get_block+0x2cc/0x11c8 fs/hfsplus/extents.c:260 block_read_full_page+0x298/0xc40 fs/buffer.c:2290 hfsplus_readpage+0x28/0x38 fs/hfsplus/inode.c:28 do_read_cache_page+0x5f4/0x8f8 mm/filemap.c:-1 read_cache_page+0x68/0x88 mm/filemap.c:3574 read_mapping_page include/linux/pagemap.h:515 [inline] hfsplus_block_allocate+0xd8/0x7bc fs/hfsplus/bitmap.c:37 hfsplus_file_extend+0x738/0x12f4 fs/hfsplus/extents.c:469 hfsplus_get_block+0x324/0x11c8 fs/hfsplus/extents.c:245 __block_write_begin_int+0x3e4/0x1588 fs/buffer.c:2012 __block_write_begin fs/buffer.c:2062 [inline] block_write_begin fs/buffer.c:2122 [inline] cont_write_begin+0x4e4/0x6dc fs/buffer.c:2471 hfsplus_write_begin+0xa8/0xf8 fs/hfsplus/inode.c:53 pagecache_write_begin+0xa0/0xc0 mm/filemap.c:3608 __page_symlink+0x114/0x23c fs/namei.c:5191 page_symlink+0x88/0xac fs/namei.c:5214 hfsplus_symlink+0xb8/0x214 fs/hfsplus/dir.c:449 vfs_symlink+0x238/0x3b0 fs/namei.c:4429 do_symlinkat+0x184/0x5a8 fs/namei.c:4458 __do_sys_symlinkat fs/namei.c:4475 [inline] __se_sys_symlinkat fs/namei.c:4472 [inline] __arm64_sys_symlinkat+0xa4/0xbc fs/namei.c:4472 __invoke_syscall arch/arm64/kernel/syscall.c:38 [inline] invoke_syscall+0x98/0x2b8 arch/arm64/kernel/syscall.c:52 el0_svc_common+0x138/0x258 arch/arm64/kernel/syscall.c:142 do_el0_svc+0x58/0x14c arch/arm64/kernel/syscall.c:181 el0_svc+0x78/0x1e0 arch/arm64/kernel/entry-common.c:608 el0t_64_sync_handler+0xcc/0xe4 arch/arm64/kernel/entry-common.c:626 el0t_64_sync+0x1a0/0x1a4 arch/arm64/kernel/entry.S:584