====================================================== WARNING: possible circular locking dependency detected 6.1.23-syzkaller #0 Not tainted ------------------------------------------------------ syz-executor.2/15114 is trying to acquire lock: ffff0000d95138f8 (&sbi->alloc_mutex){+.+.}-{3:3}, at: hfsplus_block_allocate+0x90/0x818 fs/hfsplus/bitmap.c:35 but task is already holding lock: ffff0001057122c8 (&HFSPLUS_I(inode)->extents_lock){+.+.}-{3:3}, at: hfsplus_file_extend+0x198/0x14cc fs/hfsplus/extents.c:457 which lock already depends on the new lock. the existing dependency chain (in reverse order) is: -> #1 (&HFSPLUS_I(inode)->extents_lock){+.+.}-{3:3}: __mutex_lock_common+0x190/0x21a0 kernel/locking/mutex.c:603 __mutex_lock kernel/locking/mutex.c:747 [inline] mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799 hfsplus_get_block+0x2c4/0x1168 fs/hfsplus/extents.c:260 block_read_full_folio+0x2f4/0x98c fs/buffer.c:2271 hfsplus_read_folio+0x28/0x38 fs/hfsplus/inode.c:28 filemap_read_folio+0x14c/0x39c mm/filemap.c:2407 do_read_cache_folio+0x24c/0x544 mm/filemap.c:3535 do_read_cache_page mm/filemap.c:3577 [inline] read_cache_page+0x6c/0x180 mm/filemap.c:3586 read_mapping_page include/linux/pagemap.h:756 [inline] hfsplus_block_allocate+0xe0/0x818 fs/hfsplus/bitmap.c:37 hfsplus_file_extend+0x770/0x14cc fs/hfsplus/extents.c:468 hfsplus_get_block+0x398/0x1168 fs/hfsplus/extents.c:245 __block_write_begin_int+0x340/0x13b4 fs/buffer.c:1991 __block_write_begin fs/buffer.c:2041 [inline] block_write_begin fs/buffer.c:2102 [inline] cont_write_begin+0x5c0/0x7d8 fs/buffer.c:2456 hfsplus_write_begin+0x98/0xe4 fs/hfsplus/inode.c:52 page_symlink+0x278/0x4a4 fs/namei.c:5121 hfsplus_symlink+0xc0/0x224 fs/hfsplus/dir.c:449 vfs_symlink+0x244/0x3a8 fs/namei.c:4401 do_symlinkat+0x364/0x6b0 fs/namei.c:4430 __do_sys_symlinkat fs/namei.c:4447 [inline] __se_sys_symlinkat fs/namei.c:4444 [inline] __arm64_sys_symlinkat+0xa4/0xbc fs/namei.c:4444 __invoke_syscall arch/arm64/kernel/syscall.c:38 [inline] invoke_syscall+0x98/0x2c0 arch/arm64/kernel/syscall.c:52 el0_svc_common+0x138/0x258 arch/arm64/kernel/syscall.c:142 do_el0_svc+0x64/0x218 arch/arm64/kernel/syscall.c:206 el0_svc+0x58/0x168 arch/arm64/kernel/entry-common.c:637 el0t_64_sync_handler+0x84/0xf0 arch/arm64/kernel/entry-common.c:655 el0t_64_sync+0x18c/0x190 arch/arm64/kernel/entry.S:581 -> #0 (&sbi->alloc_mutex){+.+.}-{3:3}: check_prev_add kernel/locking/lockdep.c:3098 [inline] check_prevs_add kernel/locking/lockdep.c:3217 [inline] validate_chain kernel/locking/lockdep.c:3832 [inline] __lock_acquire+0x3338/0x764c kernel/locking/lockdep.c:5056 lock_acquire+0x26c/0x7cc kernel/locking/lockdep.c:5669 __mutex_lock_common+0x190/0x21a0 kernel/locking/mutex.c:603 __mutex_lock kernel/locking/mutex.c:747 [inline] mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799 hfsplus_block_allocate+0x90/0x818 fs/hfsplus/bitmap.c:35 hfsplus_file_extend+0x770/0x14cc fs/hfsplus/extents.c:468 hfsplus_bmap_reserve+0xec/0x474 fs/hfsplus/btree.c:358 hfsplus_create_cat+0x18c/0x1330 fs/hfsplus/catalog.c:272 hfsplus_mknod+0x14c/0x268 fs/hfsplus/dir.c:494 hfsplus_create+0x38/0x4c fs/hfsplus/dir.c:523 lookup_open fs/namei.c:3413 [inline] open_last_lookups fs/namei.c:3481 [inline] path_openat+0xeac/0x2548 fs/namei.c:3711 do_filp_open+0x1bc/0x3cc fs/namei.c:3741 do_sys_openat2+0x128/0x3d8 fs/open.c:1310 do_sys_open fs/open.c:1326 [inline] __do_sys_openat fs/open.c:1342 [inline] __se_sys_openat fs/open.c:1337 [inline] __arm64_sys_openat+0x1f0/0x240 fs/open.c:1337 __invoke_syscall arch/arm64/kernel/syscall.c:38 [inline] invoke_syscall+0x98/0x2c0 arch/arm64/kernel/syscall.c:52 el0_svc_common+0x138/0x258 arch/arm64/kernel/syscall.c:142 do_el0_svc+0x64/0x218 arch/arm64/kernel/syscall.c:206 el0_svc+0x58/0x168 arch/arm64/kernel/entry-common.c:637 el0t_64_sync_handler+0x84/0xf0 arch/arm64/kernel/entry-common.c:655 el0t_64_sync+0x18c/0x190 arch/arm64/kernel/entry.S:581 other info that might help us debug this: Possible unsafe locking scenario: CPU0 CPU1 ---- ---- lock(&HFSPLUS_I(inode)->extents_lock); lock(&sbi->alloc_mutex); lock(&HFSPLUS_I(inode)->extents_lock); lock(&sbi->alloc_mutex); *** DEADLOCK *** 5 locks held by syz-executor.2/15114: #0: ffff0000d674a460 (sb_writers#12){.+.+}-{0:0}, at: mnt_want_write+0x44/0x9c fs/namespace.c:393 #1: ffff000105711740 (&type->i_mutex_dir_key#9){++++}-{3:3}, at: inode_lock include/linux/fs.h:756 [inline] #1: ffff000105711740 (&type->i_mutex_dir_key#9){++++}-{3:3}, at: open_last_lookups fs/namei.c:3478 [inline] #1: ffff000105711740 (&type->i_mutex_dir_key#9){++++}-{3:3}, at: path_openat+0x5ec/0x2548 fs/namei.c:3711 #2: ffff0000d9513998 (&sbi->vh_mutex){+.+.}-{3:3}, at: hfsplus_mknod+0x84/0x268 fs/hfsplus/dir.c:486 #3: ffff0000cbb3e0b0 (&tree->tree_lock){+.+.}-{3:3}, at: hfsplus_find_init+0x144/0x1bc #4: ffff0001057122c8 (&HFSPLUS_I(inode)->extents_lock){+.+.}-{3:3}, at: hfsplus_file_extend+0x198/0x14cc fs/hfsplus/extents.c:457 stack backtrace: CPU: 1 PID: 15114 Comm: syz-executor.2 Not tainted 6.1.23-syzkaller #0 Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 03/02/2023 Call trace: dump_backtrace+0x1c8/0x1f4 arch/arm64/kernel/stacktrace.c:158 show_stack+0x2c/0x3c arch/arm64/kernel/stacktrace.c:165 __dump_stack lib/dump_stack.c:88 [inline] dump_stack_lvl+0x108/0x170 lib/dump_stack.c:106 dump_stack+0x1c/0x5c lib/dump_stack.c:113 print_circular_bug+0x150/0x1b8 kernel/locking/lockdep.c:2056 check_noncircular+0x2cc/0x378 kernel/locking/lockdep.c:2178 check_prev_add kernel/locking/lockdep.c:3098 [inline] check_prevs_add kernel/locking/lockdep.c:3217 [inline] validate_chain kernel/locking/lockdep.c:3832 [inline] __lock_acquire+0x3338/0x764c kernel/locking/lockdep.c:5056 lock_acquire+0x26c/0x7cc kernel/locking/lockdep.c:5669 __mutex_lock_common+0x190/0x21a0 kernel/locking/mutex.c:603 __mutex_lock kernel/locking/mutex.c:747 [inline] mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799 hfsplus_block_allocate+0x90/0x818 fs/hfsplus/bitmap.c:35 hfsplus_file_extend+0x770/0x14cc fs/hfsplus/extents.c:468 hfsplus_bmap_reserve+0xec/0x474 fs/hfsplus/btree.c:358 hfsplus_create_cat+0x18c/0x1330 fs/hfsplus/catalog.c:272 hfsplus_mknod+0x14c/0x268 fs/hfsplus/dir.c:494 hfsplus_create+0x38/0x4c fs/hfsplus/dir.c:523 lookup_open fs/namei.c:3413 [inline] open_last_lookups fs/namei.c:3481 [inline] path_openat+0xeac/0x2548 fs/namei.c:3711 do_filp_open+0x1bc/0x3cc fs/namei.c:3741 do_sys_openat2+0x128/0x3d8 fs/open.c:1310 do_sys_open fs/open.c:1326 [inline] __do_sys_openat fs/open.c:1342 [inline] __se_sys_openat fs/open.c:1337 [inline] __arm64_sys_openat+0x1f0/0x240 fs/open.c:1337 __invoke_syscall arch/arm64/kernel/syscall.c:38 [inline] invoke_syscall+0x98/0x2c0 arch/arm64/kernel/syscall.c:52 el0_svc_common+0x138/0x258 arch/arm64/kernel/syscall.c:142 do_el0_svc+0x64/0x218 arch/arm64/kernel/syscall.c:206 el0_svc+0x58/0x168 arch/arm64/kernel/entry-common.c:637 el0t_64_sync_handler+0x84/0xf0 arch/arm64/kernel/entry-common.c:655 el0t_64_sync+0x18c/0x190 arch/arm64/kernel/entry.S:581 bio_check_eod: 133 callbacks suppressed syz-executor.2: attempt to access beyond end of device loop2: rw=0, sector=201326592, nr_sectors = 2 limit=1024 buffer_io_error: 14 callbacks suppressed Buffer I/O error on dev loop2, logical block 100663296, async page read syz-executor.2: attempt to access beyond end of device loop2: rw=0, sector=201326592, nr_sectors = 2 limit=1024 Buffer I/O error on dev loop2, logical block 100663296, async page read