syzbot


possible deadlock in indx_read (2)

Status: upstream: reported on 2024/05/05 22:47
Subsystems: ntfs3
[Documentation on labels]
Reported-by: syzbot+9309ec5fd67539ad7ccd@syzkaller.appspotmail.com
First crash: 44d, last: 29d
Discussions (1)
Title Replies (including bot) Last reply
[syzbot] [ntfs3?] possible deadlock in indx_read (2) 0 (1) 2024/05/05 22:47
Similar bugs (1)
Kernel Title Repro Cause bisect Fix bisect Count Last Reported Patched Status
upstream possible deadlock in indx_read ntfs3 1 234d 230d 0/27 auto-obsoleted due to no activity on 2023/12/23 22:39

Sample crash report:
======================================================
WARNING: possible circular locking dependency detected
6.9.0-syzkaller-08284-gea5f6ad9ad96 #0 Not tainted
------------------------------------------------------
syz-executor.3/24799 is trying to acquire lock:
ffff88805b7d4440 (&indx->run_lock){++++}-{3:3}, at: indx_read+0x2a5/0xc50 fs/ntfs3/index.c:1066

but task is already holding lock:
ffff88805b7d6fa0 (&ni->ni_lock/4){+.+.}-{3:3}, at: ni_lock fs/ntfs3/ntfs_fs.h:1123 [inline]
ffff88805b7d6fa0 (&ni->ni_lock/4){+.+.}-{3:3}, at: ntfs_rename+0x74a/0xd10 fs/ntfs3/namei.c:329

which lock already depends on the new lock.


the existing dependency chain (in reverse order) is:

-> #3 (&ni->ni_lock/4){+.+.}-{3:3}:
       lock_acquire+0x1ed/0x550 kernel/locking/lockdep.c:5754
       __mutex_lock_common kernel/locking/mutex.c:608 [inline]
       __mutex_lock+0x136/0xd70 kernel/locking/mutex.c:752
       ni_lock fs/ntfs3/ntfs_fs.h:1123 [inline]
       ntfs_fallocate+0xa37/0x1240 fs/ntfs3/file.c:572
       vfs_fallocate+0x564/0x6c0 fs/open.c:330
       ksys_fallocate fs/open.c:353 [inline]
       __do_sys_fallocate fs/open.c:361 [inline]
       __se_sys_fallocate fs/open.c:359 [inline]
       __x64_sys_fallocate+0xbd/0x110 fs/open.c:359
       do_syscall_x64 arch/x86/entry/common.c:52 [inline]
       do_syscall_64+0xf5/0x240 arch/x86/entry/common.c:83
       entry_SYSCALL_64_after_hwframe+0x77/0x7f

-> #2 (mapping.invalidate_lock#11){++++}-{3:3}:
       lock_acquire+0x1ed/0x550 kernel/locking/lockdep.c:5754
       down_read+0xb1/0xa40 kernel/locking/rwsem.c:1526
       filemap_invalidate_lock_shared include/linux/fs.h:846 [inline]
       filemap_fault+0x655/0x1760 mm/filemap.c:3277
       __do_fault+0x135/0x460 mm/memory.c:4526
       do_read_fault mm/memory.c:4889 [inline]
       do_fault mm/memory.c:5019 [inline]
       do_pte_missing mm/memory.c:3875 [inline]
       handle_pte_fault mm/memory.c:5295 [inline]
       __handle_mm_fault+0x45fe/0x7250 mm/memory.c:5436
       handle_mm_fault+0x27f/0x770 mm/memory.c:5601
       do_user_addr_fault arch/x86/mm/fault.c:1383 [inline]
       handle_page_fault arch/x86/mm/fault.c:1475 [inline]
       exc_page_fault+0x2a8/0x8a0 arch/x86/mm/fault.c:1533
       asm_exc_page_fault+0x26/0x30 arch/x86/include/asm/idtentry.h:623
       strncpy_from_user+0x110/0x2f0 lib/strncpy_from_user.c:139
       getname_flags+0xfa/0x4f0 fs/namei.c:150
       user_path_at_empty+0x2c/0x60 fs/namei.c:2928
       user_path_at include/linux/namei.h:58 [inline]
       ksys_umount fs/namespace.c:1916 [inline]
       __do_sys_umount fs/namespace.c:1924 [inline]
       __se_sys_umount fs/namespace.c:1922 [inline]
       __x64_sys_umount+0xf4/0x170 fs/namespace.c:1922
       do_syscall_x64 arch/x86/entry/common.c:52 [inline]
       do_syscall_64+0xf5/0x240 arch/x86/entry/common.c:83
       entry_SYSCALL_64_after_hwframe+0x77/0x7f

-> #1 (&mm->mmap_lock){++++}-{3:3}:
       lock_acquire+0x1ed/0x550 kernel/locking/lockdep.c:5754
       __might_fault+0xc6/0x120 mm/memory.c:6215
       _copy_to_user+0x2a/0xb0 lib/usercopy.c:36
       copy_to_user include/linux/uaccess.h:191 [inline]
       fiemap_fill_next_extent+0x235/0x410 fs/ioctl.c:145
       ni_fiemap+0xa5e/0x1230 fs/ntfs3/frecord.c:2065
       ntfs_fiemap+0x132/0x180 fs/ntfs3/file.c:1206
       ioctl_fiemap fs/ioctl.c:220 [inline]
       do_vfs_ioctl+0x1c07/0x2e50 fs/ioctl.c:838
       __do_sys_ioctl fs/ioctl.c:902 [inline]
       __se_sys_ioctl+0x81/0x170 fs/ioctl.c:890
       do_syscall_x64 arch/x86/entry/common.c:52 [inline]
       do_syscall_64+0xf5/0x240 arch/x86/entry/common.c:83
       entry_SYSCALL_64_after_hwframe+0x77/0x7f

-> #0 (&indx->run_lock){++++}-{3:3}:
       check_prev_add kernel/locking/lockdep.c:3134 [inline]
       check_prevs_add kernel/locking/lockdep.c:3253 [inline]
       validate_chain+0x18cb/0x58e0 kernel/locking/lockdep.c:3869
       __lock_acquire+0x1346/0x1fd0 kernel/locking/lockdep.c:5137
       lock_acquire+0x1ed/0x550 kernel/locking/lockdep.c:5754
       down_read+0xb1/0xa40 kernel/locking/rwsem.c:1526
       indx_read+0x2a5/0xc50 fs/ntfs3/index.c:1066
       indx_find+0x47a/0xbf0 fs/ntfs3/index.c:1181
       indx_insert_entry+0x527/0x790 fs/ntfs3/index.c:1959
       ni_add_name+0xa24/0xe90 fs/ntfs3/frecord.c:3055
       ni_rename+0xc2/0x1e0 fs/ntfs3/frecord.c:3087
       ntfs_rename+0x7c1/0xd10 fs/ntfs3/namei.c:334
       vfs_rename+0xbdb/0xf00 fs/namei.c:4887
       do_renameat2+0xd94/0x13f0 fs/namei.c:5044
       __do_sys_rename fs/namei.c:5091 [inline]
       __se_sys_rename fs/namei.c:5089 [inline]
       __x64_sys_rename+0x86/0xa0 fs/namei.c:5089
       do_syscall_x64 arch/x86/entry/common.c:52 [inline]
       do_syscall_64+0xf5/0x240 arch/x86/entry/common.c:83
       entry_SYSCALL_64_after_hwframe+0x77/0x7f

other info that might help us debug this:

Chain exists of:
  &indx->run_lock --> mapping.invalidate_lock#11 --> &ni->ni_lock/4

 Possible unsafe locking scenario:

       CPU0                    CPU1
       ----                    ----
  lock(&ni->ni_lock/4);
                               lock(mapping.invalidate_lock#11);
                               lock(&ni->ni_lock/4);
  rlock(&indx->run_lock);

 *** DEADLOCK ***

4 locks held by syz-executor.3/24799:
 #0: ffff888068c92420 (sb_writers#28){.+.+}-{0:0}, at: mnt_want_write+0x3f/0x90 fs/namespace.c:409
 #1: ffff88805b7d4600 (&type->i_mutex_dir_key#18/1){+.+.}-{3:3}, at: inode_lock_nested include/linux/fs.h:826 [inline]
 #1: ffff88805b7d4600 (&type->i_mutex_dir_key#18/1){+.+.}-{3:3}, at: lock_rename fs/namei.c:3074 [inline]
 #1: ffff88805b7d4600 (&type->i_mutex_dir_key#18/1){+.+.}-{3:3}, at: do_renameat2+0x62c/0x13f0 fs/namei.c:4979
 #2: ffff88805b7d4360 (&ni->ni_lock/5){+.+.}-{3:3}, at: ni_lock_dir fs/ntfs3/ntfs_fs.h:1128 [inline]
 #2: ffff88805b7d4360 (&ni->ni_lock/5){+.+.}-{3:3}, at: ntfs_rename+0x72f/0xd10 fs/ntfs3/namei.c:328
 #3: ffff88805b7d6fa0 (&ni->ni_lock/4){+.+.}-{3:3}, at: ni_lock fs/ntfs3/ntfs_fs.h:1123 [inline]
 #3: ffff88805b7d6fa0 (&ni->ni_lock/4){+.+.}-{3:3}, at: ntfs_rename+0x74a/0xd10 fs/ntfs3/namei.c:329

stack backtrace:
CPU: 1 PID: 24799 Comm: syz-executor.3 Not tainted 6.9.0-syzkaller-08284-gea5f6ad9ad96 #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 04/02/2024
Call Trace:
 <TASK>
 __dump_stack lib/dump_stack.c:88 [inline]
 dump_stack_lvl+0x241/0x360 lib/dump_stack.c:114
 check_noncircular+0x36a/0x4a0 kernel/locking/lockdep.c:2187
 check_prev_add kernel/locking/lockdep.c:3134 [inline]
 check_prevs_add kernel/locking/lockdep.c:3253 [inline]
 validate_chain+0x18cb/0x58e0 kernel/locking/lockdep.c:3869
 __lock_acquire+0x1346/0x1fd0 kernel/locking/lockdep.c:5137
 lock_acquire+0x1ed/0x550 kernel/locking/lockdep.c:5754
 down_read+0xb1/0xa40 kernel/locking/rwsem.c:1526
 indx_read+0x2a5/0xc50 fs/ntfs3/index.c:1066
 indx_find+0x47a/0xbf0 fs/ntfs3/index.c:1181
 indx_insert_entry+0x527/0x790 fs/ntfs3/index.c:1959
 ni_add_name+0xa24/0xe90 fs/ntfs3/frecord.c:3055
 ni_rename+0xc2/0x1e0 fs/ntfs3/frecord.c:3087
 ntfs_rename+0x7c1/0xd10 fs/ntfs3/namei.c:334
 vfs_rename+0xbdb/0xf00 fs/namei.c:4887
 do_renameat2+0xd94/0x13f0 fs/namei.c:5044
 __do_sys_rename fs/namei.c:5091 [inline]
 __se_sys_rename fs/namei.c:5089 [inline]
 __x64_sys_rename+0x86/0xa0 fs/namei.c:5089
 do_syscall_x64 arch/x86/entry/common.c:52 [inline]
 do_syscall_64+0xf5/0x240 arch/x86/entry/common.c:83
 entry_SYSCALL_64_after_hwframe+0x77/0x7f
RIP: 0033:0x7f0f6987cee9
Code: 28 00 00 00 75 05 48 83 c4 28 c3 e8 e1 20 00 00 90 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 c7 c1 b0 ff ff ff f7 d8 64 89 01 48
RSP: 002b:00007f0f6a65e0c8 EFLAGS: 00000246 ORIG_RAX: 0000000000000052
RAX: ffffffffffffffda RBX: 00007f0f699abf80 RCX: 00007f0f6987cee9
RDX: 0000000000000000 RSI: 0000000020000f00 RDI: 0000000020000400
RBP: 00007f0f698c949e R08: 0000000000000000 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000000
R13: 000000000000000b R14: 00007f0f699abf80 R15: 00007ffdd998d638
 </TASK>

Crashes (4):
Time Kernel Commit Syzkaller Config Log Report Syz repro C repro VM info Assets (help?) Manager Title
2024/05/17 00:34 upstream ea5f6ad9ad96 c2e07261 .config console log report info [disk image] [vmlinux] [kernel image] ci-upstream-kasan-gce-smack-root possible deadlock in indx_read
2024/05/15 17:08 upstream 1b294a1f3561 0b3dad46 .config console log report info [disk image] [vmlinux] [kernel image] ci-upstream-kasan-gce-smack-root possible deadlock in indx_read
2024/05/02 00:31 upstream 0106679839f7 3ba885bc .config console log report info [disk image] [vmlinux] [kernel image] ci-upstream-kasan-gce-smack-root possible deadlock in indx_read
2024/05/01 22:36 upstream 18daea77cca6 3ba885bc .config console log report info [disk image] [vmlinux] [kernel image] ci-upstream-kasan-gce-smack-root possible deadlock in indx_read
* Struck through repros no longer work on HEAD.