================================================================== BUG: KASAN: slab-use-after-free in __lock_acquire+0xff/0x7c80 kernel/locking/lockdep.c:5005 Read of size 8 at addr ffff8880240a38d8 by task jfsCommit/112 CPU: 0 PID: 112 Comm: jfsCommit Not tainted syzkaller #0 Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 10/25/2025 Call Trace: dump_stack_lvl+0x16c/0x230 lib/dump_stack.c:106 print_address_description mm/kasan/report.c:364 [inline] print_report+0xac/0x220 mm/kasan/report.c:468 kasan_report+0x117/0x150 mm/kasan/report.c:581 __lock_acquire+0xff/0x7c80 kernel/locking/lockdep.c:5005 lock_acquire+0x197/0x410 kernel/locking/lockdep.c:5754 __raw_spin_lock include/linux/spinlock_api_smp.h:133 [inline] _raw_spin_lock+0x2e/0x40 kernel/locking/spinlock.c:154 __mutex_lock_common kernel/locking/mutex.c:617 [inline] __mutex_lock+0x3b4/0xcc0 kernel/locking/mutex.c:747 jfs_syncpt+0x25/0x90 fs/jfs/jfs_logmgr.c:1039 txEnd+0x2e5/0x520 fs/jfs/jfs_txnmgr.c:550 txLazyCommit fs/jfs/jfs_txnmgr.c:2685 [inline] jfs_lazycommit+0x5a6/0xa60 fs/jfs/jfs_txnmgr.c:2733 kthread+0x2fa/0x390 kernel/kthread.c:388 ret_from_fork+0x48/0x80 arch/x86/kernel/process.c:152 ret_from_fork_asm+0x11/0x20 arch/x86/entry/entry_64.S:293 Allocated by task 14785: kasan_save_stack mm/kasan/common.c:46 [inline] kasan_set_track+0x4e/0x70 mm/kasan/common.c:53 ____kasan_kmalloc mm/kasan/common.c:375 [inline] __kasan_kmalloc+0x8f/0xa0 mm/kasan/common.c:384 kmalloc include/linux/slab.h:600 [inline] kzalloc include/linux/slab.h:721 [inline] open_inline_log fs/jfs/jfs_logmgr.c:1159 [inline] lmLogOpen+0x2df/0xfb0 fs/jfs/jfs_logmgr.c:1069 jfs_mount_rw+0xea/0x670 fs/jfs/jfs_mount.c:257 jfs_fill_super+0x592/0xac0 fs/jfs/super.c:565 mount_bdev+0x22b/0x2d0 fs/super.c:1643 legacy_get_tree+0xea/0x180 fs/fs_context.c:662 vfs_get_tree+0x8c/0x280 fs/super.c:1764 do_new_mount+0x24b/0xa40 fs/namespace.c:3386 do_mount fs/namespace.c:3726 [inline] __do_sys_mount fs/namespace.c:3935 [inline] __se_sys_mount+0x2da/0x3c0 fs/namespace.c:3912 do_syscall_x64 arch/x86/entry/common.c:46 [inline] do_syscall_64+0x55/0xb0 arch/x86/entry/common.c:76 entry_SYSCALL_64_after_hwframe+0x68/0xd2 Freed by task 5769: kasan_save_stack mm/kasan/common.c:46 [inline] kasan_set_track+0x4e/0x70 mm/kasan/common.c:53 kasan_save_free_info+0x2e/0x50 mm/kasan/generic.c:522 ____kasan_slab_free+0x126/0x1e0 mm/kasan/common.c:237 kasan_slab_free include/linux/kasan.h:164 [inline] slab_free_hook mm/slub.c:1811 [inline] slab_free_freelist_hook+0x130/0x1b0 mm/slub.c:1837 slab_free mm/slub.c:3830 [inline] __kmem_cache_free+0xba/0x1f0 mm/slub.c:3843 lmLogClose+0x297/0x520 fs/jfs/jfs_logmgr.c:-1 jfs_umount+0x2ef/0x3c0 fs/jfs/jfs_umount.c:114 jfs_put_super+0x8c/0x190 fs/jfs/super.c:194 generic_shutdown_super+0x134/0x2b0 fs/super.c:693 kill_block_super+0x44/0x90 fs/super.c:1660 deactivate_locked_super+0x97/0x100 fs/super.c:481 cleanup_mnt+0x429/0x4c0 fs/namespace.c:1259 task_work_run+0x1ce/0x250 kernel/task_work.c:245 resume_user_mode_work include/linux/resume_user_mode.h:49 [inline] exit_to_user_mode_loop+0xe6/0x110 kernel/entry/common.c:177 exit_to_user_mode_prepare+0xf6/0x180 kernel/entry/common.c:210 __syscall_exit_to_user_mode_work kernel/entry/common.c:291 [inline] syscall_exit_to_user_mode+0x1a/0x50 kernel/entry/common.c:302 do_syscall_64+0x61/0xb0 arch/x86/entry/common.c:82 entry_SYSCALL_64_after_hwframe+0x68/0xd2 Last potentially related work creation: kasan_save_stack+0x3e/0x60 mm/kasan/common.c:46 __kasan_record_aux_stack+0xaf/0xc0 mm/kasan/generic.c:492 kvfree_call_rcu+0xee/0x780 kernel/rcu/tree.c:3452 neigh_periodic_work+0x3f7/0xd70 net/core/neighbour.c:1010 process_one_work kernel/workqueue.c:2634 [inline] process_scheduled_works+0xa45/0x15b0 kernel/workqueue.c:2711 worker_thread+0xa55/0xfc0 kernel/workqueue.c:2792 kthread+0x2fa/0x390 kernel/kthread.c:388 ret_from_fork+0x48/0x80 arch/x86/kernel/process.c:152 ret_from_fork_asm+0x11/0x20 arch/x86/entry/entry_64.S:293 The buggy address belongs to the object at ffff8880240a3800 which belongs to the cache kmalloc-1k of size 1024 The buggy address is located 216 bytes inside of freed 1024-byte region [ffff8880240a3800, ffff8880240a3c00) The buggy address belongs to the physical page: page:ffffea0000902800 refcount:1 mapcount:0 mapping:0000000000000000 index:0x0 pfn:0x240a0 head:ffffea0000902800 order:3 entire_mapcount:0 nr_pages_mapped:0 pincount:0 flags: 0xfff00000000840(slab|head|node=0|zone=1|lastcpupid=0x7ff) page_type: 0xffffffff() raw: 00fff00000000840 ffff888017841dc0 ffffea0001db0800 dead000000000002 raw: 0000000000000000 0000000000100010 00000001ffffffff 0000000000000000 page dumped because: kasan: bad access detected page_owner tracks the page as allocated page last allocated via order 3, migratetype Unmovable, gfp_mask 0x1d20c0(__GFP_IO|__GFP_FS|__GFP_NOWARN|__GFP_NORETRY|__GFP_COMP|__GFP_NOMEMALLOC|__GFP_HARDWALL), pid 6242, tgid 6241 (syz.2.177), ts 112788087199, free_ts 112731543054 set_page_owner include/linux/page_owner.h:31 [inline] post_alloc_hook+0x1cd/0x210 mm/page_alloc.c:1554 prep_new_page mm/page_alloc.c:1561 [inline] get_page_from_freelist+0x195c/0x19f0 mm/page_alloc.c:3191 __alloc_pages+0x1e3/0x460 mm/page_alloc.c:4457 alloc_slab_page+0x5d/0x170 mm/slub.c:1881 allocate_slab mm/slub.c:2028 [inline] new_slab+0x87/0x2e0 mm/slub.c:2081 ___slab_alloc+0xc6d/0x1300 mm/slub.c:3253 __slab_alloc mm/slub.c:3339 [inline] __slab_alloc_node mm/slub.c:3392 [inline] slab_alloc_node mm/slub.c:3485 [inline] __kmem_cache_alloc_node+0x1a2/0x260 mm/slub.c:3534 __do_kmalloc_node mm/slab_common.c:1006 [inline] __kmalloc_node+0xa4/0x230 mm/slab_common.c:1014 kmalloc_node include/linux/slab.h:620 [inline] kzalloc_node include/linux/slab.h:732 [inline] qdisc_alloc+0x94/0xa50 net/sched/sch_generic.c:951 qdisc_create_dflt+0x63/0x430 net/sched/sch_generic.c:1012 attach_one_default_qdisc net/sched/sch_generic.c:1176 [inline] netdev_for_each_tx_queue include/linux/netdevice.h:2520 [inline] attach_default_qdiscs net/sched/sch_generic.c:1194 [inline] dev_activate+0x397/0x11a0 net/sched/sch_generic.c:1253 __dev_open+0x338/0x430 net/core/dev.c:1541 __dev_change_flags+0x20e/0x6a0 net/core/dev.c:8716 rtnl_configure_link net/core/rtnetlink.c:3328 [inline] rtnl_newlink_create net/core/rtnetlink.c:3525 [inline] __rtnl_newlink net/core/rtnetlink.c:3737 [inline] rtnl_newlink+0x18a5/0x2020 net/core/rtnetlink.c:3750 rtnetlink_rcv_msg+0x7c7/0xf10 net/core/rtnetlink.c:6472 netlink_rcv_skb+0x216/0x480 net/netlink/af_netlink.c:2545 page last free stack trace: reset_page_owner include/linux/page_owner.h:24 [inline] free_pages_prepare mm/page_alloc.c:1154 [inline] free_unref_page_prepare+0x7ce/0x8e0 mm/page_alloc.c:2336 free_unref_page+0x32/0x2e0 mm/page_alloc.c:2429 discard_slab mm/slub.c:2127 [inline] __unfreeze_partials+0x1cf/0x210 mm/slub.c:2667 put_cpu_partial+0x17c/0x250 mm/slub.c:2743 __slab_free+0x31d/0x410 mm/slub.c:3700 qlink_free mm/kasan/quarantine.c:166 [inline] qlist_free_all+0x75/0xe0 mm/kasan/quarantine.c:185 kasan_quarantine_reduce+0x143/0x160 mm/kasan/quarantine.c:292 __kasan_slab_alloc+0x22/0x80 mm/kasan/common.c:306 kasan_slab_alloc include/linux/kasan.h:188 [inline] slab_post_alloc_hook+0x6e/0x4d0 mm/slab.h:767 slab_alloc_node mm/slub.c:3495 [inline] slab_alloc mm/slub.c:3503 [inline] __kmem_cache_alloc_lru mm/slub.c:3510 [inline] kmem_cache_alloc+0x11e/0x2e0 mm/slub.c:3519 kmem_cache_zalloc include/linux/slab.h:711 [inline] __kernfs_new_node+0xd8/0x7e0 fs/kernfs/dir.c:624 kernfs_new_node+0x14c/0x260 fs/kernfs/dir.c:700 kernfs_create_dir_ns+0x44/0x120 fs/kernfs/dir.c:1061 sysfs_create_dir_ns+0x124/0x280 fs/sysfs/dir.c:59 create_dir lib/kobject.c:73 [inline] kobject_add_internal+0x6b8/0xc70 lib/kobject.c:240 kobject_add_varg lib/kobject.c:374 [inline] kobject_init_and_add+0x126/0x190 lib/kobject.c:457 Memory state around the buggy address: ffff8880240a3780: fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc ffff8880240a3800: fa fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb >ffff8880240a3880: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb ^ ffff8880240a3900: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb ffff8880240a3980: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb ==================================================================