==================================================================
BUG: KASAN: slab-use-after-free in lbmIODone+0xed4/0x11f0 fs/jfs/jfs_logmgr.c:2184
Read of size 4 at addr ffff888065e04118 by task ksoftirqd/3/32

CPU: 3 PID: 32 Comm: ksoftirqd/3 Not tainted 6.8.0-rc3-syzkaller-00215-ge6f39a90de92 #0
Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014
Call Trace:
 <TASK>
 __dump_stack lib/dump_stack.c:88 [inline]
 dump_stack_lvl+0xd9/0x1b0 lib/dump_stack.c:106
 print_address_description mm/kasan/report.c:377 [inline]
 print_report+0xc4/0x620 mm/kasan/report.c:488
 kasan_report+0xda/0x110 mm/kasan/report.c:601
 lbmIODone+0xed4/0x11f0 fs/jfs/jfs_logmgr.c:2184
 bio_endio+0x59c/0x6b0 block/bio.c:1608
 req_bio_endio block/blk-mq.c:792 [inline]
 blk_update_request+0x635/0x1710 block/blk-mq.c:937
 blk_mq_end_request+0x4f/0x90 block/blk-mq.c:1060
 lo_complete_rq+0x232/0x2f0 drivers/block/loop.c:366
 blk_complete_reqs+0xae/0xf0 block/blk-mq.c:1135
 __do_softirq+0x21c/0x8e7 kernel/softirq.c:553
 run_ksoftirqd kernel/softirq.c:921 [inline]
 run_ksoftirqd+0x35/0x60 kernel/softirq.c:913
 smpboot_thread_fn+0x669/0xa20 kernel/smpboot.c:164
 kthread+0x2c6/0x3b0 kernel/kthread.c:388
 ret_from_fork+0x45/0x80 arch/x86/kernel/process.c:147
 ret_from_fork_asm+0x1b/0x30 arch/x86/entry/entry_64.S:242
 </TASK>

Allocated by task 9132:
 kasan_save_stack+0x33/0x60 mm/kasan/common.c:47
 kasan_save_track+0x14/0x30 mm/kasan/common.c:68
 poison_kmalloc_redzone mm/kasan/common.c:372 [inline]
 __kasan_kmalloc+0xaa/0xb0 mm/kasan/common.c:389
 kmalloc include/linux/slab.h:590 [inline]
 lbmLogInit fs/jfs/jfs_logmgr.c:1822 [inline]
 lmLogInit+0x3ec/0x1a90 fs/jfs/jfs_logmgr.c:1270
 open_inline_log fs/jfs/jfs_logmgr.c:1175 [inline]
 lmLogOpen+0x7ea/0x1410 fs/jfs/jfs_logmgr.c:1069
 jfs_mount_rw+0x2ea/0x700 fs/jfs/jfs_mount.c:257
 jfs_fill_super+0x9d6/0xd20 fs/jfs/super.c:565
 mount_bdev+0x1e3/0x2d0 fs/super.c:1663
 legacy_get_tree+0x109/0x220 fs/fs_context.c:662
 vfs_get_tree+0x8f/0x380 fs/super.c:1784
 do_new_mount fs/namespace.c:3352 [inline]
 path_mount+0x14ea/0x1f20 fs/namespace.c:3679
 do_mount fs/namespace.c:3692 [inline]
 __do_sys_mount fs/namespace.c:3898 [inline]
 __se_sys_mount fs/namespace.c:3875 [inline]
 __x64_sys_mount+0x297/0x320 fs/namespace.c:3875
 do_syscall_x64 arch/x86/entry/common.c:52 [inline]
 do_syscall_64+0xd5/0x270 arch/x86/entry/common.c:83
 entry_SYSCALL_64_after_hwframe+0x6f/0x77

Freed by task 879:
 kasan_save_stack+0x33/0x60 mm/kasan/common.c:47
 kasan_save_track+0x14/0x30 mm/kasan/common.c:68
 kasan_save_free_info+0x3f/0x60 mm/kasan/generic.c:640
 poison_slab_object mm/kasan/common.c:241 [inline]
 __kasan_slab_free+0x121/0x1c0 mm/kasan/common.c:257
 kasan_slab_free include/linux/kasan.h:184 [inline]
 slab_free_hook mm/slub.c:2121 [inline]
 slab_free mm/slub.c:4299 [inline]
 kfree+0x124/0x370 mm/slub.c:4409
 lbmLogShutdown fs/jfs/jfs_logmgr.c:1865 [inline]
 lmLogShutdown+0x379/0x780 fs/jfs/jfs_logmgr.c:1684
 lmLogClose+0x57b/0x710 fs/jfs/jfs_logmgr.c:1460
 jfs_umount+0x2f0/0x440 fs/jfs/jfs_umount.c:114
 jfs_put_super+0x88/0x1d0 fs/jfs/super.c:194
 generic_shutdown_super+0x159/0x3d0 fs/super.c:646
 kill_block_super+0x3b/0x90 fs/super.c:1680
 deactivate_locked_super+0xbe/0x1a0 fs/super.c:477
 deactivate_super+0xde/0x100 fs/super.c:510
 cleanup_mnt+0x222/0x450 fs/namespace.c:1267
 task_work_run+0x14f/0x250 kernel/task_work.c:180
 resume_user_mode_work include/linux/resume_user_mode.h:50 [inline]
 exit_to_user_mode_loop kernel/entry/common.c:108 [inline]
 exit_to_user_mode_prepare include/linux/entry-common.h:328 [inline]
 __syscall_exit_to_user_mode_work kernel/entry/common.c:201 [inline]
 syscall_exit_to_user_mode+0x281/0x2b0 kernel/entry/common.c:212
 do_syscall_64+0xe5/0x270 arch/x86/entry/common.c:89
 entry_SYSCALL_64_after_hwframe+0x6f/0x77

The buggy address belongs to the object at ffff888065e04110
 which belongs to the cache kmalloc-192 of size 192
The buggy address is located 8 bytes inside of
 freed 192-byte region [ffff888065e04110, ffff888065e041d0)

The buggy address belongs to the physical page:
page:ffffea0001978100 refcount:1 mapcount:0 mapping:0000000000000000 index:0x0 pfn:0x65e04
head:ffffea0001978100 order:1 entire_mapcount:0 nr_pages_mapped:0 pincount:0
anon flags: 0xfff00000000840(slab|head|node=0|zone=1|lastcpupid=0x7ff)
page_type: 0xffffffff()
raw: 00fff00000000840 ffff888014c42a00 0000000000000000 dead000000000001
raw: 0000000000000000 00000000801e001e 00000001ffffffff 0000000000000000
page dumped because: kasan: bad access detected
page_owner tracks the page as allocated
page last allocated via order 1, migratetype Unmovable, gfp_mask 0x1d20c0(__GFP_IO|__GFP_FS|__GFP_NOWARN|__GFP_NORETRY|__GFP_COMP|__GFP_NOMEMALLOC|__GFP_HARDWALL), pid 17668, tgid 17666 (syz-executor.1), ts 2517755755591, free_ts 2509538183211
 set_page_owner include/linux/page_owner.h:31 [inline]
 post_alloc_hook+0x2d4/0x350 mm/page_alloc.c:1533
 prep_new_page mm/page_alloc.c:1540 [inline]
 get_page_from_freelist+0xa28/0x3780 mm/page_alloc.c:3311
 __alloc_pages+0x22f/0x2440 mm/page_alloc.c:4567
 __alloc_pages_node include/linux/gfp.h:238 [inline]
 alloc_pages_node include/linux/gfp.h:261 [inline]
 alloc_slab_page mm/slub.c:2190 [inline]
 allocate_slab mm/slub.c:2354 [inline]
 new_slab+0xcc/0x3a0 mm/slub.c:2407
 ___slab_alloc+0x4af/0x19a0 mm/slub.c:3540
 __slab_alloc.constprop.0+0x56/0xb0 mm/slub.c:3625
 __slab_alloc_node mm/slub.c:3678 [inline]
 slab_alloc_node mm/slub.c:3850 [inline]
 kmalloc_trace+0x30b/0x340 mm/slub.c:4007
 kmalloc include/linux/slab.h:590 [inline]
 __iomap_dio_rw+0x2a7/0x1bd0 fs/iomap/direct-io.c:563
 iomap_dio_rw+0x40/0xa0 fs/iomap/direct-io.c:748
 ext4_dio_read_iter fs/ext4/file.c:94 [inline]
 ext4_file_read_iter+0x4e1/0x6c0 fs/ext4/file.c:145
 call_read_iter include/linux/fs.h:2079 [inline]
 copy_splice_read+0x617/0xb80 fs/splice.c:365
 do_splice_read fs/splice.c:984 [inline]
 do_splice_read+0x2cf/0x380 fs/splice.c:959
 splice_direct_to_actor+0x2a5/0xa40 fs/splice.c:1089
 do_splice_direct_actor fs/splice.c:1207 [inline]
 do_splice_direct+0x17e/0x250 fs/splice.c:1233
 do_sendfile+0xaac/0xdb0 fs/read_write.c:1295
 __do_sys_sendfile64 fs/read_write.c:1362 [inline]
 __se_sys_sendfile64 fs/read_write.c:1348 [inline]
 __x64_sys_sendfile64+0x1da/0x220 fs/read_write.c:1348
page last free pid 17172 tgid 17170 stack trace:
 reset_page_owner include/linux/page_owner.h:24 [inline]
 free_pages_prepare mm/page_alloc.c:1140 [inline]
 free_unref_page_prepare+0x527/0xb10 mm/page_alloc.c:2346
 free_unref_page+0x33/0x3c0 mm/page_alloc.c:2486
 kasan_depopulate_vmalloc_pte+0x63/0x80 mm/kasan/shadow.c:415
 apply_to_pte_range mm/memory.c:2619 [inline]
 apply_to_pmd_range mm/memory.c:2663 [inline]
 apply_to_pud_range mm/memory.c:2699 [inline]
 apply_to_p4d_range mm/memory.c:2735 [inline]
 __apply_to_page_range+0x57e/0xdb0 mm/memory.c:2769
 kasan_release_vmalloc+0xac/0xc0 mm/kasan/shadow.c:532
 __purge_vmap_area_lazy+0x8b9/0x2170 mm/vmalloc.c:1770
 _vm_unmap_aliases+0x693/0x9b0 mm/vmalloc.c:2306
 change_page_attr_set_clr+0x24c/0x490 arch/x86/mm/pat/set_memory.c:1838
 change_page_attr_clear arch/x86/mm/pat/set_memory.c:1886 [inline]
 set_memory_ro+0x7c/0xb0 arch/x86/mm/pat/set_memory.c:2077
 bpf_prog_lock_ro include/linux/filter.h:889 [inline]
 bpf_prog_select_runtime+0x4fb/0x640 kernel/bpf/core.c:2395
 bpf_prog_load+0x1514/0x2300 kernel/bpf/syscall.c:2773
 __sys_bpf+0xbfb/0x4a30 kernel/bpf/syscall.c:5457
 __do_sys_bpf kernel/bpf/syscall.c:5561 [inline]
 __se_sys_bpf kernel/bpf/syscall.c:5559 [inline]
 __x64_sys_bpf+0x78/0xc0 kernel/bpf/syscall.c:5559
 do_syscall_x64 arch/x86/entry/common.c:52 [inline]
 do_syscall_64+0xd5/0x270 arch/x86/entry/common.c:83
 entry_SYSCALL_64_after_hwframe+0x6f/0x77

Memory state around the buggy address:
 ffff888065e04000: 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
 ffff888065e04080: 00 00 00 00 fc fc fc fc fc fc fc fc fc fc fc fc
>ffff888065e04100: fc fc fa fb fb fb fb fb fb fb fb fb fb fb fb fb
                            ^
 ffff888065e04180: fb fb fb fb fb fb fb fb fb fb fc fc fc fc fc fc
 ffff888065e04200: fc fc fc fc 00 00 00 00 00 00 00 00 00 00 00 00
==================================================================