F2FS-fs (loop2): Magic Mismatch, valid(0xf2f52010) - read(0x0) attempt to access beyond end of device loop2: rw=12288, want=8200, limit=20 attempt to access beyond end of device loop2: rw=12288, want=12296, limit=20 INFO: task syz-executor6:24623 blocked for more than 140 seconds. Not tainted 4.18.0-next-20180817+ #42 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. syz-executor6 D22152 24623 4518 0x00000004 Call Trace: context_switch kernel/sched/core.c:2825 [inline] __schedule+0x87c/0x1df0 kernel/sched/core.c:3473 schedule+0xfb/0x450 kernel/sched/core.c:3517 schedule_timeout+0x1cc/0x260 kernel/time/timer.c:1780 do_wait_for_common kernel/sched/completion.c:83 [inline] __wait_for_common kernel/sched/completion.c:104 [inline] wait_for_common kernel/sched/completion.c:115 [inline] wait_for_completion+0x430/0x8d0 kernel/sched/completion.c:136 flush_work+0x531/0x900 kernel/workqueue.c:2918 __cancel_work_timer+0x4bd/0x830 kernel/workqueue.c:2989 cancel_work_sync+0x17/0x20 kernel/workqueue.c:3025 p9_conn_destroy net/9p/trans_fd.c:865 [inline] p9_fd_close+0x376/0x5c0 net/9p/trans_fd.c:891 p9_client_create+0x9bc/0x150c net/9p/client.c:1065 v9fs_session_init+0x21a/0x1a80 fs/9p/v9fs.c:400 v9fs_mount+0x7c/0x900 fs/9p/vfs_super.c:135 legacy_get_tree+0x131/0x460 fs/fs_context.c:732 vfs_get_tree+0x1cb/0x5c0 fs/super.c:1746 do_new_mount fs/namespace.c:2627 [inline] do_mount+0x6f9/0x1e30 fs/namespace.c:2951 ksys_mount+0x12d/0x140 fs/namespace.c:3167 __do_sys_mount fs/namespace.c:3181 [inline] __se_sys_mount fs/namespace.c:3178 [inline] __x64_sys_mount+0xbe/0x150 fs/namespace.c:3178 do_syscall_64+0x1b9/0x820 arch/x86/entry/common.c:290 entry_SYSCALL_64_after_hwframe+0x49/0xbe RIP: 0033:0x457089 Code: 08 e8 3b a9 fb ff 48 8b 44 24 20 48 8b 4c 24 18 48 8b 54 24 38 48 8b 7c 24 48 eb 8d 48 8b 6c 24 50 48 83 c4 58 c3 48 89 04 24 <48> 89 5c 24 08 e8 0d a9 fb ff 48 8b 44 24 40 48 8b 4c 24 48 48 8b RSP: 002b:00007fb700233c78 EFLAGS: 00000246 ORIG_RAX: 00000000000000a5 RAX: ffffffffffffffda RBX: 00007fb7002346d4 RCX: 0000000000457089 RDX: 0000000020000340 RSI: 00000000200000c0 RDI: 0000000000000000 RBP: 0000000000930140 R08: 00000000200001c0 R09: 0000000000000000 R10: 0000000000000000 R11: 0000000000000246 R12: 00000000ffffffff R13: 00000000004d2768 R14: 00000000004c7be5 R15: 0000000000000001 INFO: task syz-executor6:24658 blocked for more than 140 seconds. Not tainted 4.18.0-next-20180817+ #42 "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. syz-executor6 D23912 24658 4518 0x00000004 Call Trace: context_switch kernel/sched/core.c:2825 [inline] __schedule+0x87c/0x1df0 kernel/sched/core.c:3473 schedule+0xfb/0x450 kernel/sched/core.c:3517 schedule_timeout+0x1cc/0x260 kernel/time/timer.c:1780 do_wait_for_common kernel/sched/completion.c:83 [inline] __wait_for_common kernel/sched/completion.c:104 [inline] wait_for_common kernel/sched/completion.c:115 [inline] wait_for_completion+0x430/0x8d0 kernel/sched/completion.c:136 flush_work+0x531/0x900 kernel/workqueue.c:2918 __cancel_work_timer+0x4bd/0x830 kernel/workqueue.c:2989 cancel_work_sync+0x17/0x20 kernel/workqueue.c:3025 p9_conn_destroy net/9p/trans_fd.c:865 [inline] p9_fd_close+0x376/0x5c0 net/9p/trans_fd.c:891 p9_client_create+0x9bc/0x150c net/9p/client.c:1065 v9fs_session_init+0x21a/0x1a80 fs/9p/v9fs.c:400 v9fs_mount+0x7c/0x900 fs/9p/vfs_super.c:135 legacy_get_tree+0x131/0x460 fs/fs_context.c:732 vfs_get_tree+0x1cb/0x5c0 fs/super.c:1746 do_new_mount fs/namespace.c:2627 [inline] do_mount+0x6f9/0x1e30 fs/namespace.c:2951 ksys_mount+0x12d/0x140 fs/namespace.c:3167 __do_sys_mount fs/namespace.c:3181 [inline] __se_sys_mount fs/namespace.c:3178 [inline] __x64_sys_mount+0xbe/0x150 fs/namespace.c:3178 do_syscall_64+0x1b9/0x820 arch/x86/entry/common.c:290 entry_SYSCALL_64_after_hwframe+0x49/0xbe RIP: 0033:0x457089 Code: 08 e8 3b a9 fb ff 48 8b 44 24 20 48 8b 4c 24 18 48 8b 54 24 38 48 8b 7c 24 48 eb 8d 48 8b 6c 24 50 48 83 c4 58 c3 48 89 04 24 <48> 89 5c 24 08 e8 0d a9 fb ff 48 8b 44 24 40 48 8b 4c 24 48 48 8b RSP: 002b:00007fb7001f1c78 EFLAGS: 00000246 ORIG_RAX: 00000000000000a5 RAX: ffffffffffffffda RBX: 00007fb7001f26d4 RCX: 0000000000457089 RDX: 0000000020000900 RSI: 0000000020000000 RDI: 0000000000000000 RBP: 0000000000930280 R08: 0000000020000840 R09: 0000000000000000 R10: 0000000000000000 R11: 0000000000000246 R12: 00000000ffffffff R13: 00000000004d2768 R14: 00000000004c7be5 R15: 0000000000000003 Showing all locks held in the system: 2 locks held by kworker/0:0/5: #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: __write_once_size include/linux/compiler.h:215 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: arch_atomic64_set arch/x86/include/asm/atomic64_64.h:34 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: atomic64_set include/asm-generic/atomic-instrumented.h:40 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: atomic_long_set include/asm-generic/atomic-long.h:59 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: set_work_data kernel/workqueue.c:617 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: set_work_pool_and_clear_pending kernel/workqueue.c:644 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: process_one_work+0xb44/0x1aa0 kernel/workqueue.c:2124 #1: 000000001efc1200 ((work_completion)(&m->wq)){+.+.}, at: process_one_work+0xb9b/0x1aa0 kernel/workqueue.c:2128 1 lock held by khungtaskd/777: #0: 0000000067675b7f (rcu_read_lock){....}, at: debug_show_all_locks+0xd0/0x428 kernel/locking/lockdep.c:4436 2 locks held by getty/4464: #0: 00000000bbc77a33 (&tty->ldisc_sem){++++}, at: ldsem_down_read+0x37/0x40 drivers/tty/tty_ldsem.c:353 #1: 0000000067db7146 (&ldata->atomic_read_lock){+.+.}, at: n_tty_read+0x335/0x1ce0 drivers/tty/n_tty.c:2140 2 locks held by getty/4465: #0: 0000000068b6131c (&tty->ldisc_sem){++++}, at: ldsem_down_read+0x37/0x40 drivers/tty/tty_ldsem.c:353 #1: 00000000a8d77d3e (&ldata->atomic_read_lock){+.+.}, at: n_tty_read+0x335/0x1ce0 drivers/tty/n_tty.c:2140 2 locks held by getty/4466: #0: 00000000170d810c (&tty->ldisc_sem){++++}, at: ldsem_down_read+0x37/0x40 drivers/tty/tty_ldsem.c:353 #1: 00000000a6d1008a (&ldata->atomic_read_lock){+.+.}, at: n_tty_read+0x335/0x1ce0 drivers/tty/n_tty.c:2140 2 locks held by getty/4467: #0: 0000000001d505c9 (&tty->ldisc_sem){++++}, at: ldsem_down_read+0x37/0x40 drivers/tty/tty_ldsem.c:353 #1: 00000000c8d68c7a (&ldata->atomic_read_lock){+.+.}, at: n_tty_read+0x335/0x1ce0 drivers/tty/n_tty.c:2140 2 locks held by getty/4468: #0: 000000000a34b339 (&tty->ldisc_sem){++++}, at: ldsem_down_read+0x37/0x40 drivers/tty/tty_ldsem.c:353 #1: 000000002cc7cec0 (&ldata->atomic_read_lock){+.+.}, at: n_tty_read+0x335/0x1ce0 drivers/tty/n_tty.c:2140 2 locks held by getty/4469: #0: 000000001eca6514 (&tty->ldisc_sem){++++}, at: ldsem_down_read+0x37/0x40 drivers/tty/tty_ldsem.c:353 #1: 00000000c26c11bd (&ldata->atomic_read_lock){+.+.}, at: n_tty_read+0x335/0x1ce0 drivers/tty/n_tty.c:2140 2 locks held by getty/4470: #0: 000000000091c85d (&tty->ldisc_sem){++++}, at: ldsem_down_read+0x37/0x40 drivers/tty/tty_ldsem.c:353 #1: 0000000083aa9868 (&ldata->atomic_read_lock){+.+.}, at: n_tty_read+0x335/0x1ce0 drivers/tty/n_tty.c:2140 2 locks held by kworker/0:3/5265: 2 locks held by kworker/0:4/9201: #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: __write_once_size include/linux/compiler.h:215 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: arch_atomic64_set arch/x86/include/asm/atomic64_64.h:34 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: atomic64_set include/asm-generic/atomic-instrumented.h:40 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: atomic_long_set include/asm-generic/atomic-long.h:59 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: set_work_data kernel/workqueue.c:617 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: set_work_pool_and_clear_pending kernel/workqueue.c:644 [inline] #0: 0000000064ede54e ((wq_completion)"events"){+.+.}, at: process_one_work+0xb44/0x1aa0 kernel/workqueue.c:2124 #1: 00000000e790af32 ((work_completion)(&m->wq)){+.+.}, at: process_one_work+0xb9b/0x1aa0 kernel/workqueue.c:2128 ============================================= NMI backtrace for cpu 1 CPU: 1 PID: 777 Comm: khungtaskd Not tainted 4.18.0-next-20180817+ #42 Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 01/01/2011 Call Trace: __dump_stack lib/dump_stack.c:77 [inline] dump_stack+0x1c9/0x2b4 lib/dump_stack.c:113 nmi_cpu_backtrace.cold.3+0x48/0x88 lib/nmi_backtrace.c:101 nmi_trigger_cpumask_backtrace+0x151/0x192 lib/nmi_backtrace.c:62 arch_trigger_cpumask_backtrace+0x14/0x20 arch/x86/kernel/apic/hw_nmi.c:38 trigger_all_cpu_backtrace include/linux/nmi.h:144 [inline] check_hung_uninterruptible_tasks kernel/hung_task.c:204 [inline] watchdog+0xb39/0x1040 kernel/hung_task.c:265 kthread+0x35a/0x420 kernel/kthread.c:246 ret_from_fork+0x3a/0x50 arch/x86/entry/entry_64.S:415 Sending NMI from CPU 1 to CPUs 0: NMI backtrace for cpu 0 CPU: 0 PID: 5265 Comm: kworker/0:3 Not tainted 4.18.0-next-20180817+ #42 Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 01/01/2011 Workqueue: (null) (events) RIP: 0010:lockdep_hardirqs_off+0x10c/0x2c0 kernel/locking/lockdep.c:2901 Code: 08 00 00 48 b8 00 00 00 00 00 fc ff df 48 89 fa 48 c1 ea 03 0f b6 04 02 84 c0 74 08 3c 03 0f 8e 61 01 00 00 8b 83 08 08 00 00 <85> c0 0f 84 a4 00 00 00 48 8d bb f8 07 00 00 48 b8 00 00 00 00 00 RSP: 0018:ffff88018cb37578 EFLAGS: 00000046 RAX: 0000000000000000 RBX: ffff8801d84ec180 RCX: 0000000000000000 RDX: 1ffff1003b09d931 RSI: 0000000000000000 RDI: ffff8801d84ec988 RBP: ffff88018cb37590 R08: ffff8801d84ec180 R09: fffffbfff111aecc R10: fffffbfff111aecc R11: ffffffff888d7667 R12: ffff8801d84ec180 R13: ffffffff86b6ba54 R14: ffffffff87f64880 R15: dffffc0000000000 FS: 0000000000000000(0000) GS:ffff8801db000000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 CR2: 0000000000000130 CR3: 00000001b1236000 CR4: 00000000001406f0 Call Trace: trace_hardirqs_off+0xb8/0x2b0 kernel/trace/trace_preemptirq.c:43 __raw_spin_lock_irqsave include/linux/spinlock_api_smp.h:108 [inline] _raw_spin_lock_irqsave+0x74/0xc0 kernel/locking/spinlock.c:152 debug_object_deactivate+0xf7/0x450 lib/debugobjects.c:541 debug_work_deactivate kernel/workqueue.c:496 [inline] process_one_work+0x594/0x1aa0 kernel/workqueue.c:2084 worker_thread+0x189/0x13c0 kernel/workqueue.c:2296 kthread+0x35a/0x420 kernel/kthread.c:246 ret_from_fork+0x3a/0x50 arch/x86/entry/entry_64.S:415