5ba95adf5b5f8299585e97372eb668ab24f3d8a5803b12dd9a3391ec499234226020345a57a9ecaf512dedfb5162388a8a1794d4b22951bc35bf810a78294131282fefc53c8efc4176060d53b1048a3dbb8b570b9b9d49751c4a8ceef0633db7e5bb8c848b8f8834e45e028b6069e7b9924170f8e5c15fe8508fa33ab9568b775b070b03bfd4646133a5a18427be914ef9cfaebe5d3cd55d1e552b3f30bfc4fe2717fcb9ca03d4861dd648d5354e24bd70ed5662b3ec1c208a17afb003dacf5467bdc732cac731cdb21b06f63313c756dfceaa218543dd1ab885266a0d98db8082a01881ac9c05c34a8619710c6173cf9bb7c6ae6412bdde536d2a49a35382258a62834ecb4cf996efa320f66281a121a2db3259cc386796c6b10bf21a86263c5b8ccf74fdc5269f79cf8e8440eea1562a87d832988acc221d59761ea210962a7a387894cc7703e4b382e92ebd0ada29e32b7c6e10103b4497d73adfd6d816d7686a339c5ee82d15bf1045f61a10928cbc9b6e9c7a9150ac1833bee5a7c29e1ece6be8936de32e640fd35115bc4004765d988f366d1666236badc7c54749bd32e638114e52ed99b3c555c4d600026918a3c39e1525a24044940f80a6841217c772cfc56eb5cff1bb8ec4f87d4a5aedbc3594319ff3298f7254066b7807ffbd144735d98172b70cb7ba9660ba86fbf3ad709d8dfb17b2002baac25791000a234d1f9e79015103fa055d613a2a31f2e0e54fd6152c8a8c09cc31db849f7e45f7ae0b17d7fac9a25ecdd00c7cc53416873fdf0dafa71271dd12ae147f1e986b8c7c5d070024efe66c8eea26f37794aeebff40413daa07c9110e6fa8a243e7c43b36d23e0bd8ce93a993a32f0e69953f843aca4010b3d871c6c84850fc8f884f33f3adf0f5c6466641a1d9062f0f48df4d2836633f9bbcbea0632490dec93ffe712da54c5ecbc56a57dd29790079cad18679be3d43adec9b63b6fe5244b5393eba4ed3304729a4add724a3969be0106bcba908e3c7aa3015b7414a31feb1d4d6d148c44197f073578ac318f68b9237984eb7e3f6f336bd48152b0ce5d691794bf91bc56d5758feec381e01f2988cdf24c7943a76012dee6e24841d0d123d05014748a31609dff3bc5abce737d2b7b2459cd67db5395d3da7bc1653f7768a5d0f960ae0f88e5c4318342798c8868c800b3b273423b66437acc4bffc157096ca16d47562c48f08f6006000b2909f05c589e6dd110598d71135297044ae2228f0bb7bdfc699b50e79fe104098d7d43d7e29a55d7a85238d5112885efa4f25d120de0ac861aef70c12c0a354e60ebf9cd41928545822bb6fdc00aeefb46458d751e0a9efe4f1daa2951ab022c644485fc6ccca76e942cdc1bbf459ef58617c42a2ae899a15f176358e5e79c7159b47bbf1d84d2fe153780507da7af671dfbcd2dcf5cb35f0f444452457fcda81d0bcd693e2639dde64827cfb3e30cd2d82981c9ff7eb3a63ffe5b6fce029b349158b5c98f2be7ddcd4b5fc09832eafc253b0d9b77059e30b5636ac14495e76adb3b1fd8551bc07a76df3a50e4344dee9f0598696a3595ac5ae25076831b22221363fa289abfc234c77f1199500cc91dc362f1d4abc37f1f5e8ff617a6f03fa7cc173563f77fb5bec7641c77f5610bf5517f56ca90aa603657baa1374bda17385eec7b9800cde4b099e59ffd624ce12fd", 0x1000}, {&(0x7f00000025c0)="610f50fb9263fa89028694521fc3cded16d51a7330f4a5b2145ce2f972c3d3d1aef9c07867df081327fc9632f58871c422f0d7bffe548c32f32ec29e57f36586b96306f27b401e3d6f5eb7a46f725987a4636e8a7234a130e0870dfa2dc586f5144e3137cd268d2629d263d463c862a8886aa1f41e2aaa15cbb42173e547c75f336a3a82be74c42954944747a54ee5e87206283cdafe8cf3f529746b5c83442b487e6c8274ffc6", 0xa7}, {&(0x7f0000002680)="6246d1827c39ab5e3793d1c7db13768db6cf30337444a3d0f39f3a5657e5d6efbdb4c9afeee263e692b5a62f453c09378656f1f34bd165d681db547ad284e0b37da987dba37678", 0x47}], 0x4, &(0x7f0000002740)=[@rights={{0x1c, 0x1, 0x1, [r1, r0, r1]}}, @rights={{0x1c, 0x1, 0x1, [r1, r21, 0xffffffffffffffff]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r22}}}], 0x60, 0x800}}], 0x5, 0x8000) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:42 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x7f020000) 00:28:42 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0xe00}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:42 executing program 3: sigaltstack(&(0x7f0000000240)={&(0x7f0000000180)=""/170, 0x80000002, 0xaa}, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) r0 = add_key$fscrypt_v1(&(0x7f0000000000), &(0x7f0000000040)={'fscrypt:', @desc2}, &(0x7f0000000080)={0x0, "90f352e54acbb500953aa3f312922f63fdcdef04a9bc505ca06452b18885376a912d89dc56bb1c077fd9f501b2b166fb65c818d24cad10b8243d95a16f2b92bd", 0x29}, 0x48, 0xfffffffffffffffc) r1 = add_key$keyring(&(0x7f0000000100), &(0x7f0000000140)={'syz', 0x0}, 0x0, 0x0, 0xfffffffffffffffc) keyctl$link(0x8, r0, r1) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) r2 = add_key$fscrypt_v1(&(0x7f0000000480), &(0x7f00000004c0)={'fscrypt:', @desc3}, &(0x7f0000000500)={0x0, "85d71362e3d9ad3c71804c36061af5d79c0b58f087be4005a6d7406c375ea05379542602015ab751c651646eeefd5b39717a8f08ce678b7b61fb025b140cf746", 0x2a}, 0x48, 0xfffffffffffffff8) keyctl$clear(0x7, r2) r3 = getuid() keyctl$get_persistent(0x16, r3, r0) 00:28:42 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0xf00}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:42 executing program 1: sigaltstack(&(0x7f0000000240)={&(0x7f0000000180)=""/170, 0x80000002, 0xaa}, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) r0 = add_key$fscrypt_v1(&(0x7f0000000000), &(0x7f0000000040)={'fscrypt:', @desc2}, &(0x7f0000000080)={0x0, "90f352e54acbb500953aa3f312922f63fdcdef04a9bc505ca06452b18885376a912d89dc56bb1c077fd9f501b2b166fb65c818d24cad10b8243d95a16f2b92bd", 0x29}, 0x48, 0xfffffffffffffffc) r1 = add_key$keyring(&(0x7f0000000100), &(0x7f0000000140)={'syz', 0x0}, 0x0, 0x0, 0xfffffffffffffffc) keyctl$link(0x8, r0, r1) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) r2 = add_key$fscrypt_v1(&(0x7f0000000480), &(0x7f00000004c0)={'fscrypt:', @desc3}, &(0x7f0000000500)={0x0, "85d71362e3d9ad3c71804c36061af5d79c0b58f087be4005a6d7406c375ea05379542602015ab751c651646eeefd5b39717a8f08ce678b7b61fb025b140cf746", 0x2a}, 0x48, 0xfffffffffffffff8) keyctl$clear(0x7, r2) r3 = getuid() keyctl$get_persistent(0x16, r3, r0) 00:28:42 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4040) getpid() getpid() getuid() getuid() 00:28:42 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4800) getpid() getpid() getuid() getuid() 00:28:43 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x88030000) 00:28:43 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1100}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:43 executing program 0: sigaltstack(&(0x7f0000000240)={&(0x7f0000000180)=""/170, 0x80000002, 0xaa}, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) r0 = add_key$fscrypt_v1(&(0x7f0000000000), &(0x7f0000000040)={'fscrypt:', @desc2}, &(0x7f0000000080)={0x0, "90f352e54acbb500953aa3f312922f63fdcdef04a9bc505ca06452b18885376a912d89dc56bb1c077fd9f501b2b166fb65c818d24cad10b8243d95a16f2b92bd", 0x29}, 0x48, 0xfffffffffffffffc) r1 = add_key$keyring(&(0x7f0000000100), &(0x7f0000000140)={'syz', 0x0}, 0x0, 0x0, 0xfffffffffffffffc) keyctl$link(0x8, r0, r1) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) r2 = add_key$fscrypt_v1(&(0x7f0000000480), &(0x7f00000004c0)={'fscrypt:', @desc3}, &(0x7f0000000500)={0x0, "85d71362e3d9ad3c71804c36061af5d79c0b58f087be4005a6d7406c375ea05379542602015ab751c651646eeefd5b39717a8f08ce678b7b61fb025b140cf746", 0x2a}, 0x48, 0xfffffffffffffff8) keyctl$clear(0x7, r2) r3 = getuid() keyctl$get_persistent(0x16, r3, r0) 00:28:43 executing program 3: socketpair(0x29, 0x6, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socket$unix(0x1, 0x1, 0x0) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:43 executing program 2: sigaltstack(&(0x7f0000000240)={&(0x7f0000000180)=""/170, 0x80000002, 0xaa}, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) (async) r0 = add_key$fscrypt_v1(&(0x7f0000000000), &(0x7f0000000040)={'fscrypt:', @desc2}, &(0x7f0000000080)={0x0, "90f352e54acbb500953aa3f312922f63fdcdef04a9bc505ca06452b18885376a912d89dc56bb1c077fd9f501b2b166fb65c818d24cad10b8243d95a16f2b92bd", 0x29}, 0x48, 0xfffffffffffffffc) r1 = add_key$keyring(&(0x7f0000000100), &(0x7f0000000140)={'syz', 0x0}, 0x0, 0x0, 0xfffffffffffffffc) keyctl$link(0x8, r0, r1) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) (async) r2 = add_key$fscrypt_v1(&(0x7f0000000480), &(0x7f00000004c0)={'fscrypt:', @desc3}, &(0x7f0000000500)={0x0, "85d71362e3d9ad3c71804c36061af5d79c0b58f087be4005a6d7406c375ea05379542602015ab751c651646eeefd5b39717a8f08ce678b7b61fb025b140cf746", 0x2a}, 0x48, 0xfffffffffffffff8) keyctl$clear(0x7, r2) r3 = getuid() keyctl$get_persistent(0x16, r3, r0) 00:28:43 executing program 4: r0 = syz_init_net_socket$bt_sco(0x1f, 0x5, 0x2) connect$bt_sco(r0, &(0x7f0000000000)={0x1f, @fixed={'\xaa\xaa\xaa\xaa\xaa', 0x11}}, 0x8) connect$bt_sco(r0, &(0x7f0000000040)={0x1f, @none}, 0x8) bpf$MAP_CREATE(0x0, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) 00:28:43 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x94160000) 00:28:43 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1200}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:43 executing program 2: sigaltstack(&(0x7f0000000240)={&(0x7f0000000180)=""/170, 0x80000002, 0xaa}, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) r0 = add_key$fscrypt_v1(&(0x7f0000000000), &(0x7f0000000040)={'fscrypt:', @desc2}, &(0x7f0000000080)={0x0, "90f352e54acbb500953aa3f312922f63fdcdef04a9bc505ca06452b18885376a912d89dc56bb1c077fd9f501b2b166fb65c818d24cad10b8243d95a16f2b92bd", 0x29}, 0x48, 0xfffffffffffffffc) r1 = add_key$keyring(&(0x7f0000000100), &(0x7f0000000140)={'syz', 0x0}, 0x0, 0x0, 0xfffffffffffffffc) keyctl$link(0x8, r0, r1) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) r2 = add_key$fscrypt_v1(&(0x7f0000000480), &(0x7f00000004c0)={'fscrypt:', @desc3}, &(0x7f0000000500)={0x0, "85d71362e3d9ad3c71804c36061af5d79c0b58f087be4005a6d7406c375ea05379542602015ab751c651646eeefd5b39717a8f08ce678b7b61fb025b140cf746", 0x2a}, 0x48, 0xfffffffffffffff8) keyctl$clear(0x7, r2) r3 = getuid() keyctl$get_persistent(0x16, r3, r0) 00:28:43 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r1 = socket$unix(0x1, 0x5, 0x0) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = getpid() r4 = getpid() r5 = getuid() r6 = getuid() r7 = fsopen(&(0x7f0000000000)='vfat\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r7, 0x6, 0x0, 0x0, 0x0) socketpair(0x6, 0x6, 0x3f, &(0x7f0000000240)={0xffffffffffffffff, 0xffffffffffffffff}) ioctl$VIDIOC_EXPBUF(0xffffffffffffffff, 0xc0405610, &(0x7f0000000280)={0x6, 0x5, 0x1, 0x80, 0xffffffffffffffff}) r11 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000000c00)={0x2, 0x4, 0x8, 0x1, 0x80, 0xffffffffffffffff, 0x20, '\x00', 0x0, 0xffffffffffffffff, 0x3, 0x0, 0x3}, 0x48) r12 = bpf$MAP_CREATE_RINGBUF(0x0, &(0x7f0000000c80)={0x1b, 0x0, 0x0, 0x3, 0x0, 0xffffffffffffffff, 0x7, '\x00', 0x0, 0xffffffffffffffff, 0x3, 0x5, 0x5}, 0x48) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r14 = geteuid() r15 = getpid() sendmmsg$unix(0xffffffffffffffff, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [0xffffffffffffffff, 0xffffffffffffffff]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r13, r14}}}, @cred={{0x1c, 0x1, 0x2, {r15, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) r16 = syz_init_net_socket$x25(0x9, 0x5, 0x0) r17 = getpid() r18 = getuid() r19 = bpf$PROG_LOAD(0x5, &(0x7f0000003c80)={0x10, 0x17, &(0x7f0000003a40)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0x1, 0x0, 0x0, 0x0, 0x2}, {{0x18, 0x1, 0x1, 0x0, r12}}, {}, [@btf_id={0x18, 0x1, 0x3, 0x0, 0x3}, @ringbuf_query={{0x18, 0x1, 0x1, 0x0, 0x1}}, @map_idx={0x18, 0x9, 0x5, 0x0, 0xf}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f0000003b00)='syzkaller\x00', 0x277f, 0x1f, &(0x7f0000003b40)=""/31, 0x40f00, 0x0, '\x00', 0x0, 0x29, r12, 0x8, &(0x7f0000003b80)={0xa, 0x5}, 0x8, 0x10, &(0x7f0000003bc0)={0x4, 0xb, 0x5, 0x310c}, 0x10, 0x0, 0xffffffffffffffff, 0x4, &(0x7f0000003c00)=[r12], &(0x7f0000003c40)=[{0x2, 0x92, 0x4, 0x7}, {0x0, 0x3, 0xe, 0xb}, {0x2, 0x4, 0x3, 0xb}, {0x2, 0x3, 0x2, 0xb}], 0x10, 0x1}, 0x90) sendmmsg$unix(r9, &(0x7f0000007700)=[{{&(0x7f0000002e40)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000012c0)=[{&(0x7f0000002ec0)="b0e16da6098009b54b15f303a3e3c7c9af386d4bbd176f901785f51b54dbe63eb11a7d35c1b8e153bb31ac1737225ca9214156478a1bb6c914c5e29b3bcf04c60155a2663d89d897732314f0cf4ffb5d68f358f07bdd076d513c4aa4e81064442f85fc0db6833be40a59655f47ec2663daca1f3809b4fb016d41e5b99645ce8b76c7e0438e1985a09c2f5f2b45e9700dcacdde68723ae54a4619b16318a2244594207b562de1bd92", 0xa8}], 0x1, &(0x7f0000002f80)=[@cred={{0x1c, 0x1, 0x2, {r4, r6, 0xee01}}}], 0x20, 0x20000040}}, {{0x0, 0x0, &(0x7f0000003440)=[{&(0x7f0000002fc0)="357f18426c3c398352766062a76ec801a813da3c4830462f067adfbdd6ca1ed6644d97e1209c41f5e03231c8a1ee136b144d2baa46f95a74abcf4330fd994ba48cbc28f063e50c5b58e46b01ea32132dc8c038eee25c8d6dbc999ec0d270b9b51c21423fca84e80370a2c7b67a0e98a090006b4460f5e09bf0b44addb4476e396251a50b0f7b722fb700dbb59587442fccac3683a08a2f017ba0cc01ea6634c688f95afd0e37a79c75194b99c3680e35ce33b1666f61864c036a9208c3b5638fd3c8f9b095eafd9229e58f3f79", 0xcd}, {&(0x7f00000030c0)="f1846a7f0701a18a8f76f6022f9710269bccdf8ebebf29a366839b84eb1c684522bec6ba7f50bf9dfc0fb9f276b193e7a27616f67a58f2d0203dbc567771681f27da53ee75380123c3f496d745fe5c38f9f659649a0d6da66ec6daf6347e08c8a8c9ff4e738619b419dcc7bf1c231a414e38ce0d9d07c442aef4277b0b2fe647738b13dd807d29d23fd6ad4889a23a8a1af88e9048bee28fdef29af615306a815262f9ece47e029ccac249919763e9b03ff16db546d8b7bb075284359892577d85a31e1fb60469879a8e5ffc723fe2b199023da93f1be14483402810a194cf8ce5d298c0ca3c51e4bad4eaa90d523c1d6488a7ea86914c5aaf1d22ba05", 0xfd}, {&(0x7f00000031c0)="1227e7ef8e347a2ec0cdea9abccf00ff6a660a0bf824000427129459b5fee80b41e6547b6a72a4a5c25082c867e7eca439a514fd125c9fcdb962407e281a2be2f227bf6c96cb86c2b3bad6d03d0e3d81b72041f669528159a40d175fccf3fb7f875f3564175d30f37f288e00531eb6c17429a1d83c2f836263a13a86d84472edd1b70664d9258e60c7aff6376318fd1cc68cce7f1ab9017b49a03854b751a3e29ef434d7c08feee809beb070c03306cecfe94419882673e3cefc8bea80aacffefd6b9c", 0xc3}, {&(0x7f00000032c0)="4d628b4a8381cfe436e33aae938105c35fc26b06904a2f856cab71d2f01f06cc889edd2201fa805a49ebad28126344b240c287828b65b8a6f3cc1dd0cc08e0662c3c0ffafad60273bb4414b733f76a0063322f9c83802200fecdaca1eaab9527257ca5f3137fbfe9c470059ce462b2120e19bf0ee94531751cbbd8eb0b835f98a8a2374b6a03338baaffabebd75556935dafa842fa4fdb410b2e0521a049f70adc", 0xa1}, {&(0x7f0000003380)="1047f7dc19974b67eef423410097761bad791dbb64d97f3ae42129fbcd7f721d94b0376c2b1564e7ae80bdcbf6af9e83066da8cb3824f71a7a43b81bddc470df787e53138a9dd9f759cfe42d355cafe4b4e4eec9fcae97cf37b89fa2da8dbf95bca80e9beb1d37cc7939ba73e3fc57b2c68eb1b52caa6f4e8c13ff2f969f98194e19ed691087a3006239dee85c67917719172aeadc9c4ea714f7", 0x9a}], 0x5, &(0x7f00000034c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r6, 0xee01}}}, @rights={{0x34, 0x1, 0x1, [r11, r0, r1, r12, r9, r9, r1, r9, r11]}}], 0x58, 0x4c001}}, {{&(0x7f0000003540)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000003740)=[{&(0x7f00000035c0)="e25866c8e7c3e7f8e3146109609d7f9338f67dc8c886e2b29cbcca634e1e3c2354444b438ea9697554437be54def453910e495397581d2883f44714e9311fc0e5c712c1520b5853de50703a58a70d839a774e722053e2efa1b94ec8088bdd53d31525f48d4abab4de929a7aed2f2363dcea6cd00abdf40d16d066a904cae7cd8d722c8ccd1662541e8d94a31b9ab8692c09c1d552a1c0710ba94cc97a1539f5a", 0xa0}, {&(0x7f0000003680)="2efaad7e1214fe779f3d2af07ab4ddad5b95f40450d9c97b2caf236ac00a51f64f1549cbbfc7840f5d7a24d25502a060e0c67e4c790f709c8c39b1511087b5f941e80153ba3008e3425a64745b894567ed02f1ba0113ffb71da9ed1880bac1c09375b4f636e4b3af094fedf76202946a75e1a6d1d13e148a7dc539888b17978f266b00d24db6d0487563fa091479aad8a9f336ec63c42f8587145211e68208ee724a451be2cf36218ae6df78ea0280b22dac97d0", 0xb4}], 0x2, &(0x7f00000037c0)=[@cred={{0x1c, 0x1, 0x2, {r4, r6, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0x0, r14, 0xee00}}}, @rights={{0x2c, 0x1, 0x1, [r7, r11, r0, r12, r16, r7, r12]}}, @cred={{0x1c, 0x1, 0x2, {r17, r5, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r3, r6, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r4, r18}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee01}}}], 0xf0, 0x40114}}, {{&(0x7f00000038c0)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000003a00)=[{&(0x7f0000003940)="b20227afb07c8442e194a6210cb1950d3ee4fb2f5a77483bb8b7feb9c49f5cddf5a3eb24d566c19bc065556d50ac316a6c03fdd5a48742846fe10bc101b75dbb81b826df456cbadacd8d89afa86ce5301c33e47419fcd5c5b46404bd07f9b84fa269f545aef8eebbded73360edd0223cc29faf7f3be387f2439ae9263bc25c93ab40badbaea2d8bf36d063b7a046f7d97cc33ff82328c33d9132335ea5cc9d", 0x9f}], 0x1, &(0x7f0000003d40)=[@cred={{0x1c, 0x1, 0x2, {r4, r6, 0xee00}}}, @rights={{0x28, 0x1, 0x1, [r11, r8, r19, r7, r10, r8]}}], 0x48, 0x8000}}, {{&(0x7f0000003ec0)=@abs={0x0, 0x0, 0x4e20}, 0x6e, &(0x7f0000006280)=[{&(0x7f0000003f40)="22c75a15dcd98e5a6208aac534a7c2b812672a1b951ceb68ea0108172faaa64e45d6b75d7a46a3d78268c86e308c370d5a74ffb22b981d43bbfca27a963bde1b9938ac6722d67a0477a4ad6275a1266b3062dd9cb74bb5310d69734802ac5db6c952de6c56b6b75f732f68831d3cf41f58bb6569759c9001c592fe8e074f7f652417b2dc23e629c7dc52ae838aef51f0f8ede5be6d3fe9e7f7592b9dde58fe7775ced63f192a8826f8771452e7c1865addf8756043608d9660801776d46402dd780fd11c067050ebe775d43e04f4192d7d2c15691f003f1266651f1cafea560dd33acb9fb393d8c7cc2471212d64e3bb25a845711ba99b6d2a20678846cb7b172c9a86bf9a875044f149043298596616c197cd51fdaae95cc8448d6047af272b3adbb65eb40b0c7ce163a4cda024b87e048fa9d73d77a15eb7d5c84cb4ca3dbd60e7305ea4e7e4a38b2cba8db3344b831dd1996d7b6bea61689e56281fad367769ad44f7616f853ab6e2b789482fe682d9198e4014ffdea7dfa5a83dbd71d2c9c12df15a9b958226e2a468d416b5980be020ce855852b84532ba586d9e341ffc70da24774cbe07e5a0658e4b71e579e470b10959120483bbb8b5e0ec4349134906f6cca9fb007727a633a88fc2d99b4a6d177018974e12be1177bb2aa6ceb144b6a7ee14982b49f566440ca7b66fbfdfed39d732e8e1672817e27aff74289b0f935d654e9efe28c3b2a15818a447575f8a9a6e54e5a74e2cdd7e22a6d4f0c5a4cf9b52f571febce885054d7b8e93f4e3848c199a7a9ec79456666505e024d57eae952091cbbd46e7fa0feb280f8ba5aad254fc2a8caba3105aa6b28eaa66356d4a77e776460ac64c6913638bd0e16548df93c9d9134dd86f3bbefae634ef7d2b5890760987b8d6610fa814d8caf47e2d3dc0df1df3b8d93e23841eff10f4d032448d87f15bd25730eae01dcd0ec59f4f5f4d77dbc202f0144f14ad91f7ce9c43364f1d6010c963f7eb2039fd2938686c046c5f41468c023cbaa4c2afdf413fc0611ff493fec016ee10bffb4519f8bfa9329b655b548113414f92e1ebffb4e634adc478f15c71075fa02c3562c1fb82780709bea9af2b958844c9182c7da624cf6e5cb775c4359336a2b2bc279bed1d42d0e50572f6b111224ea421fc65e5a1c81e4f2ccd2d526594b9c7889269c4e0fb27f4063b3b662b7108629732f48ba3d244591397b1dcef945a5fd84ae7117cf60518c9e42030668e91f773ab3bf80ca02395f27ee520c9bad2e08c9a0bec6e81bb0267b48129178e37d8f5c9a207b2df156d1f8404d75c646ade11d1345b77ed03defb035c4633bfce40083c7a443d09fda345d84a473da6f7069ef6c0de50755cc58f325fef71439d935df66efc012999f6c6fad7c9b3e4e12379d965784d1fe130a54ce6b70649a9682d997939f42f835613b21a5b67b4a6dddd52d0789956e5b61d8bf5e02dc015801978acd544896fd4b2b9a927533bbcec59da0618f092e7df744ec4d6f3a4ae5edc63f54c32d3eabefa9d9e3fa677665a4bd86b24c48e01f68114af947b7d7634c7fcbb05d373588c43a53a632add12617b3b794beb2e6cb3adf93e0009cc137c05ad3badb54c92580d3ae3dd19eb94de72013f2fdb912d3d2f6c1dacb5202038e66dbfe06b1c17db933272c711129e2201055548b285331198ab2fa7e8b70b1c6dbc3769c9c265ea54410d23939dd332cbd81de08d1c9dc77d582622b77964f1edd2e53e09a905d2ce17eb365081ecb16c28ebed293fea7913bab3b007d768ff228ace35aa91ad6921001f3ee7f029c7353b5f1989547dfa31d5b50a70b7c8536ed5216cbbcfa603162e5459d719e14f7d43b0144148f6ce15f846ab640de07f9c381c45d952b1db73720edac50364a9a4f65bb4f6656f501a4e1fb181f60948464a8c525971527ae7de85b6f41baaab9689f83e7ddb7d1635f2d0cfec6ea72a8ef8e29e1b5a8a2328722ed136ac7fcfa3a09a58ef853197cb03ebf400678fb5dbd8e6e0f8fae0590a5cbf83406ee8a9cb50319fdc93101e16ce999d3819743554f590f81af6de47a849881f3c4316e36755cdb460cfda2ecf757273b1249547de06d02980b1b18f456279f0dc8a59f18cf8bf54c0c4ab6d874f4003c5875ae17b0f2f9ecc3ead09fe1791101572b67ad156e64523ae70a86c7e4653feb6313203b0809c2286f7c2d37c1df2b067be674a2fb372e118943536761dc69d74ac3e46ec3fb23edb3ebcc9edc2fa904530595801b16965ba825728592a15246f6c91421e2f50e47abc820234aa866627d4e69b16c325a93f277005ddbeb3770a9bc424425e3cc8ecaf2fd0df19230aa85b00059411d2300339fe71ec492176e40a5b3399edcd2cd9442b346943926c67c2da8e47fdaa6ef9f0626dbaf13a01b7a2b0a587550513342053c35132c37b4b7bef71ce417c2686a98c40e74a91ec4fbf1b90a2028200f227e869b2528b5e460307491d36acbaadeee0fc620f2e10fdc2405e4bd92a70d7a03f9ff916521563aa5ce6d057e123005045b52ebe7ac7208f233609c0b0088998c5ea7aa1daf0d1e8fc0a20900eb8bf4d751bf99fd56c1559ab392f6ebd1be7eb624d22ac6d488926196632d969f1285552767d9dc9d2f84453e426375cfb1224a457479711064e36a1cda237b8a46174ee58b83f1a3d2e7b43bbbc837ff54f433e9e5a3796cc58591907d457a085bebbc528178b7e31ea27f83921d9ea25e2deac7180c70f25db8daf528f4de3339c295e79531bc48975af08251b99561cc174d841f371e555ab6239399ec6d3e1904b564694201e32227509e7536c5cd6b0aad2ff6eb9803d6b1f63bc2bfbf108318af2ec55a753f3eefcdfe02ffd2ea69b06f93dbce5509cf9a68e908feab5be81f6b60d39974f6b8a98ff5177c872d80247f98f3d4dfbbdde04f77804bd207fe3b8683a5ced53ff8518c603db4a11bf694f93d894e97c5fc606e6a4f8832e82c75472b28e2571f5a26ba12896d0a5a79f325369088eb33b887316aaca6c5b0b6cbb90239e9b95c737803a524e3e5a3f9c24486a76abf082fc96c5726f2e64978a8579b4bc21d40b79979f054687967bc37b85b3ea111bc3614480f9ea8b3ea5a73907069b95eb7f916cd0cd81df315def1e2c9323da81ec5d5c4a1fe8414295443279b600a6e8eff16f867c98da8ec091c01b3560b82bc78d227b9ef8269d5f10af1bbe4890306548c752fca9f928cb4c5dbe5c30e813bab5c98a26ddda6c2897fe74b32ef6e7073ea94c15e695feeaded5ba0840fc139ae7790acd93f4eff8828e87e1d9cf394a7df9949a7a87be13a7cc221a8de84c9a028d61e43a8805bead494cf6e7c2fa7a210665317e23f4062206f46bc8e03bd4a0fe52718078f4a5f8e4e781288356d3caabaa1f6a7b74e5a331a50ac5788b67b0348b5ae25cf6198d123f410a4b1705dca7cff5e02c4a4b82edd29c602a695e3192fed1deaf79bed016f20250b4e31408b529a631478751c94c914f9ead9c39b5333a72053a8493d2dd77e5d23e18b067a01807a2683692d882d1d4ed967240490a775bd3385f4c7b57e933c1dee916ef855c9d77e34b7f3ebd219c8658b60bf2f719e1498d2cc4f4aad61278f571886abeeea5bf34f72d9c70406724144c926978e2eadbb9768e35341b4e285e2e3510f0615e3aefcaa0a3548df69a5a0da569e0b5c0956b8ff09d9ab877c1e0db3b96e16e91e1692fa1c5a78708dc375ff320b097f41c34e175fdd9142d03184cc56d9f50ca3af250f6edf27424ce234134d7c223aedd8bfa589d0c6b02f81977117f773866a5f31b5f6920ead42b86aadd0cc24cc8790fae11a0f59f61c1d21f0ccad3bb94a2845718f1247ffb9004f06c871e5ff4138d52f30c5a702fea3475d3b9f08688b15f7921b71a48ed90c2c014462c13873606ffc11752378df5c2a2ecd7e90fbde369b323c8000ad444dca14b5b6f7f4628b688f3b945c4bd3272d74877c3c49ac17bea1be3213dfaedfae1cee5b5ddd7e602c93b5f1626400e0d043e02f0f47bc46a79084e516a7c075326bb588b8f5a85d84edb4fbf13d687cabf556cbcdc5fb9be37db4d5cf7e95fd49b910420084b6c4e91c6a03a5ab5ffc33e9e9c6dc67d8a1a1ebf4bb7e1542a16b3333f776de3be2aebd28575d34df4086ebe2d6c5c9e359b9fa224bf26f8ea45b37a9fe15b72eeaab2c8a9b517d1962601cce471f3caeb6d16d01894616841fda6bf0b1d453b0d972a4ffdd8a5a87ee7d7bb17313eca6f8872f624b16435a8769b2608444a68e4a7790633f65c5239a8c89cbc1621f4b4815282aeeaee98b75e21d37ed6bbee6d5c288439ba035b8828f0aa008415b18eb00852404a95626aceab8d7cb236cad5460ebdf95a06103806ab486efebc7f3f1d94a2be30ea14d7814d4e48794cc7b324b7f55f38842b5af06a1541b428063ed0578d99715b992e3dd6b24d869e26ece6822c85246f6fded15d054ff2c6c780637c2e37e0a27fca4c56a8167f58a23701a854f4660609b3584cc323a0d6a8d389ca8e27126b6ea701aed79fe32df10ac58c3d3cdd37e7d9c680495c0e723ddb1d4c787030bef309f519227c7c6610ffcce313a92cc264808dc2145f5ce77f2e5d8d174171fd74ccbf1baf7d3be01b566ae0100290b45d4e1f871b9e40abf05cff07e3271eca88d9e485d55e09de9a036293f2784e4bb06ad66beba63a6a5d73c5751aa9b9b5945f789478c86078293ee1a414acb0797f97e0a2657d023ae7294e67c5d8414897d2f91af2758186f3863fdeacf5f6bedfd7796d25375837303167477ff5690cc908fc3cc584c0fb51bdc7c411d35d2e8d0c05864e7612205fa1d602654dc31f5d6ac09f43ba2ccaadca4a9fdb445e20b08ccf8a58990afa1d0cb109f7a65fe20d5dd914a54fb2fb5b8022641a4e3ae8f33967571a947b42e6b041a0d99934c796bc7458846a6459df8c6de4ecb3cae6beb64793b68f43f9f12a1e8c2eca8b81ec7b33c0028ccc40a5a91a30128d15bff27d3a7e491cd9d916b6f50c05cf29899fa2019a27fdfab39a9b0fc8b1983ef061aedbb06a4652d4c9081933ab87d9b68cae11967608eb7678f49d1472f24f61337b3733aca9062ae8b1b1d7c8c611e522f53306aaaac86b1f8fa3559bbc3594813f9b95d70897753a74d8c28d97e48e753ef84cecfbf246a58c5aec8039f8c594f2f9a63d499a718a02c519045b68c0985cb7920200acdcf6a5b094fbc054d3576315090e4177ac0acb5fec5d0c2e73f832da1dd2d2e6d2ae41c3a669302a8b3dfec8ee555a78b9392cd8affe8ba9d416310f19a4a032a22ecc326d26fe09a4a5671a7aa777312cda43089b1ccc554ab10a0cf0a936e7e531f2d3728195d0953444c7f7e4a26275132f74a85725f98366f32b121d2c49cbe3db423249e207116265224bf63570378eb7e78838e15731f957c8fd2de6ad62d396eb2dddd7493e371a7025f6d2f06ba4d682b556e4f0c32df396e6b0e78a22805291c1da587a6e34427d4fa4ea56b8c81f96867cb987bd6e65c570f4e0e731799d07cb3adbe295f94ca9e4f326664fb96a0274953aee8c240d86716f9dc2c3b06ba6dac5b5d5e75879fe1d0f053a9406440d766c3d08eca6507e077f90c76a3249a90ce253cb0e4a719fff542a0a16041588429bd23cf4933a3a8d731656c97700f0e9825bb937e01cb75d8fb7c1914305ce98bbe91d660e66075d34d441def624e2e4df7aed06d680956589b43105f2e2fccb18e0", 0x1000}, {&(0x7f0000004f40)="0e596c46979ebef535948ef7920decbcce4a46a6996f2f7b257c1ccc7da99ecdd7bea6b65b8f4fc3fc9cd33b03c815e841ea7f8c70e09e8df12319f7685935b96bdb73b2c2129615be0695426ca3e2416872e00289c1a0d8765f969103e2191036dca44c9e9df053d7e8471fbaa9e96c491630b746b4b13a93d0b25314926f430be320eddf9cde44577f7567e49c22075b40321b76c7aa7461436ea0d79af7dc8b7fa8c8d7705a23a999a932a1d28a2b41e65510160602de0b0b37a3c6e2404d42be8cfb4671b0a35e01097c9498d1e68ce178d60d88e0cb3d19f544ea1c529398aeba5d3c93b837f399ca580fbd74c9246f15", 0xf3}, {&(0x7f0000005040)="4ca956103ff00eede538b09d68387086df46b58965fb436cb5dcbbfdafc6be51d7eacee59219a067241dcbe7efe7d83b9efcae92a21ad5214f06e851651de7876992cb97b8a0914d7920f53194b6745611906063a5aec8fe8d7a8062bf16143826c589675eba3bef0ad66d7a8f0d4b55a57c1ba8a8799814447f3281d3dabdd62c", 0x81}, {&(0x7f0000005100)="202976eca3403546431664a9874725c0ae8a9c52c98ab431d9f3dd4b41b4c70cd65958a8ca169e42f11aa59848fc06221dc71a45cf395adb24931e903879b92f1663ba5f40f0f1b1da35f7487f9fe6d74612855bdb8cf727b2b362962a1aa5d3", 0x60}, {&(0x7f0000005180)="21bcde6cfe501598d1384c7217dfaa19d274cb92b6bebf98d09afc456a060761fb63c5482cefa66616b3955362f393bc1d3756d8b0f29342868af5cff825373837fff0174870c51c88c9d7dbb767edd3d84840d52877a7a96a7fa3cb464d8c373b30b60d45863a5fc67e8d74af77328d9724fab55abb49e77948f48d2c7c8526a6f19cd3eda957b4d32e5c59175ef1c828abd821b421287c7bd0caf83874262236e21c3dd9382e690d9a28440e9b24f28cd5e42bb4d3b4f60236796b3fb817e4d20c346527e39c06746000bb5c4f42fd3aea32887f447bdb4571be5b9cb421a934560515fe28c2300fe31061014ed00ca3cb8a06216821eaa8821370992a5aa4f59fe7701714915041db819a1b3a7ce08d29181f9f19c7dd43459a8fe1b8b64511afff4b3a7e9630f2f34396ee9ca8daeb27a4ac9d3c05ac69f5bea18a05ba76fc6a7a1565ab3804f22ae9f7d975b9d0091e8cd18f1f7e4dfd5406f1d1fc5209256be1ef8f86b33a61a2733d969f5d49ac05542623e432bbac70d49eb4a0862072d0d4981d932bd0c7c0fd7df228c0f1dc83ac9982b468a98bbdfc41eb31ff59ec4577c0109bf9641a787e5c9a524a48723cca6ea3f71fb6d39f66c0fdaf2c726cacca15c06b480a089396dd10ee8015714c8d97253df8d42c39e9671a178b31081057ac188c90d080c18d6704154c35eeb38886e9d9e5995db71e70822ac5bdafe401d14a37cdca71f0a8d31e24c6ad6d7f5318490af2dc91f5a28e70759224aaa4c85a10babe5f253f4cbdada8d3fc144a09da6fcd9d26696cac73ac5157071bdb185686d3344d2c08dc96e79d8b16528961a2c7180ecf2f0dd0bc8e1de0da8ce88b8a21a4b8508649c61ed07b16247ef1fd97ae4f1f83d45ee7dd449e0059ac724e8115c94b496f41c43691485da358c9a8858ab0d5b25abd90f6a22a60f8f1cc9af40a48d0f1726399b3a8d904d7a12d1ba037b74564e3be44c4aa278a8678284a10f4de57e634601416372f3cc496666225d00953c252fd20c259961526060e1f7a417cf1cb02889300c60bdfb3dde0310bce41090ba04bb9c5dc2b3883180bf161be1742f23385fc2d0ca61c5192940fb5a5ff2ecf00ea5be78b360899b9a2e57afbe510673c91b1ffddb372cd0fdc2d479f8b9504d1e49be871ce5207521a60080332feac097acdaec17d5efbd3a5e23abea8bdddd03f7975c0424180b539f979b759684faab0a7ebf70afe6e04a5072aa4e881fb9520d9a7d687c127d37d3c38164480117618f4d57aec78ec80511eebbd1a75b53e2ee34e3b4bbe590ed2de0e31c3b5c8dd44b4696b5cfa6a167e9d2a148895ce4679546aaffae32a43821dab26eaf2a70656fcda640870378ccd4e6df52799035708ffc9b7bd50af946bfe7fb8fcd91a82507b13ea1a88f1a9b3a7a22eaa257d89fec97f32c664ccf955a042a7c7a524841ea854182673abc0e2799459dece15929945f9dc391889fe5da9cf20ed09376396c76cd50b4b7ae5463a0ab8f4b93da026f9cb1668e57ade668945d3305c16c65e4b9ffebe0341bbdba149e1e8cb5b6428eb4fd8873b0edfd4772c728b722fcd7813979a87efaf0008514283a935f9454d3f18608a2fa812cc0018e56248a73ae1dd8650aa84d671dc58ab06b36ae97d238ba126d1f393c899a7a4bef268a2bd1ac4cbebc66110e37074a6263f92e4d6ad0143874c378752a108bfceccd81c170d9825e6c51f611b401acbfa456dc705ed33407865611da9da4e6337417009441b3f91121461a8783df089308eae34d813471e3d4b1279be44a8b1eaff9c98e022b27258369d06fa14f6d8faed2c88e2c8e3527ba3acecd40af36e72b0c50276a7b07d2919fdafef60827ad973b05c3584592102ea32528805c17c86aabfbcc53427c0d7f51cc35b10725786dd3cfc61802427ca37411c4cac3d5f6124dafee16056f7f6f5d8cda2ee8a1c2ba15b721a295f9ecea220a952e5834d289410728bf64c1ccbd3573e853f139fa47a67198bd6ca4e54a3b1d6f0d2e45bde056074644bd14d1e84ccc738aad3664c7ccf38e13690840d2d47174adce92cff8e8e3c957e99f84e85ba3b2c5accaca850c76e3cf1b56fada7edf2df5879300f7b57e3e0fac3ad2cf051e26ec7c50c7eadf6af37286a7bdcd16d707a574b214ce5578c788d7cb37a37b528f75c6ebc3a80cee211348c616d3b60e08817bf1cc2ed9432dedfa2475aecda38b92278e8950cb31c0b7670673b79c8921aeee03a76b28e33adb097443b8adcdc0630a7f027ac2748ba8fb359863cc21518fc53191f2c01c01b09d43c4e3ee301fec641963e2f3a6d98275932310f7ed72a4bbeab0a35d8bf09e19535aa191d05ef8def3c42e1d2b19885c26285b14e36a50f85e7cf00912d78d776c90c58e46db1911bc644a960c59c4f9baad45a92ab9bb0d745f1bfab9db0f4ce2a0bfcdd47209b1f6dd53488d389d497fa837e2ecc072c72b10ff1f8025019770619317e5fcf38047fe47965d5076b9ce97e20831a7b156b84ffe64fb3c9fcf356423ee150e0effd7160deca2d397ad004ffb2be08c9001c3d89139ce5efaa5a72278c9d7c21a3e2ba622fb59920121f1bacbc0983120886af290516f8374b07837b9c2141cf0a7711c925e66746f4ced48f4d039b20aee192f0c345076b6c020b8f2342bc2a0ed44d064cb5e388fe5c50dffa2ab70cf134b4460432b01448a4824b2fbe6e37adc5971fbb910e4363039720dd6ca50ab10a76e6019e98e6c0459958a274a94617660cc40bfcf3ad9cff41014e743119eef17dea1fa84225f34792a0ad27349a26cd03e64d31c5c60f939ccd2ba52f2da4e94702e6a714cad6f163cddbbdb30c1bcc172c760a84d81b68cba9ca086f31b494e28544eec921d9b9bb6e4be8bcc44f7bb800ce1e30ea5a8001838a562b2dca1bfd00ee1b5facd2731f8ab9edea6295adc48d5082d8d6a519071a157f7cd49bd3f236ba139594e3bf4c0bfe9e784ad815045c1fb44d20a2c6e454e90c9c981a2bc3f720b15ff4c2a7490c4576fbf88e6b1c12bea6573cf7ce6f47f004cbd3ea2b963b9a7e94c32b46b73dba33e4da84dac9c1f9203d2685b0dffb14c1d95c17e83c4aadee3ebedba83f22fdb972d5b2a231230d4a59279919ccd51ab88da699318638c5cc18fd82d84d7fe45073f75b8530ad248c416a7f2678a11f84528059d1d839bf0c3ba75c453eb6ea3195bd1ce02b29d49840998c3877141b1162c292029c17bda3825b4ee475248e38ba014d7ad88be60a8b18bc838d1501f7846d9c956d77ac6bce25b9858cfe60d276c74c38f13289c2a53343ccf69db886f517d547bcb58653d2f24eab323898744fb4d91bb76c9b901b0130cc3e6d8f181a8eb65c14ebabd37e4914f80922363aed8b968c8ca2c14eebd9c5c7a7fbcb4d40fb14cbf42faea496019669f6fce56d266eb6a58ec6f0349fa6e246f374463d46ec603303cfd7f158892d082c7c1d5590e3d329b162301534338db0378d2698da21b74493f46815b6c529a5a9f018d7e29b7ae737d2eda954175ad2b848a99b5d652c0fcb34dfd5a08a0f0272a7608dbcff46f0f093cbcd509eaaf3651c22903c47c793c72dc76c7f00de26911988c55779299ec6a714b1511f5b90e5c398c051a8f2cb65f834282f7e5f4e243b3c3692aa3aa07cbaa97232deb95553b8a490ff1ab1943ad8d99d8c7b25a79ad15a6b9310e3ec7333c00a2bee5b1a54fd4108a60449bba0fc5a64b1f4baeb3b8a554cdd6a0a6f5de209dcbdd9ca157b93796267b7a7938c7bbd20fbbe4ce45161e8b6b0d6a9d85945b2877b3b1a9d830b7b6142c141a9db49da86c982662f52f212bf881e01659022832512fa4f115c1589b9eeab0f9720d1139009f6dbd6bd7d032190698dd456529bbf51ad9499ed9af4e3abcdc71122d251816c805d1110727c8e75e13ff5d0c5551dfc263aeffe866839ead4e7823a9e810a8c2c46ea1562d69566eeeaf8b6d4782094181aaafbacd1156bf1abafd78c732caed3f7a90fca140bb8ad2d237889847592afa4d7f043c08fb6eb7b13d93bc7e4198e24f80c3ee9394d1cfbdcbd0e71193619b49cbbb6e2bc2e25127f0431d743667ad3d03ee335fb09f37f22478606e57d6d7748003c81d9b97aff77e60a4ed802765e5613dc65ef29b4ef99c5153d63543e8a438209e8ddcd851d658ee098857ac9e64cd61c80cb161bdcf91576d9f588fa4fa59ce10a2d4d3e7b16513bf5026e26f25834924f764584379c39e120fe6acfbe82d288ca431da535810405f631c1e8178b95d2071a52617256d81307f0284fc6518556b0c37ad61542c4a8b7aea6292e666b1bdbf91dbefd98e56bb3b53d19dc7e773bfbb0b71e337ded1fea2ee6bdbec49a1483b4b1e6d8e282c535c84308a6b0f2b9657170e4bcf958575e25ffa2b1167f51d17c8efcdc6ca6a4091883c48289142fce258114afc20a857bf01a153f46b5447b3a3657c4d82719e22862c10ab894d6a387a4fcbee41a7bae3664dcbc9c843d721484f41a7b32907c81b7bbccb20eb76d640adb482a729fb3cadf79e52514dc757bedb17290994feac473d5e33b4a182426cf60d67cf670db8ae943e58a2ff22de89757add9f95b18b3db393c6eebaef5b72ded18c417f7021bbcb7b30fe3ba3e7b19c67e8ec7b68d726cdb869b96113449db930b4e9fb39ad1728e7e1843981638e17c3ca66548cdc9a9f63b8b7cf4d4c9a42d551d6af8fcbdac18f09afc7a5d8c1944f289cf88df3cdb7b57ab5e8ff72e8179f8fbb5d346a148a11ab6236e47db53b0da8dbeae745a4b3a91af9ce88bbb50767f170a27ae9544caae98ecebc535146c0d669782a50f0934c03af2875ae72be1079b649b4591fb5a6b1a83a8e25b34d21f376e9ef141e508b4addf44cc29ff0007f3b7fa75950fdbb4d8f0f8f1ba8005b70639e4b9e9e39316a1dbf799a3a1003177c27093513f46a5221e7f132e1a4e821e02fc5558f8618ce6c7b1cce76296cf94e347341a2a3a904f4b509e1e6c1d86deda555a565a47fa3619cb980b01372c0d6f5f3b073b52285db716dc4b64b88d20c578a822cbc2331afa4ec563765d938de25b4b3cf09b4f701ecc6a194edbc98ffe8af7668c6a9ac222c6be7a314ec6f68fc6b9a024f306ec7e66aa4990f3cfd00ebdc46ff10ff3a00dd331c2b269f29f35755c1b847aa0fcca78e1a9794d0f6482bdddebaf214925d3e79b7c5347eafa8581b18ccdc7f95b4e753b0bcd04c48a609a64899aa20b5dd3e81a5340f1cb788eccc4f0e3961f34feb1c15b26db6cad98c012f95a9265c808bf1e73cf30c8c28bb4d4d3ff22dacbe55732d0fb65d95890a498f6134aff3af9c106b72ad9329c99b133ba35b5fd10243ac550db7d1202255fc9ebd6f6ec9e07b8b5a8f9b548f97d8d203ad0406d0f128f4786bd74c7c65c5a23afdd1f8d7275dc3dacab3b348391a0039265b14d630c28375dbb24368ce9fa0e26752c0a3a1f13839806679fe18601a6b41e407a7e7d6042d741b0dd8dac1c38257baa8e9dfdb2b086dd9b426e83a88f37fe6f273738230c133fef0f57e8033f0511c0051cf19f4285d6cf6b7a5d7f7b8a04a0bb45a03174b5c145d1b98264bd8d07247358ea46a3bc85c5ff7a5ea73cac2bb4582ecc1bc46e031a7e89260327102a7446bc1c8a6d11d2a6f377d5465656c3092a51b3cdf432dfdc965f4531c4de08bf5dcc533f3e5db203003ef4d63694507e6482e3db1d3a894a10a13b8bf218c2ea63dd55f163c450", 0x1000}, {&(0x7f0000006180)="fd9af5d9c78acfdd97f87336aeb1bb81dd377703cee6e14806ac33e886eeeb8a36f0ab62f58242abf1880b57fa2d593780e7f9fcf3f2981414d80f9af48bb91ad41d143861057c558bbd32f0b68d467a37e15700fc18769349166c430e6ebd4cede8247478a30746b4dafe153e6abc07e72ccb5af931892286fe878271d27df1b3333f2fa1603124e85cb89160a05e6a0555e507f49f62dd0c5eaca604873c61803c04e1800e11f1be959034b054960d3822162dbc1e546601fedd77d7f6bab4ff831eac8bcb711034", 0xc9}], 0x6, &(0x7f0000006300)=[@cred={{0x1c, 0x1, 0x2, {r4, r14}}}], 0x20, 0x4040040}}, {{&(0x7f0000006340)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000075c0)=[{&(0x7f00000063c0)="506b2d53d79cb56b426676734636b20a7626a8b2c7a0387c2e4e60477b2ecfe6beadf5bd3b6df83f35312dc9f3e8b3a12fb2f102bc4ed929ea17068ede7a9d08eb2f4ea4830e41652ba5fe87c407571b284069278923909c29b7d6341aa61097019405167c23f187a95172ee4da7ebbe300c4695964abd5c42a7a75e66ebc15ab55cac9b50376cf5bff90082e56bab0a03476f", 0x93}, {&(0x7f0000006480)="4e36ed56e215e97db961808c0eab0ea46e1b6dcff2c319b6c759149ce3568d890d111c66e64ea2b73e89306d90e862af77820ab6cb5c68f5af3d37a632", 0x3d}, {&(0x7f00000064c0)}, {&(0x7f0000006500)="8d4c835a809e83c5b6185905f3e54bd0d061f1eb0814058a84908aa8532f6b3278aadee9a29510a64bf59e3aecfb281f2f93346650df665759957f07a490e84b1ffdd5b977fa91133584740507de962e324e9e3b516564373d1177a0168a33c88041ce047edbcb631915cf58adb19e7acc88cf8b787076d075fe2b3c7c817714dd608065fb7e5be90f21b22e3556792317a2616b0a52738d84d6b107052e4982d7aba96e34a1e8b223a45ba96f595750ce60c2328300f9bbbbd802270e695dab4d3f7ab9c6c3ab061cbd0367f6a991455c132e02d1a22dda2394555fe9a28459d88cf9aa45812253462606dc8c119083efe3d801d5f76d2b909fb5d1bf6809e3ab143561bddab732d609105e2ddb540992c9ad0396847f04d3c70a189a831afc25666f290d1f42c1a9d9f68ca631a74329325fdd6cc0429b72bc05d33720e288013a09691983e5dd638893b856eed86314f619b6af358b36145030b69452e9f8a5bfb07694b1359b92194ae39e46e85f1d42e03c04fab17a4538775d6ffa3b3f375c92599a0cc0e2fec8d6c8f2e8c4de8a51a7b61806f48191c1e85266b284c7845e1e903af011218241111e0044637aafa09e9221d5d09aa42560a98398436f575c3f3cd08cec2809605c1068749502aad4048210e1ebeb70444726f793d3491c902dd51369a3c2df3a107851704bd5ccc6d9a19604b22e9adcbc5099e854382b4d4ac4c22fadcad72cf0ec3eb42802e42df9a00cc7defea1f3e94dac64c8abeba8b22abcf8040c137fd6ae87b5f1e9eee88025f16734dea5ff678099b84d0d2977d5161cf5f0b02c5a1fb54aba2a98535da3fe1957ba8b10b557c2675bf965af9751995433fb414f6e56239cdf23ba5a97aab88b9a3cba2b1dc29d130a786109fc2c82d214a1fd6fe372635997196bd36cc3e339006319787dd07a59da33e4db600a1eb0a9192a0606c7ef31fde1b616229a70f59f274e32fae228a9330de3bc7ea51d6dec39eaaf737e0e5b221210f826e7b8ac475d6127359cd892e08873622a3aff006bcbfa0c23727c7c71b7eaec7dfb372887f62898914116a69af3228076340e437eb6d5d2f5852baaf1382299ed80cdc4bceb68a2da53fe569b6075425ee6c064b7cefbbca45ae0f73ea925be1aaa83cdead228a068dfbf500b0fca75e893c6a714bb7b04abdcf3912a3b3df10278aa0e8a20814fe12bde1e549dd5ad88be60d9d08991952113859ebc658c2f5c9be71ed437cf085986f1a7b6db4921c9c1b8d8cd5ce69322a968ceba4da29469da994315b013ba75b6aa833b201ba5dcd5dc671edec6a10b362669313e47dbf417ba543ce02d7b36bc5b421af58776c6982f222a3aba677ddd806e92932b269cbbaa181f677f2226d0ed0897614a6d8a8082c46d295b2afdfd8ec057be13c40ae15205f5ed5cbcf576d92f01393b490b0d362839a8786aaab0425b0a15a3c31f0855e413185bfb339e09c4ed60907ed82a505ab1e5f5c8feeabc72eb9d0a4a46e0f6a24fa23f3f03ca5274ca87c5b5ecf0beec70acb262b1e030309ce0dc904f8a29ca59425557c1fc75dce563668b3bc63bdabb54f8ad3fcd699246568102a729ca28b8cf2a847e6fbfe967675034e18758e023e14b4ede080da798a0ca31a1efb2200cab02e6fba243211cc77719697ff68f33c5dc6dd92b0165aed4737849dbee3321aaef7cf9ce9d826297eddf2c8e6ced507d309cd25a593b66d54f1e217f24ca5ee466d7291af18ae7f39d27531f8ebfba1cd763958a2a85220234b6306eb9832786a1d8fa785769f33b9a74cf2991fb1e20f36384e5a7d610b84f5ee47c02d55087db734cd3597668eed121693d76421efec8ce7941fa2bd7fd58b76690d797b7120c2a65d67d9cfb5f00358446e1626d4aaf62b18047c42cb57442b1241846e3994543a4c786ca23bf6152c55dcda8812e154addeca6671408fc540b098af7c524f7d2003e2404f3bc99e0e8688d82f7b38ccfdf5f66fe928c266adf655442b24c81b32d9e5819b66215b3c46896c6db87b0aec2ee346d5d59b97992138e81b0b514bd99b84a55028382fc37a9b5bfbdb23f108f7e6c59e7efc6497a14ac41251f868a22c3011017aff619dfd958f1264578888092e9abd789880602ec62a36fcdd74c193f0d66a53da8118c0dd5d862ee28ad0e21e188a73b168e8b897dcddc4c9c2a20e028c2cc972e16f7139e2d0f0aef5851cc41ec0994156edd145595a99d20ef9f674a918d216be337e85444c1123d0a28f497d45b0e230a17d64bd750ca07a5d98baf48fd67ff7f50e74c51ee5c03f0f3e72291e0c682805b1b42e1c0dc2948e8e6b194c7a587d57e5c643bddbaecf5bdbdb10fa2fd06e0c1df0add2536f3909932d331865841125a9d52c5ef64ae32945dcff4fa93f3653dee33ef0930df71aa37ca1ad1c4e5f53fcded49d5a862548b034b90ac84862a2501d26d8fe737be7f4f67d3c03823a8d0d3f4cf22706e6592306b99b0e1e1980892fa649d282b732bc3b405734ea2f185f87b96694c3bb0781ba5bc3a18b5f1ec7b41f783516b217baca44420941a421cf876cd4b0ce86f84791c8f971cb64f961bddfd1276e21973425729e0a6bc688e6cd30fb28100b579c9ed6c94c1d89b51f16c7ccb62eb38827f5d8a10f582f333a8d26af8442d8c8b8061c0004dc6e3fb53bc9dd9cacf4fc34d5d8b7fb1e8ecbdd15e60df89aba38ae3128892205ced6555f41b4eeae552cc3b59ac6fa3338a26a3427728d599a510e0f6ffe7fa9bf33773e4355e34a853fa1b28f3b35be5b8c30b96b41e1e7abc5e519edf22cd18f0ffcac5b36dcc7b6393af083e4c45656e92b384c9211bb6171ee64ef960e7a2020d2515a71ca3249b03ea4a22fbc695a6019e58d5ec4ed6ed625f31115f2cc7350514792b4c7e09d469c270a83fbd15ec9ba23bfeaac362fb9fe8058c33e7de9557c9399cedd2960b338589333e8e359529ee0d04428afe63ab5cee2eb268acbe54f11043dd89ce1af5872f92aa3cd265a827cace6bf30ff95a640805c526830e7cb176afb7946f4e6cb4432d582e63bdf79a80cca7c923211c0539661f9420c600136708fe00784a0bd4749faafd0393d6ba9ab52e8cf40d46630fb4ba92ad4a61727a9242307b7230c029f367e67aebda84c5d29161249b3a1cac08ba36db41d405ed5791486763a20fe56684a9a2974d6305a26d5ac2382044bfaa3c5325264e9c90a90014ab5049f118d452a1f9fda05fea050ccbd6b8f9d3df44b5ef2bd394348e6206cbbf80252d7902d644220702f35d4fd3afd395d8b4531c2cc5e73aee6e0fcae6e182e24a07bcc6c4578faf684b995b0d5b37283c9e1f3a1520273e272353aa2c6dfe29276062617d555932b7d087798f183a46359a436aab5f2ebfe0ea1e4d1fd6158ba6fea8299ae069a006f37ec2bb9f03f3c78843db490e7efad8d49792428ae29663732e1a0ddcfdae32d4231989ffb6f114f8410860829855844237651819c4778da4ee20a93d85e1f333dc72d770bc06f3918129412dca763e97219f9d61b1e8e898a82cbaa22398123266cde13cebf146b5977787e09708ed427f548d52ac8f737af02e4d70dd4cc99014b97f99b9bc313ff100cfed94317184944839b41ed0d322c6939f9ba8eede2be8ac8081871ed2c40d9cd0ee3521d1c7a6cb1c4e126645e0f3a06292347aed733731e59df8a0c999011cad66599cd47ef2b5ff31f63b1386c01b00aac628dd80f940f9025b61895972c28e458a5e4630fbd4f799eda409a6428228c04cc9621ce8c81182d1b473ab8adc107875c16a3f906b9452ab367951bb826a7fa4a7e9e6d9da1d41ed8ef352a1e182b99a519f421700a3571aab78448aeb4533e8417c4e11a25927131edb96a9b4472d4b02926ae2940746be29ba36bed958a5eac111bf63c968c182cded22fec8c0db6923ab7a3b65cd5987f6c6bb5442748b746ec387f7aee08160f3f3832fabc1fbba3ebe432357e9fa0be97d5a3359e74ad86fb5365725aa5e5059ec802b80ca7adc7cf4cc837af4e8e42a2eb72a914573868385d76a1426de8733ab058b2ff388b882177e1d384691ed4292e26e53bea764fafbbf83fe80d5eaf00ef3ca750a2c7d5a5c6be2ff0cb983808e025b4807b04c8d51a480cf83ca2fc5585cc36f480ffb3b3772e6ceaa93641cbc30724e339f28de0c039b9dc6cb7be7d5988bd1a291bdcbcb3c9105481c6b15ee56b31f37dcec512127bca719de9e4f46bcae364b73f3c92b9b9c31786eb3b91d84bf4394b1cd51f5112c9e330a3a86b1fe1208e7d69fc396a952a6534a58e6401b3c0ba4bd899cf7ce94c464834a9bee07c4b5c40de0f9b1faec0fc8a5f9ee97e1a3d8e847dcae729bb6fac7a501ff7bfe847486b1c01706d6d6f5875b4738f194ace523ed13efaf5a9964b7e0da2bdfc62fb172dd7c73c5042bf767d36dfaae9b33d89fd0d482122c379cd1ff4e1f8284e23d4053d8d57f9007bdb29fa5be13d4876a11b21a5b864518638094d756449b9e2c373dd11d55f8709c1e963074a64cfc25017c8ae8ae6fa3cc6ba3342eb5bd4c20b58d3bfc5cffb89acfbd068f6e95b38c7b68ba9a4cc47d8a08a37df56e60184164c13a33821dfbb3400782a7f6893ed6f6973d56e51463aeb349232665a760cfbc0a723791857cc0f0df41b3d76d817c878b7d88a96f836fe841d94b38e2fcbfc53e0f29df0fcf371b7e2fc3e079f309e9ad4f5d48c29d4537dcc4d5844195ea0633496e18552fbd4185cebc363ad7dc197172d1d66a9f3eb60928345bf05c6dc18c11bf69eac067bd3fd7a0e63408a43de74b3ffabdb20c79c7c6607edd47f5d45a0f92c636832196dde9cb7ab5ddc98718c28e0cd7b6b21c03f8923f3e5552e4ed2f0f2809c00d4c65bf050d6e2de49fbdc056f4bf0c967aa5cadb79317cefadbaffa267cba8e95e59e5c71023110ac81a8737bbe0fd396fb9ae1583390bbeb8f8c317062fc489438c230cf7caa91ff1b04f8cd702ec7f7d6fcddb03b9f987e6f1d375c8cbe20d421866df57d8f6946ee4e87e3b2c23438b6cc7164c633e230e455748c53a74cc8b5356c7a1eb01339bb19031519bb6e916b9ce767f8eaf3e7ddd767bf9e4f316b4ea95e15943842a9cd3853f7f44f363c87b0b80ba4842bfb2b935087a2811f3890eef695098fc012b4a29098efeae24eb08dbc6a8aae9cdb21ee0fc3ae7cd7bb272b65e4221b62f3db009812d7c314bc698dc64c9fe2a647a4830c964e2cc3d0d633cea27eb24783ff8dcde6ac16d77ee26909ecb3d20a0c21b30865b20b67369e8a07877ad0a1bc0813c12ce9f6879e9ac79674bf261515ea0f007c7c39db0d81478c1ea0aa31c70d3544a5d1f93451d1ae267a796f6e05cc3cea484479e0a15aee4732700dffa0471da81e56b984eb076461b9da3f8ac21aae0457df4c6f20192fb35acfff0b98cc9dfa40d42c47d6f0948c528ba2d65f5d6e5134e4d4ff7f09861c5ffbe65a6d6ba335b8f5c92a660d711fb89ba1340340e0b16c5fc04e442092ee3ee3bc5ca4b3042df77fe98f934625a284b4c58c1403a7408185d36caa31bcbbb3f9bb4d37a017c28bed82d045776473a9fd80cbc63861f21d75d3a72d19554129ea1db44d5ec1a063304218346369897da8458e50551b746c82a292bb6c5c777d3d3e15e870d5b2b430e2c303b2f491e56bfa2de76cf14034852c0ff57d278f31d28bce96a202ffcde7e76bab3d1a7b7003f5c880a9fe27c531d2ddfaadc20f0d6d57b969e1d19", 0x1000}, {&(0x7f0000007500)="dd19afdf8a5da5a3f08db50313424d8d55f6a8963d573faeaec0c6a9e29c257b2d35acb0a87e6fb7bea8b250e1e6d4d1cc08099315408fd2da1f164463bc153bcccc67bd73647ff8464c98a92ffb26726ad4bfc6b668d3804e81241d8d4742e940905009207a472c5fd40fa3790719ecca244d8b2afc144c80f373d06cc5a4cf464c8472914b49efda31b0f07d750cb2ee", 0x91}], 0x5, &(0x7f0000007640)=[@rights={{0x20, 0x1, 0x1, [r12, r10, r16, r8]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r6, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r14, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r15, r5, 0xffffffffffffffff}}}, @rights={{0x14, 0x1, 0x1, [r11]}}], 0x98, 0x40010}}], 0x6, 0x40000) getresuid(&(0x7f0000000d00)=0x0, &(0x7f0000000d40), &(0x7f0000000d80)) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000dc0)=0x0) r22 = ioctl$UDMABUF_CREATE(0xffffffffffffffff, 0x40187542, &(0x7f0000001700)={0xffffffffffffffff, 0x1, 0xfffffffffffff000, 0x1000000000000}) r23 = socket$inet6_tcp(0xa, 0x1, 0x0) getsockopt$inet6_tcp_int(r23, 0x6, 0x4, 0x0, &(0x7f0000000040)) r24 = accept4$unix(r1, &(0x7f0000002bc0)=@abs, &(0x7f0000002c40)=0x6e, 0x800) sendmmsg$unix(r1, &(0x7f0000002cc0)=[{{0x0, 0x0, &(0x7f0000000200)=[{&(0x7f0000000000)="c02bef9485e586480b3ff511a67fbac1d8bfefda2ccee81a769ce84115cd28fc9f17f2503f8d8d74136b9695d329ec00723a67e013349b5b940981593759372c3e0d4bd25659725e1ee43a17ba3f2a4ff0ffc85750e4b52e80224dec32140a5f4296d833688d69ce70f36f67f87291d4abe3d5a8d3a3a33f6186d095c09570092106f2e1c7074d306fa7d6fde389b2450dfc9a8219d61f5fb49c96fa5918d5c2ec6c8176543fb7a6cf09f3fe15f27ed871", 0xb1}, {&(0x7f00000000c0)="4d0b8e6145ae613932d872d99c8812e5ffafac225b1d4a2fefaffe2de3e6a7188ef46e07152dfcd9f398", 0x2a}, {&(0x7f0000000100)="2960eaa18f73eaad074cdb3c5fbdf9731cb48319e712338d5af1cd80e1f5037ec3e5e7d9ddf57b7ed5d1a4f59fcc894a57e5cd8762050c00634f4bbc3fa93776727db6feb669f8e462a9e9e80eaa2744ec73edfd8482a45d3327d9083b33c9eb13b6486a07a3db378a25c2766d657a7295cb7354e13ac7c3ba630b3d632a97934d3984ea3f8aee4b4751469caee1c7d99c565afa4b9dfe99f80fa363e2e1f2edc8a530a60ab31de4a0cfd20bbaf82a20aff5184d5f1140d23b51f02116f71e1f808e06dd2825287db9cdfeea1ee0f9f607dcfe6df3cdeb7ddaaa3cc974df3549a702e8ddf5e4e1ed0939ee46b567a5971192cb6c8d6f", 0xf6}], 0x3, &(0x7f00000002c0)=[@rights={{0x18, 0x1, 0x1, [r0, r0]}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @rights={{0x18, 0x1, 0x1, [r1, r7]}}, @cred={{0x1c, 0x1, 0x2, {r4, r6, 0xffffffffffffffff}}}, @rights={{0x14, 0x1, 0x1, [r9]}}, @cred={{0x1c, 0x1, 0x2, {r4, r6, 0xffffffffffffffff}}}, @rights={{0x20, 0x1, 0x1, [r1, r10, r2, r0]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r3, r6}}}], 0x108, 0x1}}, {{0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000400)="9539f1a6492ca3da43b1e82d1716f68e486ca284c1fa282a99f52a4e1aa83ed3195339f9402577db61ed91f4d417ad93258eaff5170ef042618a9ae2ca441a0662a77299ccdd58853134af3fc9242178bd5fe4d0", 0x54}, {&(0x7f0000000480)="9791f9dd608e613ee69421713d35b938b0157df9629c6923b45d9ad1eb11d82b22753fed8fc6ae77d71b409b8fd8823021711ca161a3caf77399eb399bb46d9578233a4743ca5e2ed9852f43ade97c0834b106eccfb15629fc2c239be9ce8c46ce109d40f3562650ca9d5a631095288b10bd97477df15d3d4a6ad7281bf3c262fcf9ffc04081a2d2baca5035928a3dfafe73791b5a72f0e04c7bdcb37357c3", 0x9f}, {&(0x7f0000000540)="c4193c9f834eae0789dc663b8d918333dd988d27b06a6aaab223de39a3a32f35804f3f73661ce237f5175dcafb6da0139f2f4d580318aaa798c87151098cee62ceb969947f30f4a1dfe2f9d6422530e4582c43310f", 0x55}, {&(0x7f00000005c0)="1af816bb93a2fb9cb8fad6269fe533dacb63d83abf161eb17bcc55630305dddc3f20887a619c8f78efe3f417f87f61f3608a9bd7cc82ed4ff0709f", 0x3b}, {&(0x7f0000000600)="ffd0e7106db2a09ac252d011ab6e19f006ca25c5012da00cda44d5ed16821c5f33ae31aef6a9490254b6cb8c6dff39b10c069cf0c69a35fd695a0fa18e3d93773d49a2b34e63cc9cfe2987c371ddff6d23ab873b9ce09df3794bba24214567bcd1a62872f0", 0x65}], 0x5, &(0x7f0000000700)=[@cred={{0x1c, 0x1, 0x2, {r4, r6, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5}}}], 0x40, 0x880}}, {{&(0x7f0000000740)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000b80)=[{&(0x7f00000007c0)="7d526a8f53d1b7f9ff9c6e0cc8a348233d647bf2391a2aaaebe00fc75ebe83f4307a87a3b3e5f264f65a2cc142435e6ee5ad84e512ed1c2d2ea07d4f24d362eaaf14dccae6fe", 0x46}, {&(0x7f00000008c0)="31153960adf13dfd7d09e3b80bc58ba4f0a255a1aac80ee378e9d33ae49032", 0x1f}, {&(0x7f0000000900)="81dc4a0a25ccf70210dbcf9092ae65aff51c2dcd379ed421357c77c002f4541f6ee4afe1f918d2af2d29d722ed0ea7d887029fc65873a9489e673db46a927de4318c", 0x42}, {&(0x7f0000000980)="0e1bcbe9663bd572060fe422f00a751617b492feee8073414200ead55c6d27b9a96a84791b2675f01fc692dd2290518be47a267f8ebeeebdd58d62684b1ba56febf6f39c0b46a24d33131252b5c2abcf2f08c1eade43bff3fbc044bf6b04433ebaf76ec40243b10375db50e2648b83cc79027db792bc68d8b175bd7603ec5007280d1315cb64505d0ea87e0bfae5e78ea366fcc4597809ff6524b8db191f4003fe4f442a22e9164b7fb7f9878a693adfa0cc367278383bb92d84d29c2ce4cff8f33ba42d054a45b16ee39f593c70ec06c184f2d8367bb5c46c79c98730dc44ece919a07798807c9ede3475a8e946", 0xee}, {&(0x7f0000000a80)="07b5a345215684e06f32735fce21b4a86419e71593a3b33a7a8e51e42d6c5741df2d7757802bdc3d2bd4e6509016505aeca2a8810adfa7ef73afe07fa5707c6f75f6f17855d882f6bbb75878588ebae5f41bb4bd21362b067741b1cc18b8a49973ba64790ab3c5f8ba4eaf3253b6251bea2d", 0x72}, {&(0x7f0000000b00)="00909a243334def0fca7d663613e10f53654fa693054ec216060b2063a155586f39bd7092ab190611aafb24726fc7ace2d8b47446a7d53ce1af08f4dc39db191a9790d92057f", 0x46}], 0x6, &(0x7f0000000e00)=[@rights={{0x1c, 0x1, 0x1, [r11, r12, r1]}}, @cred={{0x1c, 0x1, 0x2, {r3, r20}}}, @rights={{0x14, 0x1, 0x1, [r0]}}, @cred={{0x1c, 0x1, 0x2, {r21, r5, 0xee00}}}], 0x78, 0x81}}, {{&(0x7f0000000e80)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f00000014c0)=[{&(0x7f0000000f00)="ad45dad886c8d44443c9293a7176cc35d006a6a5e331431c1c609f5a96203ffe9d4af95138a2e17185a53a51517083679720b5c42ce0daacdd59f138cf", 0x3d}, {&(0x7f0000000f40)="a0067ac1d3522cf57a949d57b34f1dabb7020d8790aabd6355bce48aa69fc684298238af0bc37dcf520e081cc4e20d9b96c54345ef06708c86d653c87bf73f90ba241626ee62a5ebc3db7f461ff1b8ed1fd141d1bf3d9e1158344ef9d638db88f414bd401c2584fcac7f540b7451cd580ab1cec5f4bb9e9ce0eb0870b7cbd9dcd2118f04c3e8421f1f24555c42a28819f259667d15a8169525c643383b8dfaf0b847a5eedbb5a2cb2bbbe5a9c32e", 0xae}, {&(0x7f0000001000)="c4ba7c566926434a77b6406479351183181b084b55d23d08dc12d2e6beaa73f79239f58c43e90b6f4a406aec3e8e85da6a7a61095d3cf7e2151bee09c973f524b834263468e8e2011fca6e8630607f0e6d8417033dd61101d8c2218cbc4bad0e520cb5d729d82091529ec80843b0375b6a4933830b79ebee9d014e08f96ee175b9d0fc4dbc5526495fa10d9be93b735aec2d22dcebe054a26ec3188aa6cbcc08fe3fd6967b0588135d18cb777aa291ec924c202ad2354c4d64cff3da73d3ba5aa75b2745e2782de010681df9cfe4954c0b77b24c6af5f637afe076460da65d26c18d7a7272f3f01a94a71a3060c9661e1eeb9ed1e14fdc007f", 0xf9}, {&(0x7f0000001100)="d77c4d29078a94f7229db5a56da8069ad74c97841e40dc9b76ab3317468698d3f4dad49ad137bf2814f0152332ab558cdefc5b755bc0cc787eb575d41934ea376f9e944b9bced783a9acb9b1b7f2f82dc372dcc3fe661d42970b702593a3c4fe28370a2520c7619468c2bc671d65097471f9f8f2e92a927dac6d1ce9ba8c103ed579813e01102c65c728954c8ef7986f47f80b2b6f0a089e2e2ddacc83b1f27e1f50d2", 0xa3}, {&(0x7f00000011c0)="5a03b798bdad94a79341d7546bd282c01329c2717f1cb0b5cc1a94255a93f91addcae7358a0931ef7693a1f8fd3c925766eac2b1ca3ce92d2dff066b7b5a8f5066b8bdf67ccc8b775b6d859d60813de87cdd185eb7cb9adac60ce94de8375fefaadb70f3a7114f791a787f5f7062e647a7ba7b1f4934fa791c58682f564953e435983a128c1f20b73e64add0ace629d5fba86a340f39798fedf714f7efb4dbd8d6748f14da3c18517078326056dcf089", 0xb0}, {&(0x7f0000001280)="74d8749760b627fd936ba17d7abc8768da56e7dbd3083d343aba22412813b666a5ca9979f0", 0x25}, {&(0x7f00000012c0)}, {&(0x7f0000001300)="2bdfb7d567b7dd662a24794ecd5a54ebb8631a89a2be0aaa677c0162e23d1dc460953e7086102ffa96f23722df9e5f3f1727b497d9c999956fb5334f8c8562bdde37973d01e202d1ba79bd0fcbcc80c5dd2d9de843d224280a80f76f01826eb60c8ea4233fe9fa2addf7c9173a79c56788b12c51981199ea90a302013951a55ec1137ef7e2a07214e46552", 0x8b}, {&(0x7f00000013c0)="3a7b8380197c10bc400ce042548e4b39b918bd0e8d8933a9005a9d1772ec3dfd7b61c005319f8e304a3bff96fd4b0a0af76e73a4476e9b5a4ea3e194bf4dc07a160ea079aab8c46c145a8058f448f480895d678581cd2ada1a461c9efb8d7395178ae25a1633339cf29c6d74a7fc1708d48b5a005028a67cc9fd", 0x7a}, {&(0x7f0000001440)="29c47a594d4bf858358add8cf69b145ce9ec61a5efba8fe6c3365b6fc29d60a9c5d4bf001b2b7dd7d67bf635c29d5bf15f59e9c4800678bb1c5a9631b7fb782aa54984784ce870b543aa1ecbbc5ecab793894c0b82d509936510d91eb220a3ef3f680daeddc1152dd2511ab67edcd075e8670c0dd618e3b4e4ea56bf", 0x7c}], 0xa, &(0x7f0000001580)=[@cred={{0x1c, 0x1, 0x2, {r4, r6}}}], 0x20, 0x8000}}, {{0x0, 0x0, &(0x7f00000016c0)=[{&(0x7f00000015c0)="8749764c6f24f9503afee5e988ca167857b8580bb127beee950afcd5b35b0483e96a8b3456ef815fcc5226e87ea7a1a89ba9b276fd4a541aab6b7d63d49ad074a765e26fa50e3ffdec730e14bf39f663bdc9de6a6e17f25045bfbc1254abd854cdc4eefe6055a9bb5cb6b5438c32c69121b42f3526226de7496519a4480c8c8f49433d9679641c0c380fcf06235387a65840607e96e80440b1d2de1196a6982599140c06a81a2f22c1252faaa7464049893a0fa42b9f461cfb8885f5e1d81984eddd37c3e5f09335231b84862e2205d27c85d10e720085b3e380144c7013f35487c74c24fdcc7a8ead84ec2ee2c572725fb8620260", 0xf5}], 0x1, &(0x7f0000001740)=[@rights={{0x24, 0x1, 0x1, [r2, r1, r22, r0, r2]}}, @rights={{0x20, 0x1, 0x1, [r23, r1, r1, r1]}}, @rights={{0x1c, 0x1, 0x1, [0xffffffffffffffff, r2, r2]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee01}}}], 0x88, 0x8000}}, {{&(0x7f0000001800)=@abs={0x1, 0x0, 0x4e22}, 0x6e, &(0x7f0000002b40)=[{&(0x7f0000001880)="1e1277", 0x3}, {&(0x7f00000018c0)="f96ab9533299ae9f9c7f2f3b65a18cad4b8add705650168f7986e6f4e140c839ff758b7cc14a2a4673359d2d2b4ed84f6f31be8917ee2a336e787c4d9af9cea9537bd3a6d48ebdfba470da9eb3cd21eef3a36f63a368093af7adb9d18a1b45b59fea56f4bf4cbdc23ae6da47f44fef5ed5b0baf042eb7b941909393a411d08a15f613fd23a1fc26d491ff5d2394f423189ede01e2c7b", 0x96}, {&(0x7f0000001980)="d78b74a18fe39d34964d532bbc575d829b71e41578776e95fb1ce7c1620ab2a0c067375de311a0a350a1109bf00f2dbe2cc829aa452deccef004e7cdcdf4cf41df3cf197752d7e958879165548af01cf351ef271d1bde58df1d00ea40bfeb9163aaa6d89768c55cd9197d3cdab5ddcf1f9b7d82a23dbc89f8c452861b42014c285673e89c29f5003f5daa3bc5cdeef0446c2081c3b3f54aa744970447a232c42e2934db03f690dbe0a748f525d21caf0a141a8f956413e0d6e1eafaaed5d219c0ae0da3e03c6fb7423df032c994284d0f5e50c995b83a68672e3091588c8d08eaf66eec36764672f8c37f57a86455c635b25920b3bfc260adcbef3596be7e0d7e9622bd9a4f5b1da3042da11a35c84d2d72f8c875560a1b04d494ca43073e9066f3c891de5412d20840ce42c6c0034ab52fadc4fb4af6bafbc682521d9c37985dd1794f02436fb451fb539b529dcbec3315593249b6b177c15502ab0da2149d2a5dc5feb33899230033273daddab730cd2cf07d73e5a55212ad7fed88e3d5dc5b27abff8298ae85225b6a808d7c80b431e8bd70ad514766b5cd52c4bfcf5ab6c8b828d34fc4c1e6b69a986a7f43d307ab26b04bd6880d72eed4cff9ffdb831cd0fb4b99c13ec3ba9c7c15149598cfaa0cc45457bcab11285529336544329bc0ab6eac64c6295821c11eea45340ce3c052cdc923d9cb7d498ee154fe4a4b9b3d4dfa9c08cd9e44fa1dd83928117668c51b18815f8fa92e9554a93e6431bdd53328ca9b9a8cd0a008b0742089961f3aecfe0b631d90c4416dedb744ecfa8011a3ee58d3cfa58560eccef8fa244e3460c997bb051e33255c0e8253ee1b4c3269f954ea5f7220739a0aeca687be70738a3e54389f2d863a11f8e4ceaf4d6e48f4b3474092debe8dffea5c23a72eee7bbfb8a8edc2404251559d9560c8375fda60aeb7f10224d3753a9fee5732a75fafdc5e18b7ef45ef1e1d9103b774d91a5ad536b2420f47b49be4e68c6c119aafddaaa2b85e8dedc77dc8cc7eef022066a6d4e7f080acf071b1bb40760f54f9232322e03cdbf06028bb73ff76582321f776c111e98798dd2a6aa75de0d7d68ac066faa551687a316ce18f2bec08d2e7fdf390397318cc619bff5060c344dc080869fad49179842152b7490726f1057906131825d36cf7a983e7473a215d1eed51405647663789877f67e60eb8be3e90d2cc41ef6fba68af5d9e430f02bff50851048f977c4db2d4c82a9eaedf7170dac5af13e1221cda7d81990c2a39a301761d6193b9d2e5159e19a538ebb0de75c4bb16141d9a122cc483b215a798e8fc2ad52bb693b401234beb84fd0bf4f1024049644d06abdea6a28e88411dc43af522a3c19ba253a2cf733d69edf783fc499a7e75a9f5ca545e00867602eb17d97f8f9ce76b16d0adf14a41c918711dd8d404c88de9eb9865af3024ea922a9a0ec2a265da7e6c48a9c37f5527ed9632a0c1d6565383db909cf24b70309c1ba7f11d7870cd9dacfdc770891cad17b59a06ba103aa42aa6434dee4a5abae90a2611363975f9e37ae3ac90c174f130c61b764b5938fc8da441f2305715c46203e6f1e640174b816d07e109f8488c2397a8411436b62647a85207e83fb568335f12f00b06fa008c11ec853a7b41c2f6e9d18e9ffbf62f157202b6a4bb238a81f7e4ab02c4092befabc8af206ff9efc0a80b8704385d4f250b7e138353cf859fd9b28caf72b3c392914f4afc2cdc907125dc7388537b9884d2f2f6817df12d444a1b3b2b38163e3e42e39ca4a061ec13984be010cf6a0d861307d2c2d4cf8c6906c8407023b4a4ebf2a387bb9e5e428ba5d21ed7dd0d4b0120be47e1880b3a1e69f2443b71a400f63c2a1e8ac9b938fd1b0aaa21841f55b59e2d9ab7f23ad7fd9db4ff0fdd436d84677243ae06f5ceba5f6259433db0470bccf28c79bb3c1c5f11adf1f11bb02391df0cbc339d38a4e04e0cb9d9e443a784f475c452a5e7a8fce6dc84492c90b10ae911e6394af6fe6f6a402879112a12266f7ee43ece62890d2846e06be14c5c667bb99c3e242319894aa6133be5904b1b5f56314e8dfe4cfec357f08f554cdcd6915930e84b7e5e247f5b23676511d7a4c4be19c88dc5e8baaad8ede995c50f5267b818949b132a18f52890ffe64e2a4bd865006138528793ac81480af5fb10914a243b2ebe69cc885df7ca80c98801ab87d14a2a3a169f605fe0c7318bae98fd64d37dbb2ee1f92192664bdeb9f75e06c090782f113fbc1404f47a39ce6b7a6c863c2ee509bb43491b6688f3c86b2d592ccd2183d2f197e9446e34fbb9082eb35476c60bdee319a281c26fd8576f33ff227bfef7a3cdd4b0984f2a324aa1e14472cbf64d7f430f5e86fd96a483c0cb575e26f02083f5d008957b8661ff0b0d8d1a0c1fb5a3cf935f28ae9249a4ddd9f0b87de9e353e8851d703c8c63432121b823406029970a21d4a7147c37358b8475389931709da3b1bd477f663a5165ade99159f8950a40bcd29c3978e9efb4da81551e6cd3b815a511fda02b54f9123dc6147785706b28232f9f2385806835102c1f5da78efe33f9a63aeaa74d448d2aa2ead2dd7706ca87c54cd3692aad24236bf06b54ebac4c26434cdd8173e4e83590ea5d452880fd43546ff5154993dbf19d64690ef3b486145a016be6bcb86d472ffca478bfd2158d4dcb534601f904194f719c584e1718f183a4c60de17c586570293106c8b0dcbaefcb3270ae41a0577f721d83e88847409d1375fd73e7e14522d991c5072642a03f7aeebeb4123616da7fa685cae1c01eb4fd2aafbdf9c87fb8fd64a1871bf8d8d32a3a0660a6faa6262de7a306113ff708b12b5d98c7f6967d971ed27c679c131bfd7fe87754639c0d687588b707935de5198c955a263c2b464ae14c90f6c1aeefc800ec00f5442d8d2c45c6136ea816f6779fb2dacd350bb1755045e30d5c5868f16e420c6f2c31ef5104ab5b3cef8532ae2abcfeefbc768b2408fff00ec6652b6bddcd77c20f240e05face7d7b1880ffd5a7f00b44b46d3bb19f97bd5d711e7251dbeca034ca18def136dea2d67185dd2e9284c1b081f17e22ff478720003a93f9f094b7d8da6f2610a15017f86cf0cfe4357c3f19cc0cddbba1029da3db905c03ac79c5df153a61e00b25a808be3be4adafda67cc30e96b4135a254ae0b7290876cf3071d3dc79224fc6c4544e0b96112446c5102983f28e224351259108e8e0647dfe159935de494974f566237a4a88773d1d77251f0a619782b3c769ae27693bbc63a38002cd3b424d4a0798bef2f1c44742f3216fc40cd9c7f5d1d807cce0af3c6f86651ba445ab6c0e396bf09afd8579893720d05c33959a9a59bf9b2216c1704acb3ca4ddb9990ecb805a3e852cb9732c689ffd548859b6b29a59e3820235b38683064d08c99692e85e15717596cf919312d6f6cf3b1cebdd57d6e46c2a087bd21f5e3fdf1d977eaf3d3565387617b6489900e7b498355c9e1054bc150c4f98f000b1b597146ba94b723c825a1b7ea5932332635371668baf7cd85fbf737e3b1c36bd5806af7a983b96daf9f2766855e7f6f72ade0d33bb52d220fd074a382f2adf15a17b146760cbd69a9bda1f3f3bbb7c2aeb62786699e65a912771f974637e40d993c705caf331d0c855fefa13369c002a7b92a4865bbe099525f095922b52b29e83d2a3860a2908d14992df9689c9eeefccf41632d23288fb9cb071fc0deed070e1a6e48c200a138f9dedbbe01713ae8ef1088cce621e389b493a063a798281ab78d019e4baed9c1bc84bfbfd635f59d4c1645dcdf5c2be249c018b7aef5785d5f4ece085d19eab502a0e4fd6d86acee6c6f120ad967e9ca3871bf553ef33698c41c714441d344f9963b9708b4f4c3c478470b10de77f238323a777da48a8dfc4bf5266de51b916943631a89ccb562556d43a35a4f20a3c9a83389180b4d3749cea8be823343bd8a12dfb4aa1fc21c6a0e37c3524913a088f9f5c14e7dc98acd0d10a632ac019723e4f714cf61bdf424f6f47db5fef3e1287f25d75c944a86fdda49ee16b4baf0d142b3715db583b7b369012f4e7d830bb11b92af75c949e6223ce153e8a28395602462528af198514ef7b433d673fe874c2b707287744bb94b05c2d4bd9186ab7f68d3bff4e7d6da3f07084aae0da8bc6a4e8a9fa386ceac57579060fc9dc2e2daebf998704042778563228817a1f2b5d477340d0c021a4a672c3808a2e7817d9da29b4d4687f58d266e0c920caf6b78e331c65ef2411d35289504e0743f6469594f8b976096e5b7fd694fb6e51bba98ffb01613a246feab56436f2d2b3532f3fa1d2fa5048c84ed35b5082c145c1458f95b7394ae392fdd27c37f7e8df6f03768be9bbdab97affca0f234370efda3cf2d1fdc3ef8713529c9b561fc2739b86190c7968c6cab03aa099bbb1d19dbd1dd00a1b766f440637511db1a47a55b3afca56ebd96a321710c94e86477cd87083ca7d743f9a8b2e5a5ea7b7810d31980fb1c9c1b11a5f1345e50f4301b3a69722ffdce723cd2aa0300b27c3140bb40ae25c1037c73f83015ad9916a8f39e2fc642aca5f319594f3d59859fc16b4a28433bafb342e21e7d4cf32987a4f4fed4db1b2221530ef4eea01044b37f9aedab893d3577e176253836ba813eead1cfc05b46a319584346661f33173c274803ff067f9b9095137128e3dd17b5656a06831efd3c7bcc84d3f7b31da152264c6df7911dff24e28c59a2fce4456ede804cf5f93879a32619623625ac48a628a21c2a8f1d12803a660bbde39a971d68bc3b57dd7219c440be87b9d24d68a25a4d405d896bffd0566186b65cdd73499e6c90b8fbc3f6ab200c146cbc4e19b58e9723df2e2c11aefc036b66b70aaceef92e0fc70ab97a327367a5f0783228aae256dfe56a5362b14fcfe26661c669c39b8fce3e5c5f4163a2e23d322637da8a7a819611c526670f087e0af1908c4149c3a7df38af8f671025ca666a268808fc5c529685804a7405d157e6dbc8ef274cfe1da88657e38bb00784fce3d49e1211afe77f82b33632acb1d89388ed949649b44eea06b4dae6dfada1159b635af94fce185187f0acafc7c9130cd984e98bbdf422fbd960744fb583e4cd76e815894a5f42bfb15f4277f54a96a047fe1305885d0dcdd6b37ecd2893ecbcfa5939ddf64a83e17e4bd78f4979553b1c97180270de7e9964b672b77375049aa6937490f2c81c5390e064828258db93b4696d404dec9d84c79e4b4bededf474a08080c00edb2057b8389b7100cee012b99a81cde3b7fc22a1c26af52a18340055eb7391741b30db092af1be0f0a4613be4ecb204ef24808ae4e7d0dab08b130e0a77af7b95a96c16033bc2b7231dbcb9ab8878c390796bc08a57e18e632b3f1e9dc8b6b589db96e46bc28e70833ffc512917da4bb054e0735d55fec25b61b17c0d2632467b50e673d42549e29b1bc1e528e9ce29e44cced2ca322ddf2c6bdf1bce9c99d0b58d1de7516a5ec212ba92d04316954642c1f031af8025bd048810cbf7b155f3cd0610cc4465f0f623c0515a0c760ff33504107006b88535bce4d5cee2a487a28cd4c663a82a5a7a9ffcea7879ae8e36f7315024cc96a17b8abcb11c960a21368f008b813e4b0cb8653e9a2683b17ae28cb1678353a426fc1bf56a9f506f33428210859ac1dc24b5b9b9efa6d236ecd2fe0bbf9d6cd11acee0b1ea2cffc5870aed6161736da502bf9be6417e7fb4469f0e76a6dcc29adeeaed9d7cc43dce1226e8fd78c2", 0x1000}, {&(0x7f0000002980)="f58b6251a63a0364ef6d85bf7ba95fd9c89ea6131fb5a6bf82f4cab5fcc5c4c6ef4a3b10cb84d23bcf81f944716cca974b3f4182644ea87893ee0db426d672738af5aad99caac01169b1901583658a255e3b6289cf192438b044b9ba1b6512aaf0da8dc09ea54115f138bca67b1af3d2ebd03519088fbc0b0cd4ba1d907bd0b402c6ed", 0x83}, {&(0x7f0000002a40)="7b6066e5bb915d6aec25b3b9ecd359fad143342a232c6619f415f524", 0x1c}, {&(0x7f0000002a80)="fec763a6076bddf7b82a1d95f6300e40b92190de8acbc84c41cb98bd916ea63310d097402a6ca7d102ed535b05026459", 0x30}, {&(0x7f0000002ac0)="1954c344cea3b4fa405c9aed30c3d3ca65d4c6aedb572264e09ee23049878bb5335213f22a657aabebd9efc9156c5aac4ea5aa2894e815142578cb67af5e731ba6a37b3552", 0x45}], 0x7, &(0x7f0000002c80)=[@rights={{0x28, 0x1, 0x1, [r2, r1, r2, r24, r1, r1]}}, @rights={{0x14, 0x1, 0x1, [r2]}}], 0x40, 0x11}}], 0x6, 0x40000) 00:28:43 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4c00) getpid() getpid() getuid() getuid() 00:28:43 executing program 0: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x97ffffff) 00:28:43 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1300}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) [ 1722.564848][ T4865] workqueue: Failed to create a rescuer kthread for wq "nfc3_nci_cmd_wq": -EINTR [ 1723.364679][ T2247] ieee802154 phy0 wpan0: encryption failed: -22 [ 1723.368775][ T2247] ieee802154 phy1 wpan1: encryption failed: -22 00:28:43 executing program 2: sigaltstack(&(0x7f0000000240)={&(0x7f0000000180)=""/170, 0x80000002, 0xaa}, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:43 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1600}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:43 executing program 1: socketpair(0x29, 0x6, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socket$unix(0x1, 0x1, 0x0) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) socketpair(0x29, 0x6, 0x0, &(0x7f0000000000)) (async) socket$unix(0x1, 0x1, 0x0) (async) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)}, 0x4008804) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) 00:28:43 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) (async) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) (async) r1 = socket$unix(0x1, 0x5, 0x0) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = getpid() r4 = getpid() (async) r5 = getuid() (async) r6 = getuid() r7 = fsopen(&(0x7f0000000000)='vfat\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r7, 0x6, 0x0, 0x0, 0x0) socketpair(0x6, 0x6, 0x3f, &(0x7f0000000240)={0xffffffffffffffff, 0xffffffffffffffff}) ioctl$VIDIOC_EXPBUF(0xffffffffffffffff, 0xc0405610, &(0x7f0000000280)={0x6, 0x5, 0x1, 0x80, 0xffffffffffffffff}) (async) r11 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000000c00)={0x2, 0x4, 0x8, 0x1, 0x80, 0xffffffffffffffff, 0x20, '\x00', 0x0, 0xffffffffffffffff, 0x3, 0x0, 0x3}, 0x48) (async) r12 = bpf$MAP_CREATE_RINGBUF(0x0, &(0x7f0000000c80)={0x1b, 0x0, 0x0, 0x3, 0x0, 0xffffffffffffffff, 0x7, '\x00', 0x0, 0xffffffffffffffff, 0x3, 0x5, 0x5}, 0x48) (async) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) (async) r14 = geteuid() (async) r15 = getpid() sendmmsg$unix(0xffffffffffffffff, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [0xffffffffffffffff, 0xffffffffffffffff]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r13, r14}}}, @cred={{0x1c, 0x1, 0x2, {r15, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) (async) r16 = syz_init_net_socket$x25(0x9, 0x5, 0x0) (async) r17 = getpid() r18 = getuid() r19 = bpf$PROG_LOAD(0x5, &(0x7f0000003c80)={0x10, 0x17, &(0x7f0000003a40)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0x1, 0x0, 0x0, 0x0, 0x2}, {{0x18, 0x1, 0x1, 0x0, r12}}, {}, [@btf_id={0x18, 0x1, 0x3, 0x0, 0x3}, @ringbuf_query={{0x18, 0x1, 0x1, 0x0, 0x1}}, @map_idx={0x18, 0x9, 0x5, 0x0, 0xf}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f0000003b00)='syzkaller\x00', 0x277f, 0x1f, &(0x7f0000003b40)=""/31, 0x40f00, 0x0, '\x00', 0x0, 0x29, r12, 0x8, &(0x7f0000003b80)={0xa, 0x5}, 0x8, 0x10, &(0x7f0000003bc0)={0x4, 0xb, 0x5, 0x310c}, 0x10, 0x0, 0xffffffffffffffff, 0x4, &(0x7f0000003c00)=[r12], &(0x7f0000003c40)=[{0x2, 0x92, 0x4, 0x7}, {0x0, 0x3, 0xe, 0xb}, {0x2, 0x4, 0x3, 0xb}, {0x2, 0x3, 0x2, 0xb}], 0x10, 0x1}, 0x90) sendmmsg$unix(r9, &(0x7f0000007700)=[{{&(0x7f0000002e40)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000012c0)=[{&(0x7f0000002ec0)="b0e16da6098009b54b15f303a3e3c7c9af386d4bbd176f901785f51b54dbe63eb11a7d35c1b8e153bb31ac1737225ca9214156478a1bb6c914c5e29b3bcf04c60155a2663d89d897732314f0cf4ffb5d68f358f07bdd076d513c4aa4e81064442f85fc0db6833be40a59655f47ec2663daca1f3809b4fb016d41e5b99645ce8b76c7e0438e1985a09c2f5f2b45e9700dcacdde68723ae54a4619b16318a2244594207b562de1bd92", 0xa8}], 0x1, &(0x7f0000002f80)=[@cred={{0x1c, 0x1, 0x2, {r4, r6, 0xee01}}}], 0x20, 0x20000040}}, {{0x0, 0x0, &(0x7f0000003440)=[{&(0x7f0000002fc0)="357f18426c3c398352766062a76ec801a813da3c4830462f067adfbdd6ca1ed6644d97e1209c41f5e03231c8a1ee136b144d2baa46f95a74abcf4330fd994ba48cbc28f063e50c5b58e46b01ea32132dc8c038eee25c8d6dbc999ec0d270b9b51c21423fca84e80370a2c7b67a0e98a090006b4460f5e09bf0b44addb4476e396251a50b0f7b722fb700dbb59587442fccac3683a08a2f017ba0cc01ea6634c688f95afd0e37a79c75194b99c3680e35ce33b1666f61864c036a9208c3b5638fd3c8f9b095eafd9229e58f3f79", 0xcd}, {&(0x7f00000030c0)="f1846a7f0701a18a8f76f6022f9710269bccdf8ebebf29a366839b84eb1c684522bec6ba7f50bf9dfc0fb9f276b193e7a27616f67a58f2d0203dbc567771681f27da53ee75380123c3f496d745fe5c38f9f659649a0d6da66ec6daf6347e08c8a8c9ff4e738619b419dcc7bf1c231a414e38ce0d9d07c442aef4277b0b2fe647738b13dd807d29d23fd6ad4889a23a8a1af88e9048bee28fdef29af615306a815262f9ece47e029ccac249919763e9b03ff16db546d8b7bb075284359892577d85a31e1fb60469879a8e5ffc723fe2b199023da93f1be14483402810a194cf8ce5d298c0ca3c51e4bad4eaa90d523c1d6488a7ea86914c5aaf1d22ba05", 0xfd}, {&(0x7f00000031c0)="1227e7ef8e347a2ec0cdea9abccf00ff6a660a0bf824000427129459b5fee80b41e6547b6a72a4a5c25082c867e7eca439a514fd125c9fcdb962407e281a2be2f227bf6c96cb86c2b3bad6d03d0e3d81b72041f669528159a40d175fccf3fb7f875f3564175d30f37f288e00531eb6c17429a1d83c2f836263a13a86d84472edd1b70664d9258e60c7aff6376318fd1cc68cce7f1ab9017b49a03854b751a3e29ef434d7c08feee809beb070c03306cecfe94419882673e3cefc8bea80aacffefd6b9c", 0xc3}, {&(0x7f00000032c0)="4d628b4a8381cfe436e33aae938105c35fc26b06904a2f856cab71d2f01f06cc889edd2201fa805a49ebad28126344b240c287828b65b8a6f3cc1dd0cc08e0662c3c0ffafad60273bb4414b733f76a0063322f9c83802200fecdaca1eaab9527257ca5f3137fbfe9c470059ce462b2120e19bf0ee94531751cbbd8eb0b835f98a8a2374b6a03338baaffabebd75556935dafa842fa4fdb410b2e0521a049f70adc", 0xa1}, {&(0x7f0000003380)="1047f7dc19974b67eef423410097761bad791dbb64d97f3ae42129fbcd7f721d94b0376c2b1564e7ae80bdcbf6af9e83066da8cb3824f71a7a43b81bddc470df787e53138a9dd9f759cfe42d355cafe4b4e4eec9fcae97cf37b89fa2da8dbf95bca80e9beb1d37cc7939ba73e3fc57b2c68eb1b52caa6f4e8c13ff2f969f98194e19ed691087a3006239dee85c67917719172aeadc9c4ea714f7", 0x9a}], 0x5, &(0x7f00000034c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r6, 0xee01}}}, @rights={{0x34, 0x1, 0x1, [r11, r0, r1, r12, r9, r9, r1, r9, r11]}}], 0x58, 0x4c001}}, {{&(0x7f0000003540)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000003740)=[{&(0x7f00000035c0)="e25866c8e7c3e7f8e3146109609d7f9338f67dc8c886e2b29cbcca634e1e3c2354444b438ea9697554437be54def453910e495397581d2883f44714e9311fc0e5c712c1520b5853de50703a58a70d839a774e722053e2efa1b94ec8088bdd53d31525f48d4abab4de929a7aed2f2363dcea6cd00abdf40d16d066a904cae7cd8d722c8ccd1662541e8d94a31b9ab8692c09c1d552a1c0710ba94cc97a1539f5a", 0xa0}, {&(0x7f0000003680)="2efaad7e1214fe779f3d2af07ab4ddad5b95f40450d9c97b2caf236ac00a51f64f1549cbbfc7840f5d7a24d25502a060e0c67e4c790f709c8c39b1511087b5f941e80153ba3008e3425a64745b894567ed02f1ba0113ffb71da9ed1880bac1c09375b4f636e4b3af094fedf76202946a75e1a6d1d13e148a7dc539888b17978f266b00d24db6d0487563fa091479aad8a9f336ec63c42f8587145211e68208ee724a451be2cf36218ae6df78ea0280b22dac97d0", 0xb4}], 0x2, &(0x7f00000037c0)=[@cred={{0x1c, 0x1, 0x2, {r4, r6, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0x0, r14, 0xee00}}}, @rights={{0x2c, 0x1, 0x1, [r7, r11, r0, r12, r16, r7, r12]}}, @cred={{0x1c, 0x1, 0x2, {r17, r5, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r3, r6, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r4, r18}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee01}}}], 0xf0, 0x40114}}, {{&(0x7f00000038c0)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000003a00)=[{&(0x7f0000003940)="b20227afb07c8442e194a6210cb1950d3ee4fb2f5a77483bb8b7feb9c49f5cddf5a3eb24d566c19bc065556d50ac316a6c03fdd5a48742846fe10bc101b75dbb81b826df456cbadacd8d89afa86ce5301c33e47419fcd5c5b46404bd07f9b84fa269f545aef8eebbded73360edd0223cc29faf7f3be387f2439ae9263bc25c93ab40badbaea2d8bf36d063b7a046f7d97cc33ff82328c33d9132335ea5cc9d", 0x9f}], 0x1, &(0x7f0000003d40)=[@cred={{0x1c, 0x1, 0x2, {r4, r6, 0xee00}}}, @rights={{0x28, 0x1, 0x1, [r11, r8, r19, r7, r10, r8]}}], 0x48, 0x8000}}, {{&(0x7f0000003ec0)=@abs={0x0, 0x0, 0x4e20}, 0x6e, &(0x7f0000006280)=[{&(0x7f0000003f40)="22c75a15dcd98e5a6208aac534a7c2b812672a1b951ceb68ea0108172faaa64e45d6b75d7a46a3d78268c86e308c370d5a74ffb22b981d43bbfca27a963bde1b9938ac6722d67a0477a4ad6275a1266b3062dd9cb74bb5310d69734802ac5db6c952de6c56b6b75f732f68831d3cf41f58bb6569759c9001c592fe8e074f7f652417b2dc23e629c7dc52ae838aef51f0f8ede5be6d3fe9e7f7592b9dde58fe7775ced63f192a8826f8771452e7c1865addf8756043608d9660801776d46402dd780fd11c067050ebe775d43e04f4192d7d2c15691f003f1266651f1cafea560dd33acb9fb393d8c7cc2471212d64e3bb25a845711ba99b6d2a20678846cb7b172c9a86bf9a875044f149043298596616c197cd51fdaae95cc8448d6047af272b3adbb65eb40b0c7ce163a4cda024b87e048fa9d73d77a15eb7d5c84cb4ca3dbd60e7305ea4e7e4a38b2cba8db3344b831dd1996d7b6bea61689e56281fad367769ad44f7616f853ab6e2b789482fe682d9198e4014ffdea7dfa5a83dbd71d2c9c12df15a9b958226e2a468d416b5980be020ce855852b84532ba586d9e341ffc70da24774cbe07e5a0658e4b71e579e470b10959120483bbb8b5e0ec4349134906f6cca9fb007727a633a88fc2d99b4a6d177018974e12be1177bb2aa6ceb144b6a7ee14982b49f566440ca7b66fbfdfed39d732e8e1672817e27aff74289b0f935d654e9efe28c3b2a15818a447575f8a9a6e54e5a74e2cdd7e22a6d4f0c5a4cf9b52f571febce885054d7b8e93f4e3848c199a7a9ec79456666505e024d57eae952091cbbd46e7fa0feb280f8ba5aad254fc2a8caba3105aa6b28eaa66356d4a77e776460ac64c6913638bd0e16548df93c9d9134dd86f3bbefae634ef7d2b5890760987b8d6610fa814d8caf47e2d3dc0df1df3b8d93e23841eff10f4d032448d87f15bd25730eae01dcd0ec59f4f5f4d77dbc202f0144f14ad91f7ce9c43364f1d6010c963f7eb2039fd2938686c046c5f41468c023cbaa4c2afdf413fc0611ff493fec016ee10bffb4519f8bfa9329b655b548113414f92e1ebffb4e634adc478f15c71075fa02c3562c1fb82780709bea9af2b958844c9182c7da624cf6e5cb775c4359336a2b2bc279bed1d42d0e50572f6b111224ea421fc65e5a1c81e4f2ccd2d526594b9c7889269c4e0fb27f4063b3b662b7108629732f48ba3d244591397b1dcef945a5fd84ae7117cf60518c9e42030668e91f773ab3bf80ca02395f27ee520c9bad2e08c9a0bec6e81bb0267b48129178e37d8f5c9a207b2df156d1f8404d75c646ade11d1345b77ed03defb035c4633bfce40083c7a443d09fda345d84a473da6f7069ef6c0de50755cc58f325fef71439d935df66efc012999f6c6fad7c9b3e4e12379d965784d1fe130a54ce6b70649a9682d997939f42f835613b21a5b67b4a6dddd52d0789956e5b61d8bf5e02dc015801978acd544896fd4b2b9a927533bbcec59da0618f092e7df744ec4d6f3a4ae5edc63f54c32d3eabefa9d9e3fa677665a4bd86b24c48e01f68114af947b7d7634c7fcbb05d373588c43a53a632add12617b3b794beb2e6cb3adf93e0009cc137c05ad3badb54c92580d3ae3dd19eb94de72013f2fdb912d3d2f6c1dacb5202038e66dbfe06b1c17db933272c711129e2201055548b285331198ab2fa7e8b70b1c6dbc3769c9c265ea54410d23939dd332cbd81de08d1c9dc77d582622b77964f1edd2e53e09a905d2ce17eb365081ecb16c28ebed293fea7913bab3b007d768ff228ace35aa91ad6921001f3ee7f029c7353b5f1989547dfa31d5b50a70b7c8536ed5216cbbcfa603162e5459d719e14f7d43b0144148f6ce15f846ab640de07f9c381c45d952b1db73720edac50364a9a4f65bb4f6656f501a4e1fb181f60948464a8c525971527ae7de85b6f41baaab9689f83e7ddb7d1635f2d0cfec6ea72a8ef8e29e1b5a8a2328722ed136ac7fcfa3a09a58ef853197cb03ebf400678fb5dbd8e6e0f8fae0590a5cbf83406ee8a9cb50319fdc93101e16ce999d3819743554f590f81af6de47a849881f3c4316e36755cdb460cfda2ecf757273b1249547de06d02980b1b18f456279f0dc8a59f18cf8bf54c0c4ab6d874f4003c5875ae17b0f2f9ecc3ead09fe1791101572b67ad156e64523ae70a86c7e4653feb6313203b0809c2286f7c2d37c1df2b067be674a2fb372e118943536761dc69d74ac3e46ec3fb23edb3ebcc9edc2fa904530595801b16965ba825728592a15246f6c91421e2f50e47abc820234aa866627d4e69b16c325a93f277005ddbeb3770a9bc424425e3cc8ecaf2fd0df19230aa85b00059411d2300339fe71ec492176e40a5b3399edcd2cd9442b346943926c67c2da8e47fdaa6ef9f0626dbaf13a01b7a2b0a587550513342053c35132c37b4b7bef71ce417c2686a98c40e74a91ec4fbf1b90a2028200f227e869b2528b5e460307491d36acbaadeee0fc620f2e10fdc2405e4bd92a70d7a03f9ff916521563aa5ce6d057e123005045b52ebe7ac7208f233609c0b0088998c5ea7aa1daf0d1e8fc0a20900eb8bf4d751bf99fd56c1559ab392f6ebd1be7eb624d22ac6d488926196632d969f1285552767d9dc9d2f84453e426375cfb1224a457479711064e36a1cda237b8a46174ee58b83f1a3d2e7b43bbbc837ff54f433e9e5a3796cc58591907d457a085bebbc528178b7e31ea27f83921d9ea25e2deac7180c70f25db8daf528f4de3339c295e79531bc48975af08251b99561cc174d841f371e555ab6239399ec6d3e1904b564694201e32227509e7536c5cd6b0aad2ff6eb9803d6b1f63bc2bfbf108318af2ec55a753f3eefcdfe02ffd2ea69b06f93dbce5509cf9a68e908feab5be81f6b60d39974f6b8a98ff5177c872d80247f98f3d4dfbbdde04f77804bd207fe3b8683a5ced53ff8518c603db4a11bf694f93d894e97c5fc606e6a4f8832e82c75472b28e2571f5a26ba12896d0a5a79f325369088eb33b887316aaca6c5b0b6cbb90239e9b95c737803a524e3e5a3f9c24486a76abf082fc96c5726f2e64978a8579b4bc21d40b79979f054687967bc37b85b3ea111bc3614480f9ea8b3ea5a73907069b95eb7f916cd0cd81df315def1e2c9323da81ec5d5c4a1fe8414295443279b600a6e8eff16f867c98da8ec091c01b3560b82bc78d227b9ef8269d5f10af1bbe4890306548c752fca9f928cb4c5dbe5c30e813bab5c98a26ddda6c2897fe74b32ef6e7073ea94c15e695feeaded5ba0840fc139ae7790acd93f4eff8828e87e1d9cf394a7df9949a7a87be13a7cc221a8de84c9a028d61e43a8805bead494cf6e7c2fa7a210665317e23f4062206f46bc8e03bd4a0fe52718078f4a5f8e4e781288356d3caabaa1f6a7b74e5a331a50ac5788b67b0348b5ae25cf6198d123f410a4b1705dca7cff5e02c4a4b82edd29c602a695e3192fed1deaf79bed016f20250b4e31408b529a631478751c94c914f9ead9c39b5333a72053a8493d2dd77e5d23e18b067a01807a2683692d882d1d4ed967240490a775bd3385f4c7b57e933c1dee916ef855c9d77e34b7f3ebd219c8658b60bf2f719e1498d2cc4f4aad61278f571886abeeea5bf34f72d9c70406724144c926978e2eadbb9768e35341b4e285e2e3510f0615e3aefcaa0a3548df69a5a0da569e0b5c0956b8ff09d9ab877c1e0db3b96e16e91e1692fa1c5a78708dc375ff320b097f41c34e175fdd9142d03184cc56d9f50ca3af250f6edf27424ce234134d7c223aedd8bfa589d0c6b02f81977117f773866a5f31b5f6920ead42b86aadd0cc24cc8790fae11a0f59f61c1d21f0ccad3bb94a2845718f1247ffb9004f06c871e5ff4138d52f30c5a702fea3475d3b9f08688b15f7921b71a48ed90c2c014462c13873606ffc11752378df5c2a2ecd7e90fbde369b323c8000ad444dca14b5b6f7f4628b688f3b945c4bd3272d74877c3c49ac17bea1be3213dfaedfae1cee5b5ddd7e602c93b5f1626400e0d043e02f0f47bc46a79084e516a7c075326bb588b8f5a85d84edb4fbf13d687cabf556cbcdc5fb9be37db4d5cf7e95fd49b910420084b6c4e91c6a03a5ab5ffc33e9e9c6dc67d8a1a1ebf4bb7e1542a16b3333f776de3be2aebd28575d34df4086ebe2d6c5c9e359b9fa224bf26f8ea45b37a9fe15b72eeaab2c8a9b517d1962601cce471f3caeb6d16d01894616841fda6bf0b1d453b0d972a4ffdd8a5a87ee7d7bb17313eca6f8872f624b16435a8769b2608444a68e4a7790633f65c5239a8c89cbc1621f4b4815282aeeaee98b75e21d37ed6bbee6d5c288439ba035b8828f0aa008415b18eb00852404a95626aceab8d7cb236cad5460ebdf95a06103806ab486efebc7f3f1d94a2be30ea14d7814d4e48794cc7b324b7f55f38842b5af06a1541b428063ed0578d99715b992e3dd6b24d869e26ece6822c85246f6fded15d054ff2c6c780637c2e37e0a27fca4c56a8167f58a23701a854f4660609b3584cc323a0d6a8d389ca8e27126b6ea701aed79fe32df10ac58c3d3cdd37e7d9c680495c0e723ddb1d4c787030bef309f519227c7c6610ffcce313a92cc264808dc2145f5ce77f2e5d8d174171fd74ccbf1baf7d3be01b566ae0100290b45d4e1f871b9e40abf05cff07e3271eca88d9e485d55e09de9a036293f2784e4bb06ad66beba63a6a5d73c5751aa9b9b5945f789478c86078293ee1a414acb0797f97e0a2657d023ae7294e67c5d8414897d2f91af2758186f3863fdeacf5f6bedfd7796d25375837303167477ff5690cc908fc3cc584c0fb51bdc7c411d35d2e8d0c05864e7612205fa1d602654dc31f5d6ac09f43ba2ccaadca4a9fdb445e20b08ccf8a58990afa1d0cb109f7a65fe20d5dd914a54fb2fb5b8022641a4e3ae8f33967571a947b42e6b041a0d99934c796bc7458846a6459df8c6de4ecb3cae6beb64793b68f43f9f12a1e8c2eca8b81ec7b33c0028ccc40a5a91a30128d15bff27d3a7e491cd9d916b6f50c05cf29899fa2019a27fdfab39a9b0fc8b1983ef061aedbb06a4652d4c9081933ab87d9b68cae11967608eb7678f49d1472f24f61337b3733aca9062ae8b1b1d7c8c611e522f53306aaaac86b1f8fa3559bbc3594813f9b95d70897753a74d8c28d97e48e753ef84cecfbf246a58c5aec8039f8c594f2f9a63d499a718a02c519045b68c0985cb7920200acdcf6a5b094fbc054d3576315090e4177ac0acb5fec5d0c2e73f832da1dd2d2e6d2ae41c3a669302a8b3dfec8ee555a78b9392cd8affe8ba9d416310f19a4a032a22ecc326d26fe09a4a5671a7aa777312cda43089b1ccc554ab10a0cf0a936e7e531f2d3728195d0953444c7f7e4a26275132f74a85725f98366f32b121d2c49cbe3db423249e207116265224bf63570378eb7e78838e15731f957c8fd2de6ad62d396eb2dddd7493e371a7025f6d2f06ba4d682b556e4f0c32df396e6b0e78a22805291c1da587a6e34427d4fa4ea56b8c81f96867cb987bd6e65c570f4e0e731799d07cb3adbe295f94ca9e4f326664fb96a0274953aee8c240d86716f9dc2c3b06ba6dac5b5d5e75879fe1d0f053a9406440d766c3d08eca6507e077f90c76a3249a90ce253cb0e4a719fff542a0a16041588429bd23cf4933a3a8d731656c97700f0e9825bb937e01cb75d8fb7c1914305ce98bbe91d660e66075d34d441def624e2e4df7aed06d680956589b43105f2e2fccb18e0", 0x1000}, {&(0x7f0000004f40)="0e596c46979ebef535948ef7920decbcce4a46a6996f2f7b257c1ccc7da99ecdd7bea6b65b8f4fc3fc9cd33b03c815e841ea7f8c70e09e8df12319f7685935b96bdb73b2c2129615be0695426ca3e2416872e00289c1a0d8765f969103e2191036dca44c9e9df053d7e8471fbaa9e96c491630b746b4b13a93d0b25314926f430be320eddf9cde44577f7567e49c22075b40321b76c7aa7461436ea0d79af7dc8b7fa8c8d7705a23a999a932a1d28a2b41e65510160602de0b0b37a3c6e2404d42be8cfb4671b0a35e01097c9498d1e68ce178d60d88e0cb3d19f544ea1c529398aeba5d3c93b837f399ca580fbd74c9246f15", 0xf3}, {&(0x7f0000005040)="4ca956103ff00eede538b09d68387086df46b58965fb436cb5dcbbfdafc6be51d7eacee59219a067241dcbe7efe7d83b9efcae92a21ad5214f06e851651de7876992cb97b8a0914d7920f53194b6745611906063a5aec8fe8d7a8062bf16143826c589675eba3bef0ad66d7a8f0d4b55a57c1ba8a8799814447f3281d3dabdd62c", 0x81}, {&(0x7f0000005100)="202976eca3403546431664a9874725c0ae8a9c52c98ab431d9f3dd4b41b4c70cd65958a8ca169e42f11aa59848fc06221dc71a45cf395adb24931e903879b92f1663ba5f40f0f1b1da35f7487f9fe6d74612855bdb8cf727b2b362962a1aa5d3", 0x60}, {&(0x7f0000005180)="21bcde6cfe501598d1384c7217dfaa19d274cb92b6bebf98d09afc456a060761fb63c5482cefa66616b3955362f393bc1d3756d8b0f29342868af5cff825373837fff0174870c51c88c9d7dbb767edd3d84840d52877a7a96a7fa3cb464d8c373b30b60d45863a5fc67e8d74af77328d9724fab55abb49e77948f48d2c7c8526a6f19cd3eda957b4d32e5c59175ef1c828abd821b421287c7bd0caf83874262236e21c3dd9382e690d9a28440e9b24f28cd5e42bb4d3b4f60236796b3fb817e4d20c346527e39c06746000bb5c4f42fd3aea32887f447bdb4571be5b9cb421a934560515fe28c2300fe31061014ed00ca3cb8a06216821eaa8821370992a5aa4f59fe7701714915041db819a1b3a7ce08d29181f9f19c7dd43459a8fe1b8b64511afff4b3a7e9630f2f34396ee9ca8daeb27a4ac9d3c05ac69f5bea18a05ba76fc6a7a1565ab3804f22ae9f7d975b9d0091e8cd18f1f7e4dfd5406f1d1fc5209256be1ef8f86b33a61a2733d969f5d49ac05542623e432bbac70d49eb4a0862072d0d4981d932bd0c7c0fd7df228c0f1dc83ac9982b468a98bbdfc41eb31ff59ec4577c0109bf9641a787e5c9a524a48723cca6ea3f71fb6d39f66c0fdaf2c726cacca15c06b480a089396dd10ee8015714c8d97253df8d42c39e9671a178b31081057ac188c90d080c18d6704154c35eeb38886e9d9e5995db71e70822ac5bdafe401d14a37cdca71f0a8d31e24c6ad6d7f5318490af2dc91f5a28e70759224aaa4c85a10babe5f253f4cbdada8d3fc144a09da6fcd9d26696cac73ac5157071bdb185686d3344d2c08dc96e79d8b16528961a2c7180ecf2f0dd0bc8e1de0da8ce88b8a21a4b8508649c61ed07b16247ef1fd97ae4f1f83d45ee7dd449e0059ac724e8115c94b496f41c43691485da358c9a8858ab0d5b25abd90f6a22a60f8f1cc9af40a48d0f1726399b3a8d904d7a12d1ba037b74564e3be44c4aa278a8678284a10f4de57e634601416372f3cc496666225d00953c252fd20c259961526060e1f7a417cf1cb02889300c60bdfb3dde0310bce41090ba04bb9c5dc2b3883180bf161be1742f23385fc2d0ca61c5192940fb5a5ff2ecf00ea5be78b360899b9a2e57afbe510673c91b1ffddb372cd0fdc2d479f8b9504d1e49be871ce5207521a60080332feac097acdaec17d5efbd3a5e23abea8bdddd03f7975c0424180b539f979b759684faab0a7ebf70afe6e04a5072aa4e881fb9520d9a7d687c127d37d3c38164480117618f4d57aec78ec80511eebbd1a75b53e2ee34e3b4bbe590ed2de0e31c3b5c8dd44b4696b5cfa6a167e9d2a148895ce4679546aaffae32a43821dab26eaf2a70656fcda640870378ccd4e6df52799035708ffc9b7bd50af946bfe7fb8fcd91a82507b13ea1a88f1a9b3a7a22eaa257d89fec97f32c664ccf955a042a7c7a524841ea854182673abc0e2799459dece15929945f9dc391889fe5da9cf20ed09376396c76cd50b4b7ae5463a0ab8f4b93da026f9cb1668e57ade668945d3305c16c65e4b9ffebe0341bbdba149e1e8cb5b6428eb4fd8873b0edfd4772c728b722fcd7813979a87efaf0008514283a935f9454d3f18608a2fa812cc0018e56248a73ae1dd8650aa84d671dc58ab06b36ae97d238ba126d1f393c899a7a4bef268a2bd1ac4cbebc66110e37074a6263f92e4d6ad0143874c378752a108bfceccd81c170d9825e6c51f611b401acbfa456dc705ed33407865611da9da4e6337417009441b3f91121461a8783df089308eae34d813471e3d4b1279be44a8b1eaff9c98e022b27258369d06fa14f6d8faed2c88e2c8e3527ba3acecd40af36e72b0c50276a7b07d2919fdafef60827ad973b05c3584592102ea32528805c17c86aabfbcc53427c0d7f51cc35b10725786dd3cfc61802427ca37411c4cac3d5f6124dafee16056f7f6f5d8cda2ee8a1c2ba15b721a295f9ecea220a952e5834d289410728bf64c1ccbd3573e853f139fa47a67198bd6ca4e54a3b1d6f0d2e45bde056074644bd14d1e84ccc738aad3664c7ccf38e13690840d2d47174adce92cff8e8e3c957e99f84e85ba3b2c5accaca850c76e3cf1b56fada7edf2df5879300f7b57e3e0fac3ad2cf051e26ec7c50c7eadf6af37286a7bdcd16d707a574b214ce5578c788d7cb37a37b528f75c6ebc3a80cee211348c616d3b60e08817bf1cc2ed9432dedfa2475aecda38b92278e8950cb31c0b7670673b79c8921aeee03a76b28e33adb097443b8adcdc0630a7f027ac2748ba8fb359863cc21518fc53191f2c01c01b09d43c4e3ee301fec641963e2f3a6d98275932310f7ed72a4bbeab0a35d8bf09e19535aa191d05ef8def3c42e1d2b19885c26285b14e36a50f85e7cf00912d78d776c90c58e46db1911bc644a960c59c4f9baad45a92ab9bb0d745f1bfab9db0f4ce2a0bfcdd47209b1f6dd53488d389d497fa837e2ecc072c72b10ff1f8025019770619317e5fcf38047fe47965d5076b9ce97e20831a7b156b84ffe64fb3c9fcf356423ee150e0effd7160deca2d397ad004ffb2be08c9001c3d89139ce5efaa5a72278c9d7c21a3e2ba622fb59920121f1bacbc0983120886af290516f8374b07837b9c2141cf0a7711c925e66746f4ced48f4d039b20aee192f0c345076b6c020b8f2342bc2a0ed44d064cb5e388fe5c50dffa2ab70cf134b4460432b01448a4824b2fbe6e37adc5971fbb910e4363039720dd6ca50ab10a76e6019e98e6c0459958a274a94617660cc40bfcf3ad9cff41014e743119eef17dea1fa84225f34792a0ad27349a26cd03e64d31c5c60f939ccd2ba52f2da4e94702e6a714cad6f163cddbbdb30c1bcc172c760a84d81b68cba9ca086f31b494e28544eec921d9b9bb6e4be8bcc44f7bb800ce1e30ea5a8001838a562b2dca1bfd00ee1b5facd2731f8ab9edea6295adc48d5082d8d6a519071a157f7cd49bd3f236ba139594e3bf4c0bfe9e784ad815045c1fb44d20a2c6e454e90c9c981a2bc3f720b15ff4c2a7490c4576fbf88e6b1c12bea6573cf7ce6f47f004cbd3ea2b963b9a7e94c32b46b73dba33e4da84dac9c1f9203d2685b0dffb14c1d95c17e83c4aadee3ebedba83f22fdb972d5b2a231230d4a59279919ccd51ab88da699318638c5cc18fd82d84d7fe45073f75b8530ad248c416a7f2678a11f84528059d1d839bf0c3ba75c453eb6ea3195bd1ce02b29d49840998c3877141b1162c292029c17bda3825b4ee475248e38ba014d7ad88be60a8b18bc838d1501f7846d9c956d77ac6bce25b9858cfe60d276c74c38f13289c2a53343ccf69db886f517d547bcb58653d2f24eab323898744fb4d91bb76c9b901b0130cc3e6d8f181a8eb65c14ebabd37e4914f80922363aed8b968c8ca2c14eebd9c5c7a7fbcb4d40fb14cbf42faea496019669f6fce56d266eb6a58ec6f0349fa6e246f374463d46ec603303cfd7f158892d082c7c1d5590e3d329b162301534338db0378d2698da21b74493f46815b6c529a5a9f018d7e29b7ae737d2eda954175ad2b848a99b5d652c0fcb34dfd5a08a0f0272a7608dbcff46f0f093cbcd509eaaf3651c22903c47c793c72dc76c7f00de26911988c55779299ec6a714b1511f5b90e5c398c051a8f2cb65f834282f7e5f4e243b3c3692aa3aa07cbaa97232deb95553b8a490ff1ab1943ad8d99d8c7b25a79ad15a6b9310e3ec7333c00a2bee5b1a54fd4108a60449bba0fc5a64b1f4baeb3b8a554cdd6a0a6f5de209dcbdd9ca157b93796267b7a7938c7bbd20fbbe4ce45161e8b6b0d6a9d85945b2877b3b1a9d830b7b6142c141a9db49da86c982662f52f212bf881e01659022832512fa4f115c1589b9eeab0f9720d1139009f6dbd6bd7d032190698dd456529bbf51ad9499ed9af4e3abcdc71122d251816c805d1110727c8e75e13ff5d0c5551dfc263aeffe866839ead4e7823a9e810a8c2c46ea1562d69566eeeaf8b6d4782094181aaafbacd1156bf1abafd78c732caed3f7a90fca140bb8ad2d237889847592afa4d7f043c08fb6eb7b13d93bc7e4198e24f80c3ee9394d1cfbdcbd0e71193619b49cbbb6e2bc2e25127f0431d743667ad3d03ee335fb09f37f22478606e57d6d7748003c81d9b97aff77e60a4ed802765e5613dc65ef29b4ef99c5153d63543e8a438209e8ddcd851d658ee098857ac9e64cd61c80cb161bdcf91576d9f588fa4fa59ce10a2d4d3e7b16513bf5026e26f25834924f764584379c39e120fe6acfbe82d288ca431da535810405f631c1e8178b95d2071a52617256d81307f0284fc6518556b0c37ad61542c4a8b7aea6292e666b1bdbf91dbefd98e56bb3b53d19dc7e773bfbb0b71e337ded1fea2ee6bdbec49a1483b4b1e6d8e282c535c84308a6b0f2b9657170e4bcf958575e25ffa2b1167f51d17c8efcdc6ca6a4091883c48289142fce258114afc20a857bf01a153f46b5447b3a3657c4d82719e22862c10ab894d6a387a4fcbee41a7bae3664dcbc9c843d721484f41a7b32907c81b7bbccb20eb76d640adb482a729fb3cadf79e52514dc757bedb17290994feac473d5e33b4a182426cf60d67cf670db8ae943e58a2ff22de89757add9f95b18b3db393c6eebaef5b72ded18c417f7021bbcb7b30fe3ba3e7b19c67e8ec7b68d726cdb869b96113449db930b4e9fb39ad1728e7e1843981638e17c3ca66548cdc9a9f63b8b7cf4d4c9a42d551d6af8fcbdac18f09afc7a5d8c1944f289cf88df3cdb7b57ab5e8ff72e8179f8fbb5d346a148a11ab6236e47db53b0da8dbeae745a4b3a91af9ce88bbb50767f170a27ae9544caae98ecebc535146c0d669782a50f0934c03af2875ae72be1079b649b4591fb5a6b1a83a8e25b34d21f376e9ef141e508b4addf44cc29ff0007f3b7fa75950fdbb4d8f0f8f1ba8005b70639e4b9e9e39316a1dbf799a3a1003177c27093513f46a5221e7f132e1a4e821e02fc5558f8618ce6c7b1cce76296cf94e347341a2a3a904f4b509e1e6c1d86deda555a565a47fa3619cb980b01372c0d6f5f3b073b52285db716dc4b64b88d20c578a822cbc2331afa4ec563765d938de25b4b3cf09b4f701ecc6a194edbc98ffe8af7668c6a9ac222c6be7a314ec6f68fc6b9a024f306ec7e66aa4990f3cfd00ebdc46ff10ff3a00dd331c2b269f29f35755c1b847aa0fcca78e1a9794d0f6482bdddebaf214925d3e79b7c5347eafa8581b18ccdc7f95b4e753b0bcd04c48a609a64899aa20b5dd3e81a5340f1cb788eccc4f0e3961f34feb1c15b26db6cad98c012f95a9265c808bf1e73cf30c8c28bb4d4d3ff22dacbe55732d0fb65d95890a498f6134aff3af9c106b72ad9329c99b133ba35b5fd10243ac550db7d1202255fc9ebd6f6ec9e07b8b5a8f9b548f97d8d203ad0406d0f128f4786bd74c7c65c5a23afdd1f8d7275dc3dacab3b348391a0039265b14d630c28375dbb24368ce9fa0e26752c0a3a1f13839806679fe18601a6b41e407a7e7d6042d741b0dd8dac1c38257baa8e9dfdb2b086dd9b426e83a88f37fe6f273738230c133fef0f57e8033f0511c0051cf19f4285d6cf6b7a5d7f7b8a04a0bb45a03174b5c145d1b98264bd8d07247358ea46a3bc85c5ff7a5ea73cac2bb4582ecc1bc46e031a7e89260327102a7446bc1c8a6d11d2a6f377d5465656c3092a51b3cdf432dfdc965f4531c4de08bf5dcc533f3e5db203003ef4d63694507e6482e3db1d3a894a10a13b8bf218c2ea63dd55f163c450", 0x1000}, {&(0x7f0000006180)="fd9af5d9c78acfdd97f87336aeb1bb81dd377703cee6e14806ac33e886eeeb8a36f0ab62f58242abf1880b57fa2d593780e7f9fcf3f2981414d80f9af48bb91ad41d143861057c558bbd32f0b68d467a37e15700fc18769349166c430e6ebd4cede8247478a30746b4dafe153e6abc07e72ccb5af931892286fe878271d27df1b3333f2fa1603124e85cb89160a05e6a0555e507f49f62dd0c5eaca604873c61803c04e1800e11f1be959034b054960d3822162dbc1e546601fedd77d7f6bab4ff831eac8bcb711034", 0xc9}], 0x6, &(0x7f0000006300)=[@cred={{0x1c, 0x1, 0x2, {r4, r14}}}], 0x20, 0x4040040}}, {{&(0x7f0000006340)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000075c0)=[{&(0x7f00000063c0)="506b2d53d79cb56b426676734636b20a7626a8b2c7a0387c2e4e60477b2ecfe6beadf5bd3b6df83f35312dc9f3e8b3a12fb2f102bc4ed929ea17068ede7a9d08eb2f4ea4830e41652ba5fe87c407571b284069278923909c29b7d6341aa61097019405167c23f187a95172ee4da7ebbe300c4695964abd5c42a7a75e66ebc15ab55cac9b50376cf5bff90082e56bab0a03476f", 0x93}, {&(0x7f0000006480)="4e36ed56e215e97db961808c0eab0ea46e1b6dcff2c319b6c759149ce3568d890d111c66e64ea2b73e89306d90e862af77820ab6cb5c68f5af3d37a632", 0x3d}, {&(0x7f00000064c0)}, {&(0x7f0000006500)="8d4c835a809e83c5b6185905f3e54bd0d061f1eb0814058a84908aa8532f6b3278aadee9a29510a64bf59e3aecfb281f2f93346650df665759957f07a490e84b1ffdd5b977fa91133584740507de962e324e9e3b516564373d1177a0168a33c88041ce047edbcb631915cf58adb19e7acc88cf8b787076d075fe2b3c7c817714dd608065fb7e5be90f21b22e3556792317a2616b0a52738d84d6b107052e4982d7aba96e34a1e8b223a45ba96f595750ce60c2328300f9bbbbd802270e695dab4d3f7ab9c6c3ab061cbd0367f6a991455c132e02d1a22dda2394555fe9a28459d88cf9aa45812253462606dc8c119083efe3d801d5f76d2b909fb5d1bf6809e3ab143561bddab732d609105e2ddb540992c9ad0396847f04d3c70a189a831afc25666f290d1f42c1a9d9f68ca631a74329325fdd6cc0429b72bc05d33720e288013a09691983e5dd638893b856eed86314f619b6af358b36145030b69452e9f8a5bfb07694b1359b92194ae39e46e85f1d42e03c04fab17a4538775d6ffa3b3f375c92599a0cc0e2fec8d6c8f2e8c4de8a51a7b61806f48191c1e85266b284c7845e1e903af011218241111e0044637aafa09e9221d5d09aa42560a98398436f575c3f3cd08cec2809605c1068749502aad4048210e1ebeb70444726f793d3491c902dd51369a3c2df3a107851704bd5ccc6d9a19604b22e9adcbc5099e854382b4d4ac4c22fadcad72cf0ec3eb42802e42df9a00cc7defea1f3e94dac64c8abeba8b22abcf8040c137fd6ae87b5f1e9eee88025f16734dea5ff678099b84d0d2977d5161cf5f0b02c5a1fb54aba2a98535da3fe1957ba8b10b557c2675bf965af9751995433fb414f6e56239cdf23ba5a97aab88b9a3cba2b1dc29d130a786109fc2c82d214a1fd6fe372635997196bd36cc3e339006319787dd07a59da33e4db600a1eb0a9192a0606c7ef31fde1b616229a70f59f274e32fae228a9330de3bc7ea51d6dec39eaaf737e0e5b221210f826e7b8ac475d6127359cd892e08873622a3aff006bcbfa0c23727c7c71b7eaec7dfb372887f62898914116a69af3228076340e437eb6d5d2f5852baaf1382299ed80cdc4bceb68a2da53fe569b6075425ee6c064b7cefbbca45ae0f73ea925be1aaa83cdead228a068dfbf500b0fca75e893c6a714bb7b04abdcf3912a3b3df10278aa0e8a20814fe12bde1e549dd5ad88be60d9d08991952113859ebc658c2f5c9be71ed437cf085986f1a7b6db4921c9c1b8d8cd5ce69322a968ceba4da29469da994315b013ba75b6aa833b201ba5dcd5dc671edec6a10b362669313e47dbf417ba543ce02d7b36bc5b421af58776c6982f222a3aba677ddd806e92932b269cbbaa181f677f2226d0ed0897614a6d8a8082c46d295b2afdfd8ec057be13c40ae15205f5ed5cbcf576d92f01393b490b0d362839a8786aaab0425b0a15a3c31f0855e413185bfb339e09c4ed60907ed82a505ab1e5f5c8feeabc72eb9d0a4a46e0f6a24fa23f3f03ca5274ca87c5b5ecf0beec70acb262b1e030309ce0dc904f8a29ca59425557c1fc75dce563668b3bc63bdabb54f8ad3fcd699246568102a729ca28b8cf2a847e6fbfe967675034e18758e023e14b4ede080da798a0ca31a1efb2200cab02e6fba243211cc77719697ff68f33c5dc6dd92b0165aed4737849dbee3321aaef7cf9ce9d826297eddf2c8e6ced507d309cd25a593b66d54f1e217f24ca5ee466d7291af18ae7f39d27531f8ebfba1cd763958a2a85220234b6306eb9832786a1d8fa785769f33b9a74cf2991fb1e20f36384e5a7d610b84f5ee47c02d55087db734cd3597668eed121693d76421efec8ce7941fa2bd7fd58b76690d797b7120c2a65d67d9cfb5f00358446e1626d4aaf62b18047c42cb57442b1241846e3994543a4c786ca23bf6152c55dcda8812e154addeca6671408fc540b098af7c524f7d2003e2404f3bc99e0e8688d82f7b38ccfdf5f66fe928c266adf655442b24c81b32d9e5819b66215b3c46896c6db87b0aec2ee346d5d59b97992138e81b0b514bd99b84a55028382fc37a9b5bfbdb23f108f7e6c59e7efc6497a14ac41251f868a22c3011017aff619dfd958f1264578888092e9abd789880602ec62a36fcdd74c193f0d66a53da8118c0dd5d862ee28ad0e21e188a73b168e8b897dcddc4c9c2a20e028c2cc972e16f7139e2d0f0aef5851cc41ec0994156edd145595a99d20ef9f674a918d216be337e85444c1123d0a28f497d45b0e230a17d64bd750ca07a5d98baf48fd67ff7f50e74c51ee5c03f0f3e72291e0c682805b1b42e1c0dc2948e8e6b194c7a587d57e5c643bddbaecf5bdbdb10fa2fd06e0c1df0add2536f3909932d331865841125a9d52c5ef64ae32945dcff4fa93f3653dee33ef0930df71aa37ca1ad1c4e5f53fcded49d5a862548b034b90ac84862a2501d26d8fe737be7f4f67d3c03823a8d0d3f4cf22706e6592306b99b0e1e1980892fa649d282b732bc3b405734ea2f185f87b96694c3bb0781ba5bc3a18b5f1ec7b41f783516b217baca44420941a421cf876cd4b0ce86f84791c8f971cb64f961bddfd1276e21973425729e0a6bc688e6cd30fb28100b579c9ed6c94c1d89b51f16c7ccb62eb38827f5d8a10f582f333a8d26af8442d8c8b8061c0004dc6e3fb53bc9dd9cacf4fc34d5d8b7fb1e8ecbdd15e60df89aba38ae3128892205ced6555f41b4eeae552cc3b59ac6fa3338a26a3427728d599a510e0f6ffe7fa9bf33773e4355e34a853fa1b28f3b35be5b8c30b96b41e1e7abc5e519edf22cd18f0ffcac5b36dcc7b6393af083e4c45656e92b384c9211bb6171ee64ef960e7a2020d2515a71ca3249b03ea4a22fbc695a6019e58d5ec4ed6ed625f31115f2cc7350514792b4c7e09d469c270a83fbd15ec9ba23bfeaac362fb9fe8058c33e7de9557c9399cedd2960b338589333e8e359529ee0d04428afe63ab5cee2eb268acbe54f11043dd89ce1af5872f92aa3cd265a827cace6bf30ff95a640805c526830e7cb176afb7946f4e6cb4432d582e63bdf79a80cca7c923211c0539661f9420c600136708fe00784a0bd4749faafd0393d6ba9ab52e8cf40d46630fb4ba92ad4a61727a9242307b7230c029f367e67aebda84c5d29161249b3a1cac08ba36db41d405ed5791486763a20fe56684a9a2974d6305a26d5ac2382044bfaa3c5325264e9c90a90014ab5049f118d452a1f9fda05fea050ccbd6b8f9d3df44b5ef2bd394348e6206cbbf80252d7902d644220702f35d4fd3afd395d8b4531c2cc5e73aee6e0fcae6e182e24a07bcc6c4578faf684b995b0d5b37283c9e1f3a1520273e272353aa2c6dfe29276062617d555932b7d087798f183a46359a436aab5f2ebfe0ea1e4d1fd6158ba6fea8299ae069a006f37ec2bb9f03f3c78843db490e7efad8d49792428ae29663732e1a0ddcfdae32d4231989ffb6f114f8410860829855844237651819c4778da4ee20a93d85e1f333dc72d770bc06f3918129412dca763e97219f9d61b1e8e898a82cbaa22398123266cde13cebf146b5977787e09708ed427f548d52ac8f737af02e4d70dd4cc99014b97f99b9bc313ff100cfed94317184944839b41ed0d322c6939f9ba8eede2be8ac8081871ed2c40d9cd0ee3521d1c7a6cb1c4e126645e0f3a06292347aed733731e59df8a0c999011cad66599cd47ef2b5ff31f63b1386c01b00aac628dd80f940f9025b61895972c28e458a5e4630fbd4f799eda409a6428228c04cc9621ce8c81182d1b473ab8adc107875c16a3f906b9452ab367951bb826a7fa4a7e9e6d9da1d41ed8ef352a1e182b99a519f421700a3571aab78448aeb4533e8417c4e11a25927131edb96a9b4472d4b02926ae2940746be29ba36bed958a5eac111bf63c968c182cded22fec8c0db6923ab7a3b65cd5987f6c6bb5442748b746ec387f7aee08160f3f3832fabc1fbba3ebe432357e9fa0be97d5a3359e74ad86fb5365725aa5e5059ec802b80ca7adc7cf4cc837af4e8e42a2eb72a914573868385d76a1426de8733ab058b2ff388b882177e1d384691ed4292e26e53bea764fafbbf83fe80d5eaf00ef3ca750a2c7d5a5c6be2ff0cb983808e025b4807b04c8d51a480cf83ca2fc5585cc36f480ffb3b3772e6ceaa93641cbc30724e339f28de0c039b9dc6cb7be7d5988bd1a291bdcbcb3c9105481c6b15ee56b31f37dcec512127bca719de9e4f46bcae364b73f3c92b9b9c31786eb3b91d84bf4394b1cd51f5112c9e330a3a86b1fe1208e7d69fc396a952a6534a58e6401b3c0ba4bd899cf7ce94c464834a9bee07c4b5c40de0f9b1faec0fc8a5f9ee97e1a3d8e847dcae729bb6fac7a501ff7bfe847486b1c01706d6d6f5875b4738f194ace523ed13efaf5a9964b7e0da2bdfc62fb172dd7c73c5042bf767d36dfaae9b33d89fd0d482122c379cd1ff4e1f8284e23d4053d8d57f9007bdb29fa5be13d4876a11b21a5b864518638094d756449b9e2c373dd11d55f8709c1e963074a64cfc25017c8ae8ae6fa3cc6ba3342eb5bd4c20b58d3bfc5cffb89acfbd068f6e95b38c7b68ba9a4cc47d8a08a37df56e60184164c13a33821dfbb3400782a7f6893ed6f6973d56e51463aeb349232665a760cfbc0a723791857cc0f0df41b3d76d817c878b7d88a96f836fe841d94b38e2fcbfc53e0f29df0fcf371b7e2fc3e079f309e9ad4f5d48c29d4537dcc4d5844195ea0633496e18552fbd4185cebc363ad7dc197172d1d66a9f3eb60928345bf05c6dc18c11bf69eac067bd3fd7a0e63408a43de74b3ffabdb20c79c7c6607edd47f5d45a0f92c636832196dde9cb7ab5ddc98718c28e0cd7b6b21c03f8923f3e5552e4ed2f0f2809c00d4c65bf050d6e2de49fbdc056f4bf0c967aa5cadb79317cefadbaffa267cba8e95e59e5c71023110ac81a8737bbe0fd396fb9ae1583390bbeb8f8c317062fc489438c230cf7caa91ff1b04f8cd702ec7f7d6fcddb03b9f987e6f1d375c8cbe20d421866df57d8f6946ee4e87e3b2c23438b6cc7164c633e230e455748c53a74cc8b5356c7a1eb01339bb19031519bb6e916b9ce767f8eaf3e7ddd767bf9e4f316b4ea95e15943842a9cd3853f7f44f363c87b0b80ba4842bfb2b935087a2811f3890eef695098fc012b4a29098efeae24eb08dbc6a8aae9cdb21ee0fc3ae7cd7bb272b65e4221b62f3db009812d7c314bc698dc64c9fe2a647a4830c964e2cc3d0d633cea27eb24783ff8dcde6ac16d77ee26909ecb3d20a0c21b30865b20b67369e8a07877ad0a1bc0813c12ce9f6879e9ac79674bf261515ea0f007c7c39db0d81478c1ea0aa31c70d3544a5d1f93451d1ae267a796f6e05cc3cea484479e0a15aee4732700dffa0471da81e56b984eb076461b9da3f8ac21aae0457df4c6f20192fb35acfff0b98cc9dfa40d42c47d6f0948c528ba2d65f5d6e5134e4d4ff7f09861c5ffbe65a6d6ba335b8f5c92a660d711fb89ba1340340e0b16c5fc04e442092ee3ee3bc5ca4b3042df77fe98f934625a284b4c58c1403a7408185d36caa31bcbbb3f9bb4d37a017c28bed82d045776473a9fd80cbc63861f21d75d3a72d19554129ea1db44d5ec1a063304218346369897da8458e50551b746c82a292bb6c5c777d3d3e15e870d5b2b430e2c303b2f491e56bfa2de76cf14034852c0ff57d278f31d28bce96a202ffcde7e76bab3d1a7b7003f5c880a9fe27c531d2ddfaadc20f0d6d57b969e1d19", 0x1000}, {&(0x7f0000007500)="dd19afdf8a5da5a3f08db50313424d8d55f6a8963d573faeaec0c6a9e29c257b2d35acb0a87e6fb7bea8b250e1e6d4d1cc08099315408fd2da1f164463bc153bcccc67bd73647ff8464c98a92ffb26726ad4bfc6b668d3804e81241d8d4742e940905009207a472c5fd40fa3790719ecca244d8b2afc144c80f373d06cc5a4cf464c8472914b49efda31b0f07d750cb2ee", 0x91}], 0x5, &(0x7f0000007640)=[@rights={{0x20, 0x1, 0x1, [r12, r10, r16, r8]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r6, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r14, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r15, r5, 0xffffffffffffffff}}}, @rights={{0x14, 0x1, 0x1, [r11]}}], 0x98, 0x40010}}], 0x6, 0x40000) (async) getresuid(&(0x7f0000000d00)=0x0, &(0x7f0000000d40), &(0x7f0000000d80)) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000dc0)=0x0) r22 = ioctl$UDMABUF_CREATE(0xffffffffffffffff, 0x40187542, &(0x7f0000001700)={0xffffffffffffffff, 0x1, 0xfffffffffffff000, 0x1000000000000}) (async) r23 = socket$inet6_tcp(0xa, 0x1, 0x0) getsockopt$inet6_tcp_int(r23, 0x6, 0x4, 0x0, &(0x7f0000000040)) (async) r24 = accept4$unix(r1, &(0x7f0000002bc0)=@abs, &(0x7f0000002c40)=0x6e, 0x800) sendmmsg$unix(r1, &(0x7f0000002cc0)=[{{0x0, 0x0, &(0x7f0000000200)=[{&(0x7f0000000000)="c02bef9485e586480b3ff511a67fbac1d8bfefda2ccee81a769ce84115cd28fc9f17f2503f8d8d74136b9695d329ec00723a67e013349b5b940981593759372c3e0d4bd25659725e1ee43a17ba3f2a4ff0ffc85750e4b52e80224dec32140a5f4296d833688d69ce70f36f67f87291d4abe3d5a8d3a3a33f6186d095c09570092106f2e1c7074d306fa7d6fde389b2450dfc9a8219d61f5fb49c96fa5918d5c2ec6c8176543fb7a6cf09f3fe15f27ed871", 0xb1}, {&(0x7f00000000c0)="4d0b8e6145ae613932d872d99c8812e5ffafac225b1d4a2fefaffe2de3e6a7188ef46e07152dfcd9f398", 0x2a}, {&(0x7f0000000100)="2960eaa18f73eaad074cdb3c5fbdf9731cb48319e712338d5af1cd80e1f5037ec3e5e7d9ddf57b7ed5d1a4f59fcc894a57e5cd8762050c00634f4bbc3fa93776727db6feb669f8e462a9e9e80eaa2744ec73edfd8482a45d3327d9083b33c9eb13b6486a07a3db378a25c2766d657a7295cb7354e13ac7c3ba630b3d632a97934d3984ea3f8aee4b4751469caee1c7d99c565afa4b9dfe99f80fa363e2e1f2edc8a530a60ab31de4a0cfd20bbaf82a20aff5184d5f1140d23b51f02116f71e1f808e06dd2825287db9cdfeea1ee0f9f607dcfe6df3cdeb7ddaaa3cc974df3549a702e8ddf5e4e1ed0939ee46b567a5971192cb6c8d6f", 0xf6}], 0x3, &(0x7f00000002c0)=[@rights={{0x18, 0x1, 0x1, [r0, r0]}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @rights={{0x18, 0x1, 0x1, [r1, r7]}}, @cred={{0x1c, 0x1, 0x2, {r4, r6, 0xffffffffffffffff}}}, @rights={{0x14, 0x1, 0x1, [r9]}}, @cred={{0x1c, 0x1, 0x2, {r4, r6, 0xffffffffffffffff}}}, @rights={{0x20, 0x1, 0x1, [r1, r10, r2, r0]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r3, r6}}}], 0x108, 0x1}}, {{0x0, 0x0, &(0x7f0000000680)=[{&(0x7f0000000400)="9539f1a6492ca3da43b1e82d1716f68e486ca284c1fa282a99f52a4e1aa83ed3195339f9402577db61ed91f4d417ad93258eaff5170ef042618a9ae2ca441a0662a77299ccdd58853134af3fc9242178bd5fe4d0", 0x54}, {&(0x7f0000000480)="9791f9dd608e613ee69421713d35b938b0157df9629c6923b45d9ad1eb11d82b22753fed8fc6ae77d71b409b8fd8823021711ca161a3caf77399eb399bb46d9578233a4743ca5e2ed9852f43ade97c0834b106eccfb15629fc2c239be9ce8c46ce109d40f3562650ca9d5a631095288b10bd97477df15d3d4a6ad7281bf3c262fcf9ffc04081a2d2baca5035928a3dfafe73791b5a72f0e04c7bdcb37357c3", 0x9f}, {&(0x7f0000000540)="c4193c9f834eae0789dc663b8d918333dd988d27b06a6aaab223de39a3a32f35804f3f73661ce237f5175dcafb6da0139f2f4d580318aaa798c87151098cee62ceb969947f30f4a1dfe2f9d6422530e4582c43310f", 0x55}, {&(0x7f00000005c0)="1af816bb93a2fb9cb8fad6269fe533dacb63d83abf161eb17bcc55630305dddc3f20887a619c8f78efe3f417f87f61f3608a9bd7cc82ed4ff0709f", 0x3b}, {&(0x7f0000000600)="ffd0e7106db2a09ac252d011ab6e19f006ca25c5012da00cda44d5ed16821c5f33ae31aef6a9490254b6cb8c6dff39b10c069cf0c69a35fd695a0fa18e3d93773d49a2b34e63cc9cfe2987c371ddff6d23ab873b9ce09df3794bba24214567bcd1a62872f0", 0x65}], 0x5, &(0x7f0000000700)=[@cred={{0x1c, 0x1, 0x2, {r4, r6, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5}}}], 0x40, 0x880}}, {{&(0x7f0000000740)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000b80)=[{&(0x7f00000007c0)="7d526a8f53d1b7f9ff9c6e0cc8a348233d647bf2391a2aaaebe00fc75ebe83f4307a87a3b3e5f264f65a2cc142435e6ee5ad84e512ed1c2d2ea07d4f24d362eaaf14dccae6fe", 0x46}, {&(0x7f00000008c0)="31153960adf13dfd7d09e3b80bc58ba4f0a255a1aac80ee378e9d33ae49032", 0x1f}, {&(0x7f0000000900)="81dc4a0a25ccf70210dbcf9092ae65aff51c2dcd379ed421357c77c002f4541f6ee4afe1f918d2af2d29d722ed0ea7d887029fc65873a9489e673db46a927de4318c", 0x42}, {&(0x7f0000000980)="0e1bcbe9663bd572060fe422f00a751617b492feee8073414200ead55c6d27b9a96a84791b2675f01fc692dd2290518be47a267f8ebeeebdd58d62684b1ba56febf6f39c0b46a24d33131252b5c2abcf2f08c1eade43bff3fbc044bf6b04433ebaf76ec40243b10375db50e2648b83cc79027db792bc68d8b175bd7603ec5007280d1315cb64505d0ea87e0bfae5e78ea366fcc4597809ff6524b8db191f4003fe4f442a22e9164b7fb7f9878a693adfa0cc367278383bb92d84d29c2ce4cff8f33ba42d054a45b16ee39f593c70ec06c184f2d8367bb5c46c79c98730dc44ece919a07798807c9ede3475a8e946", 0xee}, {&(0x7f0000000a80)="07b5a345215684e06f32735fce21b4a86419e71593a3b33a7a8e51e42d6c5741df2d7757802bdc3d2bd4e6509016505aeca2a8810adfa7ef73afe07fa5707c6f75f6f17855d882f6bbb75878588ebae5f41bb4bd21362b067741b1cc18b8a49973ba64790ab3c5f8ba4eaf3253b6251bea2d", 0x72}, {&(0x7f0000000b00)="00909a243334def0fca7d663613e10f53654fa693054ec216060b2063a155586f39bd7092ab190611aafb24726fc7ace2d8b47446a7d53ce1af08f4dc39db191a9790d92057f", 0x46}], 0x6, &(0x7f0000000e00)=[@rights={{0x1c, 0x1, 0x1, [r11, r12, r1]}}, @cred={{0x1c, 0x1, 0x2, {r3, r20}}}, @rights={{0x14, 0x1, 0x1, [r0]}}, @cred={{0x1c, 0x1, 0x2, {r21, r5, 0xee00}}}], 0x78, 0x81}}, {{&(0x7f0000000e80)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f00000014c0)=[{&(0x7f0000000f00)="ad45dad886c8d44443c9293a7176cc35d006a6a5e331431c1c609f5a96203ffe9d4af95138a2e17185a53a51517083679720b5c42ce0daacdd59f138cf", 0x3d}, {&(0x7f0000000f40)="a0067ac1d3522cf57a949d57b34f1dabb7020d8790aabd6355bce48aa69fc684298238af0bc37dcf520e081cc4e20d9b96c54345ef06708c86d653c87bf73f90ba241626ee62a5ebc3db7f461ff1b8ed1fd141d1bf3d9e1158344ef9d638db88f414bd401c2584fcac7f540b7451cd580ab1cec5f4bb9e9ce0eb0870b7cbd9dcd2118f04c3e8421f1f24555c42a28819f259667d15a8169525c643383b8dfaf0b847a5eedbb5a2cb2bbbe5a9c32e", 0xae}, {&(0x7f0000001000)="c4ba7c566926434a77b6406479351183181b084b55d23d08dc12d2e6beaa73f79239f58c43e90b6f4a406aec3e8e85da6a7a61095d3cf7e2151bee09c973f524b834263468e8e2011fca6e8630607f0e6d8417033dd61101d8c2218cbc4bad0e520cb5d729d82091529ec80843b0375b6a4933830b79ebee9d014e08f96ee175b9d0fc4dbc5526495fa10d9be93b735aec2d22dcebe054a26ec3188aa6cbcc08fe3fd6967b0588135d18cb777aa291ec924c202ad2354c4d64cff3da73d3ba5aa75b2745e2782de010681df9cfe4954c0b77b24c6af5f637afe076460da65d26c18d7a7272f3f01a94a71a3060c9661e1eeb9ed1e14fdc007f", 0xf9}, {&(0x7f0000001100)="d77c4d29078a94f7229db5a56da8069ad74c97841e40dc9b76ab3317468698d3f4dad49ad137bf2814f0152332ab558cdefc5b755bc0cc787eb575d41934ea376f9e944b9bced783a9acb9b1b7f2f82dc372dcc3fe661d42970b702593a3c4fe28370a2520c7619468c2bc671d65097471f9f8f2e92a927dac6d1ce9ba8c103ed579813e01102c65c728954c8ef7986f47f80b2b6f0a089e2e2ddacc83b1f27e1f50d2", 0xa3}, {&(0x7f00000011c0)="5a03b798bdad94a79341d7546bd282c01329c2717f1cb0b5cc1a94255a93f91addcae7358a0931ef7693a1f8fd3c925766eac2b1ca3ce92d2dff066b7b5a8f5066b8bdf67ccc8b775b6d859d60813de87cdd185eb7cb9adac60ce94de8375fefaadb70f3a7114f791a787f5f7062e647a7ba7b1f4934fa791c58682f564953e435983a128c1f20b73e64add0ace629d5fba86a340f39798fedf714f7efb4dbd8d6748f14da3c18517078326056dcf089", 0xb0}, {&(0x7f0000001280)="74d8749760b627fd936ba17d7abc8768da56e7dbd3083d343aba22412813b666a5ca9979f0", 0x25}, {&(0x7f00000012c0)}, {&(0x7f0000001300)="2bdfb7d567b7dd662a24794ecd5a54ebb8631a89a2be0aaa677c0162e23d1dc460953e7086102ffa96f23722df9e5f3f1727b497d9c999956fb5334f8c8562bdde37973d01e202d1ba79bd0fcbcc80c5dd2d9de843d224280a80f76f01826eb60c8ea4233fe9fa2addf7c9173a79c56788b12c51981199ea90a302013951a55ec1137ef7e2a07214e46552", 0x8b}, {&(0x7f00000013c0)="3a7b8380197c10bc400ce042548e4b39b918bd0e8d8933a9005a9d1772ec3dfd7b61c005319f8e304a3bff96fd4b0a0af76e73a4476e9b5a4ea3e194bf4dc07a160ea079aab8c46c145a8058f448f480895d678581cd2ada1a461c9efb8d7395178ae25a1633339cf29c6d74a7fc1708d48b5a005028a67cc9fd", 0x7a}, {&(0x7f0000001440)="29c47a594d4bf858358add8cf69b145ce9ec61a5efba8fe6c3365b6fc29d60a9c5d4bf001b2b7dd7d67bf635c29d5bf15f59e9c4800678bb1c5a9631b7fb782aa54984784ce870b543aa1ecbbc5ecab793894c0b82d509936510d91eb220a3ef3f680daeddc1152dd2511ab67edcd075e8670c0dd618e3b4e4ea56bf", 0x7c}], 0xa, &(0x7f0000001580)=[@cred={{0x1c, 0x1, 0x2, {r4, r6}}}], 0x20, 0x8000}}, {{0x0, 0x0, &(0x7f00000016c0)=[{&(0x7f00000015c0)="8749764c6f24f9503afee5e988ca167857b8580bb127beee950afcd5b35b0483e96a8b3456ef815fcc5226e87ea7a1a89ba9b276fd4a541aab6b7d63d49ad074a765e26fa50e3ffdec730e14bf39f663bdc9de6a6e17f25045bfbc1254abd854cdc4eefe6055a9bb5cb6b5438c32c69121b42f3526226de7496519a4480c8c8f49433d9679641c0c380fcf06235387a65840607e96e80440b1d2de1196a6982599140c06a81a2f22c1252faaa7464049893a0fa42b9f461cfb8885f5e1d81984eddd37c3e5f09335231b84862e2205d27c85d10e720085b3e380144c7013f35487c74c24fdcc7a8ead84ec2ee2c572725fb8620260", 0xf5}], 0x1, &(0x7f0000001740)=[@rights={{0x24, 0x1, 0x1, [r2, r1, r22, r0, r2]}}, @rights={{0x20, 0x1, 0x1, [r23, r1, r1, r1]}}, @rights={{0x1c, 0x1, 0x1, [0xffffffffffffffff, r2, r2]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee01}}}], 0x88, 0x8000}}, {{&(0x7f0000001800)=@abs={0x1, 0x0, 0x4e22}, 0x6e, &(0x7f0000002b40)=[{&(0x7f0000001880)="1e1277", 0x3}, {&(0x7f00000018c0)="f96ab9533299ae9f9c7f2f3b65a18cad4b8add705650168f7986e6f4e140c839ff758b7cc14a2a4673359d2d2b4ed84f6f31be8917ee2a336e787c4d9af9cea9537bd3a6d48ebdfba470da9eb3cd21eef3a36f63a368093af7adb9d18a1b45b59fea56f4bf4cbdc23ae6da47f44fef5ed5b0baf042eb7b941909393a411d08a15f613fd23a1fc26d491ff5d2394f423189ede01e2c7b", 0x96}, {&(0x7f0000001980)="d78b74a18fe39d34964d532bbc575d829b71e41578776e95fb1ce7c1620ab2a0c067375de311a0a350a1109bf00f2dbe2cc829aa452deccef004e7cdcdf4cf41df3cf197752d7e958879165548af01cf351ef271d1bde58df1d00ea40bfeb9163aaa6d89768c55cd9197d3cdab5ddcf1f9b7d82a23dbc89f8c452861b42014c285673e89c29f5003f5daa3bc5cdeef0446c2081c3b3f54aa744970447a232c42e2934db03f690dbe0a748f525d21caf0a141a8f956413e0d6e1eafaaed5d219c0ae0da3e03c6fb7423df032c994284d0f5e50c995b83a68672e3091588c8d08eaf66eec36764672f8c37f57a86455c635b25920b3bfc260adcbef3596be7e0d7e9622bd9a4f5b1da3042da11a35c84d2d72f8c875560a1b04d494ca43073e9066f3c891de5412d20840ce42c6c0034ab52fadc4fb4af6bafbc682521d9c37985dd1794f02436fb451fb539b529dcbec3315593249b6b177c15502ab0da2149d2a5dc5feb33899230033273daddab730cd2cf07d73e5a55212ad7fed88e3d5dc5b27abff8298ae85225b6a808d7c80b431e8bd70ad514766b5cd52c4bfcf5ab6c8b828d34fc4c1e6b69a986a7f43d307ab26b04bd6880d72eed4cff9ffdb831cd0fb4b99c13ec3ba9c7c15149598cfaa0cc45457bcab11285529336544329bc0ab6eac64c6295821c11eea45340ce3c052cdc923d9cb7d498ee154fe4a4b9b3d4dfa9c08cd9e44fa1dd83928117668c51b18815f8fa92e9554a93e6431bdd53328ca9b9a8cd0a008b0742089961f3aecfe0b631d90c4416dedb744ecfa8011a3ee58d3cfa58560eccef8fa244e3460c997bb051e33255c0e8253ee1b4c3269f954ea5f7220739a0aeca687be70738a3e54389f2d863a11f8e4ceaf4d6e48f4b3474092debe8dffea5c23a72eee7bbfb8a8edc2404251559d9560c8375fda60aeb7f10224d3753a9fee5732a75fafdc5e18b7ef45ef1e1d9103b774d91a5ad536b2420f47b49be4e68c6c119aafddaaa2b85e8dedc77dc8cc7eef022066a6d4e7f080acf071b1bb40760f54f9232322e03cdbf06028bb73ff76582321f776c111e98798dd2a6aa75de0d7d68ac066faa551687a316ce18f2bec08d2e7fdf390397318cc619bff5060c344dc080869fad49179842152b7490726f1057906131825d36cf7a983e7473a215d1eed51405647663789877f67e60eb8be3e90d2cc41ef6fba68af5d9e430f02bff50851048f977c4db2d4c82a9eaedf7170dac5af13e1221cda7d81990c2a39a301761d6193b9d2e5159e19a538ebb0de75c4bb16141d9a122cc483b215a798e8fc2ad52bb693b401234beb84fd0bf4f1024049644d06abdea6a28e88411dc43af522a3c19ba253a2cf733d69edf783fc499a7e75a9f5ca545e00867602eb17d97f8f9ce76b16d0adf14a41c918711dd8d404c88de9eb9865af3024ea922a9a0ec2a265da7e6c48a9c37f5527ed9632a0c1d6565383db909cf24b70309c1ba7f11d7870cd9dacfdc770891cad17b59a06ba103aa42aa6434dee4a5abae90a2611363975f9e37ae3ac90c174f130c61b764b5938fc8da441f2305715c46203e6f1e640174b816d07e109f8488c2397a8411436b62647a85207e83fb568335f12f00b06fa008c11ec853a7b41c2f6e9d18e9ffbf62f157202b6a4bb238a81f7e4ab02c4092befabc8af206ff9efc0a80b8704385d4f250b7e138353cf859fd9b28caf72b3c392914f4afc2cdc907125dc7388537b9884d2f2f6817df12d444a1b3b2b38163e3e42e39ca4a061ec13984be010cf6a0d861307d2c2d4cf8c6906c8407023b4a4ebf2a387bb9e5e428ba5d21ed7dd0d4b0120be47e1880b3a1e69f2443b71a400f63c2a1e8ac9b938fd1b0aaa21841f55b59e2d9ab7f23ad7fd9db4ff0fdd436d84677243ae06f5ceba5f6259433db0470bccf28c79bb3c1c5f11adf1f11bb02391df0cbc339d38a4e04e0cb9d9e443a784f475c452a5e7a8fce6dc84492c90b10ae911e6394af6fe6f6a402879112a12266f7ee43ece62890d2846e06be14c5c667bb99c3e242319894aa6133be5904b1b5f56314e8dfe4cfec357f08f554cdcd6915930e84b7e5e247f5b23676511d7a4c4be19c88dc5e8baaad8ede995c50f5267b818949b132a18f52890ffe64e2a4bd865006138528793ac81480af5fb10914a243b2ebe69cc885df7ca80c98801ab87d14a2a3a169f605fe0c7318bae98fd64d37dbb2ee1f92192664bdeb9f75e06c090782f113fbc1404f47a39ce6b7a6c863c2ee509bb43491b6688f3c86b2d592ccd2183d2f197e9446e34fbb9082eb35476c60bdee319a281c26fd8576f33ff227bfef7a3cdd4b0984f2a324aa1e14472cbf64d7f430f5e86fd96a483c0cb575e26f02083f5d008957b8661ff0b0d8d1a0c1fb5a3cf935f28ae9249a4ddd9f0b87de9e353e8851d703c8c63432121b823406029970a21d4a7147c37358b8475389931709da3b1bd477f663a5165ade99159f8950a40bcd29c3978e9efb4da81551e6cd3b815a511fda02b54f9123dc6147785706b28232f9f2385806835102c1f5da78efe33f9a63aeaa74d448d2aa2ead2dd7706ca87c54cd3692aad24236bf06b54ebac4c26434cdd8173e4e83590ea5d452880fd43546ff5154993dbf19d64690ef3b486145a016be6bcb86d472ffca478bfd2158d4dcb534601f904194f719c584e1718f183a4c60de17c586570293106c8b0dcbaefcb3270ae41a0577f721d83e88847409d1375fd73e7e14522d991c5072642a03f7aeebeb4123616da7fa685cae1c01eb4fd2aafbdf9c87fb8fd64a1871bf8d8d32a3a0660a6faa6262de7a306113ff708b12b5d98c7f6967d971ed27c679c131bfd7fe87754639c0d687588b707935de5198c955a263c2b464ae14c90f6c1aeefc800ec00f5442d8d2c45c6136ea816f6779fb2dacd350bb1755045e30d5c5868f16e420c6f2c31ef5104ab5b3cef8532ae2abcfeefbc768b2408fff00ec6652b6bddcd77c20f240e05face7d7b1880ffd5a7f00b44b46d3bb19f97bd5d711e7251dbeca034ca18def136dea2d67185dd2e9284c1b081f17e22ff478720003a93f9f094b7d8da6f2610a15017f86cf0cfe4357c3f19cc0cddbba1029da3db905c03ac79c5df153a61e00b25a808be3be4adafda67cc30e96b4135a254ae0b7290876cf3071d3dc79224fc6c4544e0b96112446c5102983f28e224351259108e8e0647dfe159935de494974f566237a4a88773d1d77251f0a619782b3c769ae27693bbc63a38002cd3b424d4a0798bef2f1c44742f3216fc40cd9c7f5d1d807cce0af3c6f86651ba445ab6c0e396bf09afd8579893720d05c33959a9a59bf9b2216c1704acb3ca4ddb9990ecb805a3e852cb9732c689ffd548859b6b29a59e3820235b38683064d08c99692e85e15717596cf919312d6f6cf3b1cebdd57d6e46c2a087bd21f5e3fdf1d977eaf3d3565387617b6489900e7b498355c9e1054bc150c4f98f000b1b597146ba94b723c825a1b7ea5932332635371668baf7cd85fbf737e3b1c36bd5806af7a983b96daf9f2766855e7f6f72ade0d33bb52d220fd074a382f2adf15a17b146760cbd69a9bda1f3f3bbb7c2aeb62786699e65a912771f974637e40d993c705caf331d0c855fefa13369c002a7b92a4865bbe099525f095922b52b29e83d2a3860a2908d14992df9689c9eeefccf41632d23288fb9cb071fc0deed070e1a6e48c200a138f9dedbbe01713ae8ef1088cce621e389b493a063a798281ab78d019e4baed9c1bc84bfbfd635f59d4c1645dcdf5c2be249c018b7aef5785d5f4ece085d19eab502a0e4fd6d86acee6c6f120ad967e9ca3871bf553ef33698c41c714441d344f9963b9708b4f4c3c478470b10de77f238323a777da48a8dfc4bf5266de51b916943631a89ccb562556d43a35a4f20a3c9a83389180b4d3749cea8be823343bd8a12dfb4aa1fc21c6a0e37c3524913a088f9f5c14e7dc98acd0d10a632ac019723e4f714cf61bdf424f6f47db5fef3e1287f25d75c944a86fdda49ee16b4baf0d142b3715db583b7b369012f4e7d830bb11b92af75c949e6223ce153e8a28395602462528af198514ef7b433d673fe874c2b707287744bb94b05c2d4bd9186ab7f68d3bff4e7d6da3f07084aae0da8bc6a4e8a9fa386ceac57579060fc9dc2e2daebf998704042778563228817a1f2b5d477340d0c021a4a672c3808a2e7817d9da29b4d4687f58d266e0c920caf6b78e331c65ef2411d35289504e0743f6469594f8b976096e5b7fd694fb6e51bba98ffb01613a246feab56436f2d2b3532f3fa1d2fa5048c84ed35b5082c145c1458f95b7394ae392fdd27c37f7e8df6f03768be9bbdab97affca0f234370efda3cf2d1fdc3ef8713529c9b561fc2739b86190c7968c6cab03aa099bbb1d19dbd1dd00a1b766f440637511db1a47a55b3afca56ebd96a321710c94e86477cd87083ca7d743f9a8b2e5a5ea7b7810d31980fb1c9c1b11a5f1345e50f4301b3a69722ffdce723cd2aa0300b27c3140bb40ae25c1037c73f83015ad9916a8f39e2fc642aca5f319594f3d59859fc16b4a28433bafb342e21e7d4cf32987a4f4fed4db1b2221530ef4eea01044b37f9aedab893d3577e176253836ba813eead1cfc05b46a319584346661f33173c274803ff067f9b9095137128e3dd17b5656a06831efd3c7bcc84d3f7b31da152264c6df7911dff24e28c59a2fce4456ede804cf5f93879a32619623625ac48a628a21c2a8f1d12803a660bbde39a971d68bc3b57dd7219c440be87b9d24d68a25a4d405d896bffd0566186b65cdd73499e6c90b8fbc3f6ab200c146cbc4e19b58e9723df2e2c11aefc036b66b70aaceef92e0fc70ab97a327367a5f0783228aae256dfe56a5362b14fcfe26661c669c39b8fce3e5c5f4163a2e23d322637da8a7a819611c526670f087e0af1908c4149c3a7df38af8f671025ca666a268808fc5c529685804a7405d157e6dbc8ef274cfe1da88657e38bb00784fce3d49e1211afe77f82b33632acb1d89388ed949649b44eea06b4dae6dfada1159b635af94fce185187f0acafc7c9130cd984e98bbdf422fbd960744fb583e4cd76e815894a5f42bfb15f4277f54a96a047fe1305885d0dcdd6b37ecd2893ecbcfa5939ddf64a83e17e4bd78f4979553b1c97180270de7e9964b672b77375049aa6937490f2c81c5390e064828258db93b4696d404dec9d84c79e4b4bededf474a08080c00edb2057b8389b7100cee012b99a81cde3b7fc22a1c26af52a18340055eb7391741b30db092af1be0f0a4613be4ecb204ef24808ae4e7d0dab08b130e0a77af7b95a96c16033bc2b7231dbcb9ab8878c390796bc08a57e18e632b3f1e9dc8b6b589db96e46bc28e70833ffc512917da4bb054e0735d55fec25b61b17c0d2632467b50e673d42549e29b1bc1e528e9ce29e44cced2ca322ddf2c6bdf1bce9c99d0b58d1de7516a5ec212ba92d04316954642c1f031af8025bd048810cbf7b155f3cd0610cc4465f0f623c0515a0c760ff33504107006b88535bce4d5cee2a487a28cd4c663a82a5a7a9ffcea7879ae8e36f7315024cc96a17b8abcb11c960a21368f008b813e4b0cb8653e9a2683b17ae28cb1678353a426fc1bf56a9f506f33428210859ac1dc24b5b9b9efa6d236ecd2fe0bbf9d6cd11acee0b1ea2cffc5870aed6161736da502bf9be6417e7fb4469f0e76a6dcc29adeeaed9d7cc43dce1226e8fd78c2", 0x1000}, {&(0x7f0000002980)="f58b6251a63a0364ef6d85bf7ba95fd9c89ea6131fb5a6bf82f4cab5fcc5c4c6ef4a3b10cb84d23bcf81f944716cca974b3f4182644ea87893ee0db426d672738af5aad99caac01169b1901583658a255e3b6289cf192438b044b9ba1b6512aaf0da8dc09ea54115f138bca67b1af3d2ebd03519088fbc0b0cd4ba1d907bd0b402c6ed", 0x83}, {&(0x7f0000002a40)="7b6066e5bb915d6aec25b3b9ecd359fad143342a232c6619f415f524", 0x1c}, {&(0x7f0000002a80)="fec763a6076bddf7b82a1d95f6300e40b92190de8acbc84c41cb98bd916ea63310d097402a6ca7d102ed535b05026459", 0x30}, {&(0x7f0000002ac0)="1954c344cea3b4fa405c9aed30c3d3ca65d4c6aedb572264e09ee23049878bb5335213f22a657aabebd9efc9156c5aac4ea5aa2894e815142578cb67af5e731ba6a37b3552", 0x45}], 0x7, &(0x7f0000002c80)=[@rights={{0x28, 0x1, 0x1, [r2, r1, r2, r24, r1, r1]}}, @rights={{0x14, 0x1, 0x1, [r2]}}], 0x40, 0x11}}], 0x6, 0x40000) 00:28:43 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x98020000) 00:28:43 executing program 3: sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:43 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1700}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:43 executing program 0: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xa5010000) 00:28:43 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r0 = socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) sendmsg$unix(r0, &(0x7f0000000700)={&(0x7f0000000000)=@file={0x1, './file0aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'}, 0x6e, &(0x7f0000000640)=[{&(0x7f0000000080)="e788da0e8e3dd3e1214ac8b8fcbc8e8ebcf3ad7f27912de41534fdd91f407accc469d2ae43724dd669090531270031367a9e69608640e01f2a88c17ea7825ca2a7465154956759ed879553a94e24eb7a37cb30abb6dacd5bb3653ec1165f743b895643197860000f2b1f80603b65f80124234a6e341fd099fb458b4e82e67914bdefd352b388aa18fa6ac3", 0x8b}, {&(0x7f0000000140)="430fbec6b6579e10aab2a393ac54d78a7516da14851ff9be9d75abebc0a8cddc2238f5f5ab0016817c042a1714b1cab8651735217eb2e1692c7b6da59570a288c1b78fc64c6614bef5fba93d066ace8e3757bc4b35b47280d9e53d3687247f582f992a520cf9b9ae3c724d8733127cb4db11b99a312872838eec892dbd7b955e97b596dc39e258e3", 0x88}, {&(0x7f0000000200)="55361c3ad3719e092ae24d35e52648f4ad3aea31e5235fe0ea3c8934b506e42142507f11be0489a05c16e644d340b8a592a35700e83e8b3dcde10d3e9fd475b6b4d8251ee711e8ffaaea5b05d8adbf1c8f7718e7988a4c178436a8bf9ddcc6ed47300ddaea407a5345034070d5deb2638a7c351581cd4add7672d68f9aaf1780298dc3d7d67d822757869da570cf687c522f1838e3074c3b6ddf69d3d25fdad50e96690cd2967604d6dcbaedf8c5e9be2a1d360c151781cad8d21deef236ac8c72d930b07766b7a1938501c0a5b2066a29d4c88ba413fd874f7676c68fcf98f2d944e1bf7fa62029dc8fef7c85", 0xed}, {&(0x7f0000000300)="e828d896f715c6d15809cd113f47a5ef8aae158e3739770396bd32d3b9b85adee260f09312fedf2203f6008415dc82e21a5113ea54d98e2daf62a05a9205fdbddd9eed4e8ca64ce78a8cad401297c5e14a8cac05df73251418dc704a1180f5c07f6d", 0x62}, {&(0x7f0000000380)="76cedcac2912ec32bdc2f31969975861654fb5e932791b0d7446d0f278580d0c20ba19bda58a8050f5083b240823b8712474902fcb083ee88837add4d2dce7c351b54e58023c7872cb82cdba144678ac1e91a37f8657cad4e18bf9566fd2cf04d4bb994e4510d9bd92c12e4a799f1d8529d5719e2a85a6980a3017509dd2b8f386247d58ea24126b585599167dac5200bbed120d6f703463047a83b38112bde1b79b16d4dda670f98afeda76", 0xac}, {&(0x7f00000008c0)="fc0f4ae8500aea363dbbb44a85fb519d50efc83e50fdd1f1c9b56197e54c2f7eca6abb2a58b769805efa19b9693fd16b19ed04e039a81b4e97045284f1f01aa2299e60e719b2ff31ceff21f29fc6220b9d940a024dfb3553147c6d4bf69ba8a7534b7ff8bef95ff8127ab1702d049a7c7cc04639f272672f21f2958820728a161196c1172fa3ea5dc1099b9bdf7c090bf7fc6242713c09a7125e3eeb86f9f1c3f9ac505f39b308152631e81145d2fbdf02f83516d92746f4a79bcfd32a634abd0f7f28009d8fb5e6b870f54baa7f8e21eacf074ce374b92056b08b59a41375a83676ae7b03f68f94e1d3d4a8c92fb68f5a4df65af1774343943dda891a1c8ee55458bc72f28fca9e36e24828f7ada607a87ed922ba6df9dc6cea472624b5abc80407905bc352bd77a3cf208b6cadd8d492c8f20ba1ce7df84ff7edf3fd4bfe50113916e1702825a12159663aadc98712b7f070b9f0e6c43e7974eeaba0b45459441498a0bf7fceef0c6e5a50ace72ebaa6326d83efbe7e74d4955dccc140eeee560aedf59b92ccf9341faf73260ec016bd5e7d4ba4dde3fdf586eae5278b82e87f1d107a1ee1871b28081ebe595adb9d3ac788fc894ba03ad2e925027adc2b5c4451ae1f4953435368ed6bc24c837056deca98e5a6270f05dcc0ab1a53ce3700455f603718ecb3a1a88d5e5f2c80fa39faebc4067e406f025f64a18ebc0fa763aeebf075e1484117241f3731f8edec92b90ff10ab703698730f71193a297cb2c317eef784d3f1492e79d0198c4f76d9e4e2885e846e86529cff869370918ffd88f80da007a1202e1fdca7aa0751008e008ad50ebbe16991e2b1f33dcd4a05a842102d8c0c4db1d02f92dbf73a7f152a5490c53e89d917097a56e053f9cdc66f925625dc53a583edd6fed4ec70b92ac08dc5aeb269eb7e063fe86a4a06441b0eab41e4597fea30fb2fc177889ac4a961f7d891f5e7554359b989653155a01022ee9b9bf46f6e843b3b5845952061558a71d8b8ba091b0826545fe768c34d69982ec9fe2c8899185d10f775b8a975dd92efca8a2d1732194c5ffe779e4591d44f3e54a4afeba865595344876e661b8eeaa5be401ca9d273de8b2f4f3629442aafc11e1108f996ce921f82643bc1331b8b6fc39cad3ed065f6a4372f6a6ab399fa415fca488dfd67c46fda3a03266c0a85a2685dbcc77bd55971fd85495bd03f55c38c41f427a88b5cf0174482a8a7ba80202fb25922ebdb88aed4154e32852b956acd968173b5b081367c5f972ed74b0f9c1176dfe93b7518ed45ce14c120ae3be42f23b5b4fc35902bd64c1526aeb5368da25d7367dc2e18927eac3193e01a3da341629dedd44434109f3e6fffc88bba961563caf05755d184813501b93f503afb20b8027325eaab784f5729c9af5ed57e06aea86a66d8e7b57f95d83161946584b0d22c1b2fb3dbcce1ea5d23f34bb7ebb6e47d44a5ac1287e594af1c38c1533fc0a034a1de07ba6d61db73fe5059988c7a9c7b6b6a972c6e2a441978a4b31fe155c70e2c0a0a808f8a2d22326a672998fd90323c94ed8dcab9e325d420e23564617fbbc15b8d47ab657094d83a995ae6c29cfd645e410e6c71208ebc155a7da80f27373d242509f27eed918da96f1087f31c2afffa085d80cd1fd038aed4d40799b17e71196994e594ffc565d1aeaec6b92def971e9e28522f14195f2c46f756bb3e59b119ac5dbb6e65d6a0b91a6ede787fec7704edce8af0f3283db6282e00bbe1b77084353cc0697da5cad40fddf099d321fdf13c2a83d2996d6de4bdfa2762601e82ab5208b92d08309cdac39c3e26b6b9815e896d5522133ffd34263a4387bb43ebc8e0e77acd7ac51061c3432be87ab95fba045c1a44caaeb157f878628a97150483c7a813d3e2fd8f4647a2ea6b82de3045465a8ef71862b406b9958e4dad8bda91fe30d242185931531b87f3b47e198e452a1866faab99e902ea5a6907cb969432f05080c0acf1b1c7a0ba3134bc6607932ad94121e9484527a3a060524baa2857be7827cad0364c3334773c654688fa5672db827ade5ba5a527722b08ad4598127b8e4f96b1ccb119bc6727191461359ca25b21c0bca86bdc5e753c6509e3a23786244aea98120f78c5b43f879bdaf2f59d22829e7e4ec6791fe40d245f881e33816800cff7b00c232b4bac74d5d7eab0c3e477555e9c20b71866dd90c4949dc5829a0c8392130888eebc4fa752abab419a6568afdf3404609cbe63d3fafdef4a557cab617e944e002d768f0123018bdd7b7d27b9bc145f13abdc53bc979e0de421033c69443c660fbe955c62bae22186b4761666e1348202a64b4893e92ff785ff877a426e99fa329b66c5057337013b5f6e7ac4da159664d792100832e8600223aa536e441bb2bc514cae0084784f05e579221ae0b4672e68a8d8bc1117d477ae1e40f30fb6863f001e34c80611a3d920d80f60256dd21b8754764e4041855aa694da1f524a51caecd678589ff9ddaa8516f15abac0c0d45ccd5df1061cf45c8a82b78263e4107517d6ad2b03f6a6ac5be38714e956ec431a71f6dace96f1970e7f52dd3bff553b15e6ef5d8d3b997b941d8affbbcfc920a6ff78316ed32fe07e6acee645aa18044e9d4836f5e8ef1b74e95f34dedc790c0a44d8ff1fb5acedd624aa2fea58ff12bfe9f31a7618ec875186355fa941608c3a022d07b696bb026e2f43913bf1964545d233c34d48ea0bd209dc44e11fe1f2bb0de16c2a87bac8e373325dc2696bde07b20da2e6733da66cb2917bd108226c51df8a4b8c6f1a3d46b7bb4a64220772b65c5d56c59f5cfe900508ec9ce8b960997926005eaa44a429a1a8bc61d73582d5c88932a457e4b033e7b6b367a771f57b5d28b3f2b8419c22bcd1c190f703f54b9932a479f06367381cfd95de098947f742ef5a709c19223aab7a48d1aa6098ba319b0b4bf43393e230401f75a7dba3b0425be9312f732fafdef82311bb489ee4d687141c61c8cd0301094d351c1327f84bb1a7cf142c708375b946e4daffa457b6c9bec70a6660c25d79f8cc1b0017899d8ab7bf03ee7f216e09c762a1ca5f4889b98fa351d85567d6cf12c5306232f3a9b62eb13c834f787d7b4ad93f201f694a0f146da66526460b2cfe72118d07ba6b7e9530e4675eb9a92be5a83707e101ae32453dbde5ceb35d2eac7ef624bdbdfef04686cd155c861a577776972b54d75e7e99376fd5a402947682d1f84c1d14f5c0bc2b045a086a97460adf1a63c342196dbd3c41ce270544c7c275f8156c9e244543bc5e78017579000d4c643899563c1ae57c246539fa4a1bafb6b3c5f7ed308fdb5cd15eda28bc4fa269c762b87a289afbd31eda08469129a2e3d28f955c42c78d2100ce534fca836d2145813afea576c59e75c078540a9a92adc12e1a265e918855c8a174e0252ab5f8114625fe78eb07f3588f728cac6dc5873bd2715ec0bb1852d33256b036bd14ccdd44b0d04be0bc17a0580828d7a24a17524b3a6442392eca1a844b5ae03b174761343ce1ad85e4ae1d7bb8d4eef8c52718c341e33ea5e8dbf8b053fdf6fe5a9e153b6a9ff1c24d8f70c8f756525f38876889d652c8baa920f228b91744f238ec3dd653df1298a73f58bac6bb0c0317030186b592cc64fa1c467f04eec09f3bf48d6adbc8eb7b1ae3195d08168e65acf5f5c4c2028c3c798fb0c89edba2f67d76ec469bb076733423318182b1ef7fa759dbccee15a2ec9db424aafcb52862baa9f7d7db3e9ee7417d9467f6604d505166ba9d787dedffe761a797fb13de536c4d0e69d85bb159dd8c895c93b52044905dd3c4f6685a522c3776782f4174df7f80cfcac747d5d4d77a1b904557b973e220d0053cfc91e3bfc03420db7de22e972516f692f12e010039308b5698c4e8f7224788cdca6c6534f88c8c9d16eea65b66f1780bcb9cbe593ab44e282acd19c39af0583f144620837a12b04174d19d2867b777f32061ed284f86287ec44433645a694580cf4db7d99b96a7646fd6e1bb5e1ae29853c501a2a8021522620d6d1e22194b28cac92201e80865d3126df5dc34115aac20711573d96564db335225d865c3096b9181329e488d39ec19f7fa5ec1bd9f0ef7736018df725bb0418cead8b2cf5465ff83e2a39fe5a5ed7f704a9cc5af43f3a488c36c0bc6d192452f7e0a2002f65718db8cb6aedb53855ff4d2ae67c09c3f1f0356354b117ecec06b27b3521065270a1f52f5d8486ed00feb26a537a20b03d27357823580bc74e60874cc6843211ce182b656d8dab56b8bb38ae118686d8736b574588d77ff094974b43d974962c6cc86b5012c7d1b0e6ee7d1d690dfa1288b05023823c21ae929271bb48aa2f1ec455980b732d3d1735eff6b7ad58e96273e453255e7174f8c4368ac18c4115cc5c6f9aded9b6856132950680b236d83db1a9d06b172b0ffcc3571a68d34cdf7ad74eb2a4676022e8c20258b02064c26d87e2a3861e6f353f4e81d1f0d50586321db6c6d7c6f5e874a80d0b8df661d9b2d517aaec5ccecfdbc7bd51d30929a71467fe102532cd455bd353bacf4f6669eb6cd8c0fbfe8306e0574cfac2c8a24aea1e1fba10a69e61e33e72325d1ab29eb93b5e160d08a403ecf2d504c9d5fd52a88e1d25ab939dd8bd0533974c1dac22b7cce4b6032d0256e6a9bd053d8b4dd6ed5d198d1bbdf05a7408e18f434fd5200baaca15e3f2cbcb272c7e2d9f2b6a61cadba7ec4c27c449d88cb47be40cb8145e31a4669b9f7be3a6a4847f69dc25e5c6093a37a2ae273a4666af9c4131a4c2d28628542e8a1d99080d04313f30b7bac98213b1080c4230264adc2aa1e44bcb1180f521379dc9e70e2c05399e56b4bdce62635f3c3f89e57b3dbd9d86d1832e0c6f6ed0f158e8480e947fdfeeaf8b14665f6837c86ce32c0102a71a14151ed3691c79554dbb881af990f5b0402b7ffc23037f01a484d222c780d6f2e6dda2923a6254214808cc5a65fc0d20bd8910f8feb179894f0c6203dbe3f67fd023ca3c8df11b54e56800b6fc10fc13394c4c11944485af20f0e23babe6f6e7ebba9411c7de927ce1ccd4dc467bba5dc464552790ac9ff9736779b9cb25dee8811a4c1a9ff1101f0fa5338c9739140142683fc9e2775c09d3284ad29b62dc3390f062fd58316180bf792e18cfdeb1f275b5c7d12f9af17967f55bce7ff68a229b4d2bc9427ad23734452ef1c2b8bb98724b146a8f6718009464dae11cfd23bde6f7401730b3b688beadb481aab1c87f29c8e7508b7b0a4caaa7dd497aa39bff912487c8c60f30d34ff47cdfef581c0fee60f599aab8cf1dbdadde70dcfbe9a6d1a4eded8406a6062014b911704952c6c7eb64d4aff0292ce73df0400273089b161f63c3ce155c66e51edb3eca9835ca89f9cbdfa48c014086b3ae5c8a50cacb63a0b6387e6c0581ada4f7c392f8e0b8b08d047f30e798a322802db81bb8ae177876020f63efc3ad02186b66aba129a7be7e6df4a8bc17bcb2ac7ef191cbd0601eee8bf696c62d6036d67f317b58681abedec64f5b71be7c3dafd4a220b16df5424ad6d37bf9b0d46b32c182fafd242a4d05aef5a5e9661f5ff98f077d7734a0a52a5f729b33ceb49f10488186a17c8c990cdc06216bea5837e6e8fbe85a4e09b0ce8ab9d6b4d038d7c841e6567fd5db2f49de514bdac138e09224a329cc82bdf43a2c0719a0d906bb8694ec9de23488fc8661bc50e766e345a9f9354f2ee6d8974b1a70441dc11a84cf85fed58152f64ba133c922374d24", 0x1000}, {&(0x7f0000000440)="dd4072b34ae0a8cc35fc1de24ffc0fcda7ed7f010bf7e39cd93bfd6ea9067fc6be4468eac79652203fe18508e0abeeda93b428d96ba1896f00c1dd7812b9651b464aa221f9cb89bbfae6", 0x4a}, {&(0x7f00000004c0)="24b3637b028edf0aac6b001f3b1c754e5019112b18724053d453f08e2deeeea049118c1a1ddb0e3259987267801bf3233be4ec8da8d9e6a5449a1f4ee7b50975c59385264091797608e4f442eaf83ca2ddeae3e3e462d40d7cfc09353882ebaad265feb7d99a4cc3bf789fba53c82d382e3396c1dce1a171285a1321cb1e12f57d0e83e75ce564f50908a5bb6e32a06e6b9568b1e9c19f1b7f3d9e926f07702490fa2ba4e718ad10a17f9ca6ec4e6a6cc2621800a27081e41bd26afad8520a9d184fa83a355d81a2abd54177edcb3b9ffdda80aaa2d726d3b0e5d121dfe2048ebec0913708962fb2b3fa27d0300bb1f01d9c2fdd06bcc4ef8b044c3e16", 0xfd}, {&(0x7f00000005c0)="3195a4e6de7f6dabf297685f86e80498b0d210edcbcd5942aa62a37e8c419e22f247dca423f92af1e1d9d22075c691112077caa7d3ccf387a85ca7d65643c8f53ca64b982ed180a423b18fc80bff71bc22438517f211f1986594db8b0817", 0x5e}], 0x9}, 0x840) getpid() getpid() getuid() getuid() 00:28:43 executing program 1: sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:43 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x5503) getpid() getpid() getuid() getuid() 00:28:43 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1800}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:43 executing program 0: sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:43 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) bind(r1, &(0x7f0000000040)=@caif=@util={0x25, "989a93b67dd8f49f40ffe5bb900c3661"}, 0x80) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f00000001c0)=0x0) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r7 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r8 = getuid() r9 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r11 = geteuid() r12 = getpid() sendmmsg$unix(r5, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r6, r4, 0xffffffffffffffff, r4, r7]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r8, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r5, r9]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r10, r11}}}, @cred={{0x1c, 0x1, 0x2, {r12, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) r13 = ioctl$UDMABUF_CREATE(0xffffffffffffffff, 0x40187542, &(0x7f0000000200)={0xffffffffffffffff, 0x1, 0x8000, 0xfffff000}) r14 = socket$inet6_udplite(0xa, 0x2, 0x88) ioctl$ifreq_SIOCGIFINDEX_vcan(r14, 0x8933, &(0x7f0000000000)={'vxcan1\x00'}) getresuid(&(0x7f0000000240)=0x0, &(0x7f0000000280), &(0x7f00000002c0)) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000300)=0x0) r17 = geteuid() r18 = openat$smackfs_ipv6host(0xffffffffffffff9c, &(0x7f0000000340), 0x2, 0x0) r19 = getpid() r20 = syz_open_pts(0xffffffffffffffff, 0x400040) getresuid(&(0x7f0000000d80), &(0x7f0000000dc0)=0x0, &(0x7f0000000e00)) r22 = getpid() tgkill(r22, r22, 0x816) r23 = getuid() sendmmsg$unix(r0, &(0x7f0000002400)=[{{&(0x7f00000000c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000180)=[{&(0x7f0000000140)="a2f0d12222dd29494720", 0xa}], 0x1, &(0x7f0000000380)=[@rights={{0x10}}, @cred={{0x1c, 0x1, 0x2, {r3, r11, 0xffffffffffffffff}}}, @rights={{0x24, 0x1, 0x1, [r2, r1, r1, r13, r1]}}, @rights={{0x24, 0x1, 0x1, [r2, r14, r0, r1, r0]}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, r15, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r16, r17, 0xffffffffffffffff}}}, @rights={{0x28, 0x1, 0x1, [r1, 0xffffffffffffffff, r2, r1, r18, r0]}}, @cred={{0x1c, 0x1, 0x2, {r19, 0xee01}}}], 0x108, 0x8c0}}, {{&(0x7f00000005c0)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000b40)=[{&(0x7f00000004c0)="774abe746686470cf2a5184e6d084ab4e6e0a7ef71f02477b3ce8cf639d4bec031680abb0c08e5c34ba3e3cb8d5e14", 0x2f}, {&(0x7f0000000540)="85ac0b4c78faf59edfaaa19ba7d874e4293614", 0x13}, {&(0x7f0000000640)="facb1498e4b77704757538110f0c36d751027c57214910bbccda159fda24e3c503a8df01c71de8f9d8", 0x29}, {&(0x7f0000000680)="0b4720db8d991fc867657defadaab95a9bc86160e23715f7a4a54e29fc84d0ec7820c70e5a8aa502d91bc4371175a51d3259839d7ad362c2e46b234feb41f8a1b0c05903fca67f3e515140c2e0bb7bf69844129b1b9df18dce96336d1810194c6fc606d8dc6509c45eefc82f64d93c993c66a71052b644d7f143fd53c89aa5b2340cc49cb44bca59e5d9ea8f8d46288db71668282b24af068e652fc0ccf9b7b1c68223b0c4d41f78b5fc7abd313cf47b6ddc0dca928620", 0xb7}, {&(0x7f0000000740)="1ab94c45d888547a7f1adb75f41d87a637f79ba33253ea1343eb645aa4f6e53306024580b39aa33f4d648154b1ebd345e4c6cd2131bb1a98fddd0422d16eaa7bdde1e4106aaf92303e304ae6d303c79bb60c5686a05bf104a05779bb251d61122270e0046483b2c10fdc5f07cc960d0f67c7e62111e035653b3140820b179abf768abeeba3a0c514ae9e22e17a24590e67209471", 0x94}, {&(0x7f00000008c0)="db1f1096417a77bec9b051e9383eb05e81a3120eb1c81a4346ccd490b58d813a6cb3456d6455195196662116a9ba9c62b014311cf64ce69d7d86ae64880069a52a7d021cbe356e1273edb81c5579c80a68c3669f314c9870d20848e4cb74da2839cbf3125be325f4174352eb2623f8f985a798cba7280b430847f1c53ce318635bf13bdf67974f01bd66f9d050bf1efd05298381178c72dda7627d4d03f41a21eb4c3f0a9aa21ef41dfacaa77511a3ac3e10b1ac91180a42683b0e49efcb76b877f389d32db16062dbde711fdb33aca86afe95f55965c6b908ea9655093b780070ea09a8739fb14e8414dd3d1f3220ed", 0xf0}, {&(0x7f00000009c0)="34bdd92d60ab7b4264d48b7ec5a9d9b24702e2424241073fcd1e29047010b7ada9b609e2dd3cb85dfca292da79bfe1854a24a86cbf9270c85961a106a1027b0ed012662f3416cd7797bbc6ecf5e54d8f057eb83051c1cc72f05d2ab02aff2e7aa99cfc87c05236cd74984c120c1c04df0391bf1a2ea33f2058fe18e4a02f11a7581c1f4115bc46a82c28c05f", 0x8c}, {&(0x7f0000000a80)="fa4f107f1a4e405b0f8c6aee172427f8a587aac17d28ec4cf440f9dd7734662d54a8311e1bd80c672f7604b5bb9c459b10132a12571b6edb2b079bf07e80b909a237ccb0e69f5c22885321781ff13892cfffc84a39079ad37914512e1ec0eb5562f09a4f4af9bf0b2f1609a4dfb74dd56c5e5a5f72f7334ae1ff6fda4db9ea4941ae052b84a83d8c5f52ce841eae04d924991a06aa", 0x95}, {&(0x7f0000000800)="5dc66e7a1a0810683168cfd46af4ffad98626025459b889c418c17bbb67c275a56e96a41d76bdc33bc91db2901efaaa08c633b5c8b00a48470707fa3113eb9e161e3e8851a8c4a0fc5bb2313bb", 0x4d}], 0x9, &(0x7f0000000c00), 0x0, 0x4}}, {{&(0x7f0000000c40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000d40)=[{&(0x7f0000000cc0)="5054747e0cae3571d87bde5d51e6c5220f73d1b2f09bf71c97d5efbf978cd49442a12a1563bc18d4f015a4694ab9c1e7e0f8cccd86f4c98eede0544bfaebdfddf46738ad6cf64f066f5fc1c7522b93e90427", 0x52}], 0x1, &(0x7f0000000e40)=[@rights={{0x24, 0x1, 0x1, [r20, r0, r0, r0, r1]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r21}}}], 0x48, 0x8081}}, {{&(0x7f0000000ec0)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000002300)=[{&(0x7f0000000f40)="f91b228507a65d8acd958d73de9502a29dd926b6869244b52cc9093a900598751b8b2e43a96967796b0989ad6aed0560efb92235da57cc7253c2e55345f8af7f0cedf62b43f7b97bc506f1f708cea1c1988dc26a75357a3739396b96d1c3d13178ec2155cc62449172dd6a88a5f3751917d2ba085d03d630a8cb49dd6447df1dc13dfe93711cd3b16044d92726e3190196ccce04e4d7d05f450ff8312d01ad4a12cf7f0d060f91ca8c3865c3b607edef128af28a9921a1e5c97a27354af6da52b34e5e027388fb1df81440394362d39ca74ba492513d17099898e539598cb25de8c7c00a6514a009195894c36b1d058d478e9341956bd3c283aeb0cb106c9ef8c73d367528728d769c84820259367ebf6817db0edb6a97f8da188f4c7e0eac0c467c90748bf3f09f159ec13d85720407506844198dea186da5efb2c739ca3fd4e40d68260efef0e678fda0dfa1a3bf1ef927f60693755f884f2ffaf0a5a13002e7de294b06b7c292f2efb592f027ac3c724f58fe8888f420bb74cc15283156b176a3d35f8dcd7356d3ba0902bfc52ac8a2dcca49a236d7f6c51c944acb46e9cc4a40607e34f6699d0a4970f3a9fdd793c96520d382851b3ffffe11e0a21358d4b5359a655c4b123cdce511fc250a8e67264791002c89ebb736cc0556deadac8d5e8e2b9c5f1fc30bd22786c58da4869d3cc6bb2843138197591fc8f484337030213093bb9369b5339b789d961af830d532a2b819e99aa9248de97bf096fd00532e943bc47db3e414040709ec88d742b8300b4694006afaf118692fb23b77c9141ad7f40db583bddf7e14dc961210195c1921a20bb4bc503b61e1835d490edcb8391cacfdadf1ff2d1ddda3d986a843cedb0e94b0dd4b6cef5cfc2a4698899dc8c7a1920167f0025f2a452137022e92456a99d8aeb65940c718b90d4ffb6e4537b599080e405cb29cfdb6149d186c07815a0daad6b4af2ea2d675df42b67b2e4f80eabaec36c6168e1d6a86b6d4172d6fa76f3df868327cbb0825c37644052750a525da021dff3a4cfb0ff32b39f7af425a2d931074307e091bcb59b3c63aef996dbcd3c238d5dd0dd0bee37efb722283bb13bb2cdc354f7eaec2ade78df44ae42491a52ca1e4b7a238ce22bf526a99a79fd3278c0195bc7dadea4063783b342a7ff016aca8a174134aad2d9ad0410510a3ae1989798846418e2edfed02b108e09a5097cb7d4132edbfe3e4dc93c8cd45d0142a886744e9da45ad2a22e832b4894a6c821fa3ceea828d1d39afb99c90ed310cba20ec7f5ea84edc0c5b0141027653d2bfe7aca63fb9c4b898aef2ba8d3e5eab9f9b83a0e402ccf8a6b93e496fef608358b08d774220928c08d06a63b7a19fd8e843752ae7d103ea7fafe82bc8a4137acfcac6dc8950751227833b6ecdca2d6f6f7d39f4f6ff699059bf7122c3fa82baabc806ddcd88528cbe02b09eceb8ff8d9055475415dd0be19f85925346d8b856791778ef722b64bfbfd3f11b5c712bb692d1b58d9ced6e3d24006cfef8f836ca7086abc5a18fa433afce08ffd971f03b185abf2938ce3312a1b1cd407126413669adf32ebce2b92bc5f9e533863b83526a80652766c0d73bd37ad9941bfcda0bd8a83b4c59eee14f99b2f92aeae4959100838a9dc002ecbcd0752782b1abc5251a70a4aa5fdcbc2b721e6b1b9e1b98f6de74a4fde2a09a4716e2bdcf72c65b0ba94d8d83ce02df70d12d6181174afef53336fcfa669e1cbdd3a159072c7e4232da5034778a1216242182dd7a71333dec28862306a3a10a836122f281a68363357dafd9eb91f3b57d2d722c79a41d072c9060783480a4fe32ce757edfee4ea2ba79c61c5de232479298749dbfc3ed47f54ee249f4aa7a51283603d8196c019706b8afd1f0da98fd62695ef81c405225cadae5b72acebc5afde67609f798462dc4fd512f9d37098d16919c24f0c32d3afc6346e242ad6aa4671352ee3017abcc287f0b641dc99e8ef9fa26e4733bbacf0197da3df7c907d822b212ee2e556ad4ac3da925ffb00aaf1957a0b62c92545adeaf93f8eaebed0a9c1875d073fc386c2fc15bb6d6eaa1837e24a9aa8a1542db86931191775d14af0992ba5f150b62b68014d604e67cf4bb566d8c3deaf45d8bfd52c7cf921fd8ee228022c4dff1c406fdd3d52d63bf685aa33254dc7fc0042f03d10150beed5eeb5c376cbb9b6905a160dd1ed8ba551331070238d935818d278bfdeff1042fe61cc741809f6663d66817bf14bfd3451f8c78ab7272a5e399710684f1a3c7ef43285622db81bc16c30cd192e1c284b2214819009715d7af97d3bd1a44a5457a43ea2d9b23f06ac87b83b413db35546e122a633a046e4ba143615efa3599000b38dcb3fa6724dc6c72854088439fde876e3a88f1bb11c89f45a5ed06fdd35ecde31d51ca25e12853d38bc38ba3c2b5cca418cf95fea956af7f8bd6af27831c97c2dd57f9fe3ec587ef4db32fcf5deb663f6449bc8ce44afaa1f3b7049f74c5e869b18ccb940fef96b983a56df6712e6fdc046141eb3e62fb8abae47177d20f035ab07085ef244f461b10d8505e841ec0cdcebd8998eed3e667fad496d9183e5971d004a85c4005b4c4bf02db32667b16450b3e491efb1099a0c2919de31b63744fa3ee9b9df3182946cfefb65ddd35bba4c156c84e5525d91ef09d320b164329b153178e7e77032552524d8436a18877c2d582a138e3b3581071d2965dd0dfcebf3a6067121eda4c39491a0dcb257399099b4511f383652f913c60253f6d6e9be16b21fffab2a90ecd0aa23e369a0b675afaafedee15a5558ecfef096de66f20710b2b22842ec0027ba80be114454cbb25bcd77edd727a7f56f0e58e6dc627f23e62175babba39b4cf4d3161299f24d09882082c82c1a144f098323866e15c1b79781dd7115f3a9b67b97b40284ebdb187d9d51a2f7a399221a2a3e9e2d6e23270d66078e8aef9369de80fab9bc0953215c83e91e64734623fc112c9abc4865405688b600dfeedd49f9e9666ca46dacbcff2d543781cb31d584aa0f0061bca93a700f8c284eabf2b7b7ecebb0ac66c575ab8565067b3f8914dde503c763bd16e94891703409e478ac02731e7563016ff5b38e5d3d9b80174bd3fcd0cb19214b7c15fbd2192af9a283c63cf42e7c42090e3fb6cb914d16141e28b24e6e47b154e528795d2dd59de21d942e2cb3fc154824ad5463090cfd00463daa2407b15508597d9b63fa2badeec049779292e4c4e952dc753d26c735a7f5abd532e34876c08ede94d2d0eaa1575cfb423d6fa1cae0034490e1c80680eba263d69a2c344e57f8edfd2d359f50128e0d5d17247cd3c2196b993ec5d33a99b2a54ecae1f60a9fee3a193d3a1d5ddeee6e42f093bd933625966aa779f70f9778e830f0e39d94a357c4078c467c863cf2d319aff7525dc9861658a91b157d026254f15a8d4481fdd285f60b3906d68bc514fb982a3f1511679d2d335fe5138e84d52c41402b9a565a249d483c6f2e14843adf62be9e64eb97818280b76b889b818d995db582bedd1e4fb8aea3118f7b493300f5b886cf2d724d428ceb591b2ef92acd93f9b1e9014bcab3c584e6b1c094d7de76d65eeae5a20ddeaa9c208ea702965bcb8979de730014fc64e841682278afcfc89c93feec7029cd73c43b300a345bebd91bc6a3c200c606b924a416a8a62321f7ee710ff0b3b655db046d3745819c6633e396ebf7b83ad153ea796e878e85bb6164d44304281ecce15944219ce86181d95eb97f5addb763ebcd6d90b60300ac4385df78ad10e515ba73fa15a8471d11a7cd557584c6fc1dd733faa8e6d0f34fbb0686a7b34a666f147fac5cfa77e641965472a46567f0433e133eb25d786e3b88f5bdd857458778c4ca07a0f3a92e49dfc47b47af5ba71daacf535e40436e6959fa1c482478a3a4bf19e29aa4492e36d859659da87613bd8696dfc1c7217c3932f70b21deb4da27f161150e0da491b8999007aa670b9defe7d595c3026d3f9a7a47db72dcb2b6d09da1fe55865b2aee3e004de8e404e986fb59522d4dae5288afd5311ec87f55291fc621a45ce6404729a93389697e5f2a4006b25bf79ef92ef928a8703ed42d78082b1011642353b8aea9dba4c1e2bce1adcc0a9b8e982aaef18b299982af7e992be24e593329e9aac877c437718b0e025b49d40c93a594244624a1179ef0502ae3aa6e845b8c2fce31a1111695fd63939921cd8b4b58cae6d553e7b8d9fbd297bd2512d63b45b96f2f268a51201b5d03cf41403493d6b2ce31322234cdb5a44028b04c65c09534b8e6ae7c0a2aba7ebeddb5f06e6fb944c661177e870545a3570d7baa6e469c4dbd895e81517c03521254508f39823873f47737f4f331e88e2ea7536e215490069d400cf6a8911c6b5573af3179aec96c20a6c4cf7b6b2157767254648116686be68251be326a93232abe0677e23934c4b72d8894292b674cf9c67085595e9deab4b862f71b7043c092504993429a13e31c5532a8c814fbb030a5e8fc14dfda8403bab82eeb9a58b627cdd008f6ff664bd1894987c4eb37e0e916b871199e300bd817dd0098dfc9d1b4155d18f612e137812315e4253ea4ca44fced536a63acea94458ed34296815fce552ac5c8d93be1fce4fe9f377215211787c956f87bb0e86a9b96a7a6fbb9d37b7423692336c3d4efcf7d463f549119832e7a12ead655faeb1ad3bbd3ff0588d16a4c3e07d6b4d7d6b33736e8a7038c16ec6c0182d2a07253c8e27f944ee8f638ee0ab1ae64b30960875c452b2144c11a7f6639d5c2cd00348a0c6277f18eeede0d52163ba282a19edd8d839cf734cb248bc3047596daed6e572522fb24c95734df5c4387c69e2854fc83e5d3ce37145e445eaf199dc4f2212c932b9cccfe82bb8a081639b3226323d8ebc6f75aff71c5edef8455ed4fad758abcc2e1333a5230efc90478b8975f04f0b05d3f7cfd41d9df357e7d197699e63afebab617fb8668010788d1f29e4bdc1f6f2855f2051a8bb0a37408f2c2e6aa74a92600c280177c515bd5ea9e1fd52e8f3cae36975be4c725e2a7c4bf56821690e093dc2cb824b2a4fc055e43a13cd03f029c85cfd19d46067d12694d9013f6fd4d9856f3f4074b582c90a559956eb385eb273d59f3292ef93a607243a97bf00f03667fd59ff3aed06fc4a99f0793614bf88377bd94e5fc59b7b06cc7f0402afd6edeffa049471f751b0c86be0acf64964c2864d3b2e7fe884d7d9827cd2b6861f32c76f0298c97fd75c84dd2f9f598e5ec1e96ce2474f50ffe892ca21f18e585972e319a8e1a3fcc65d1cd9aa206dafe1ee5fe58803428a165126306a35c1777e7dbdd833d8f9213727818910c82ef18e3b6cfb1d83c3c50fd1c58bc869206944e4ab2bea82cd86bb9aa736f4ea46ff3fc7e1138a3693e34d275642aa9e7ed0b8fd57e00791ffa243ed8cab5e6e7a52c8505c280e70d21900856690b3e35725f025523072f9ff260cc45fba52f9cf8f59c750a6c250b9f14bf151c56420ffa4a3d9212f73d7ead42626da73618994b6d5532bb00ff231228e35aaaff89f2ae4507cc727d4676403887bf8711afd70352245775f32c60238b869def27306e884705e0674dd9ebd5f939fbe25b213db2b4b72e5dbbb143fc0e375b8b1ae33e7fa558640025ad028247262a3bbf43bde6630934156655a2a602d0eb689888005e00110a0ac7afb51384e4fcd9012f08489b5dcbaa04a79bc3e0b28c15dd832daf82447083ce3483b6346736cf7b1fb", 0x1000}, {&(0x7f0000001f40)="b45a5a2307e1ac7e01f97c31cb42f2cb6eda15b9ad448f7494b2dfb79bed8ad68cf427b4242e", 0x26}, {&(0x7f0000001f80)="be2314fcd1293d74b2dce075f0e2918da58f7571a54d17768472821bb40a5e4c548ce8bda11922bad379a8", 0x2b}, {&(0x7f0000001fc0)="aeaa9b32c2b608f36f874397edbf4948fea371d0dc83f2a2bf1316f7335a1c80e2e419e4c2c6a21b0d18d8c441a8c7ccd5ae10d2725a6d440fa8b4b7a0a77c8501c2bf77495f3e4636e434a6a76f8d50059bc3f8bc7114cb62b2703edbf50145bd2d0ea92ae448a6828c011950b995c39f792f23c0f7ebb554208d46ebd7527b10f57a97357ec96aa2885edbdaf9e3d61028d14a63c976a4bc7d4bbd3914f9949b9006f0fd32c139b90eb3eb8433993b4cd93e8ee83f53012009cd023cc603a8099adaa39c9e7a23e02b536318fb", 0xce}, {&(0x7f00000020c0)="9ac51d1478126c0793642800ae4565197bb4cfec43ee5f65aa6fcf5fa33a00b456196df995daea32c3a16ff550a16fa5068348762ebde0e103cd5eaf3b1c63f504ec46d04187c1f0c7674a28ec618c368b4ecf3e56d5bbf31c3b256706731d064874cdc81aeb934060494abd40ced490356863cc41e9c60c604f77085998c777528723984903e2205813a77d9243b376df0f25815028ea1f8d637f300d17907a67dd074854c998224ab3d580b4bd1e3f98ea4a20312c0efbcc80bf0cee0f0ae5eee7ee9d1d023ac9a9fc3c7a97597775ce4d845d3ef5c117a15906d178d4dc0953", 0xe1}, {&(0x7f00000021c0)="e42d16c0f9100ebbac570b9eef9533cac7a68494625f335c662630b7c0236224a17e7c34", 0x24}, {&(0x7f0000002200)="5e6a3487aaaf2bd80d9902c749c134afa1bded518edad98d01c63b13a76ee753189b078073204eca69a144019e8c1ec576d6fc1f709e82b414b820884ebc18870cf4cb90f8a8e4a4e2becb5d4ee5ee1c6e57de5bb7f5ae63cc0f80132fa5f91c9b58e905aaabd8e93e6e680309c4d2cc66724b2d21e5c150216d5b0c80e0d9f7b01e28467113854adaa13c726963ef0166c6db4257e7b0eba874f3e8fd368597d94d8df975944ca0cdeb6fd7b1fb0ff47e488092cc4f6b5da48e753cb6af832167887411cda566314ddb64391e", 0xcd}], 0x7, &(0x7f0000002380)=[@rights={{0x1c, 0x1, 0x1, [r2, r0, r1]}}, @cred={{0x1c, 0x1, 0x2, {r22, r23, 0xffffffffffffffff}}}, @rights={{0x14, 0x1, 0x1, [r0]}}, @rights={{0x28, 0x1, 0x1, [r2, r1, r1, r2, r0, r1]}}], 0x80, 0x8015}}], 0x4, 0x1) 00:28:43 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1900}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:43 executing program 0: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xb5000000) 00:28:43 executing program 3: sigaltstack(0x0, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:43 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1a00}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:43 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r0 = socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) sendmsg$unix(r0, &(0x7f0000000700)={&(0x7f0000000000)=@file={0x1, './file0aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'}, 0x6e, &(0x7f0000000640)=[{&(0x7f0000000080)="e788da0e8e3dd3e1214ac8b8fcbc8e8ebcf3ad7f27912de41534fdd91f407accc469d2ae43724dd669090531270031367a9e69608640e01f2a88c17ea7825ca2a7465154956759ed879553a94e24eb7a37cb30abb6dacd5bb3653ec1165f743b895643197860000f2b1f80603b65f80124234a6e341fd099fb458b4e82e67914bdefd352b388aa18fa6ac3", 0x8b}, {&(0x7f0000000140)="430fbec6b6579e10aab2a393ac54d78a7516da14851ff9be9d75abebc0a8cddc2238f5f5ab0016817c042a1714b1cab8651735217eb2e1692c7b6da59570a288c1b78fc64c6614bef5fba93d066ace8e3757bc4b35b47280d9e53d3687247f582f992a520cf9b9ae3c724d8733127cb4db11b99a312872838eec892dbd7b955e97b596dc39e258e3", 0x88}, {&(0x7f0000000200)="55361c3ad3719e092ae24d35e52648f4ad3aea31e5235fe0ea3c8934b506e42142507f11be0489a05c16e644d340b8a592a35700e83e8b3dcde10d3e9fd475b6b4d8251ee711e8ffaaea5b05d8adbf1c8f7718e7988a4c178436a8bf9ddcc6ed47300ddaea407a5345034070d5deb2638a7c351581cd4add7672d68f9aaf1780298dc3d7d67d822757869da570cf687c522f1838e3074c3b6ddf69d3d25fdad50e96690cd2967604d6dcbaedf8c5e9be2a1d360c151781cad8d21deef236ac8c72d930b07766b7a1938501c0a5b2066a29d4c88ba413fd874f7676c68fcf98f2d944e1bf7fa62029dc8fef7c85", 0xed}, {&(0x7f0000000300)="e828d896f715c6d15809cd113f47a5ef8aae158e3739770396bd32d3b9b85adee260f09312fedf2203f6008415dc82e21a5113ea54d98e2daf62a05a9205fdbddd9eed4e8ca64ce78a8cad401297c5e14a8cac05df73251418dc704a1180f5c07f6d", 0x62}, {&(0x7f0000000380)="76cedcac2912ec32bdc2f31969975861654fb5e932791b0d7446d0f278580d0c20ba19bda58a8050f5083b240823b8712474902fcb083ee88837add4d2dce7c351b54e58023c7872cb82cdba144678ac1e91a37f8657cad4e18bf9566fd2cf04d4bb994e4510d9bd92c12e4a799f1d8529d5719e2a85a6980a3017509dd2b8f386247d58ea24126b585599167dac5200bbed120d6f703463047a83b38112bde1b79b16d4dda670f98afeda76", 0xac}, {&(0x7f00000008c0)="fc0f4ae8500aea363dbbb44a85fb519d50efc83e50fdd1f1c9b56197e54c2f7eca6abb2a58b769805efa19b9693fd16b19ed04e039a81b4e97045284f1f01aa2299e60e719b2ff31ceff21f29fc6220b9d940a024dfb3553147c6d4bf69ba8a7534b7ff8bef95ff8127ab1702d049a7c7cc04639f272672f21f2958820728a161196c1172fa3ea5dc1099b9bdf7c090bf7fc6242713c09a7125e3eeb86f9f1c3f9ac505f39b308152631e81145d2fbdf02f83516d92746f4a79bcfd32a634abd0f7f28009d8fb5e6b870f54baa7f8e21eacf074ce374b92056b08b59a41375a83676ae7b03f68f94e1d3d4a8c92fb68f5a4df65af1774343943dda891a1c8ee55458bc72f28fca9e36e24828f7ada607a87ed922ba6df9dc6cea472624b5abc80407905bc352bd77a3cf208b6cadd8d492c8f20ba1ce7df84ff7edf3fd4bfe50113916e1702825a12159663aadc98712b7f070b9f0e6c43e7974eeaba0b45459441498a0bf7fceef0c6e5a50ace72ebaa6326d83efbe7e74d4955dccc140eeee560aedf59b92ccf9341faf73260ec016bd5e7d4ba4dde3fdf586eae5278b82e87f1d107a1ee1871b28081ebe595adb9d3ac788fc894ba03ad2e925027adc2b5c4451ae1f4953435368ed6bc24c837056deca98e5a6270f05dcc0ab1a53ce3700455f603718ecb3a1a88d5e5f2c80fa39faebc4067e406f025f64a18ebc0fa763aeebf075e1484117241f3731f8edec92b90ff10ab703698730f71193a297cb2c317eef784d3f1492e79d0198c4f76d9e4e2885e846e86529cff869370918ffd88f80da007a1202e1fdca7aa0751008e008ad50ebbe16991e2b1f33dcd4a05a842102d8c0c4db1d02f92dbf73a7f152a5490c53e89d917097a56e053f9cdc66f925625dc53a583edd6fed4ec70b92ac08dc5aeb269eb7e063fe86a4a06441b0eab41e4597fea30fb2fc177889ac4a961f7d891f5e7554359b989653155a01022ee9b9bf46f6e843b3b5845952061558a71d8b8ba091b0826545fe768c34d69982ec9fe2c8899185d10f775b8a975dd92efca8a2d1732194c5ffe779e4591d44f3e54a4afeba865595344876e661b8eeaa5be401ca9d273de8b2f4f3629442aafc11e1108f996ce921f82643bc1331b8b6fc39cad3ed065f6a4372f6a6ab399fa415fca488dfd67c46fda3a03266c0a85a2685dbcc77bd55971fd85495bd03f55c38c41f427a88b5cf0174482a8a7ba80202fb25922ebdb88aed4154e32852b956acd968173b5b081367c5f972ed74b0f9c1176dfe93b7518ed45ce14c120ae3be42f23b5b4fc35902bd64c1526aeb5368da25d7367dc2e18927eac3193e01a3da341629dedd44434109f3e6fffc88bba961563caf05755d184813501b93f503afb20b8027325eaab784f5729c9af5ed57e06aea86a66d8e7b57f95d83161946584b0d22c1b2fb3dbcce1ea5d23f34bb7ebb6e47d44a5ac1287e594af1c38c1533fc0a034a1de07ba6d61db73fe5059988c7a9c7b6b6a972c6e2a441978a4b31fe155c70e2c0a0a808f8a2d22326a672998fd90323c94ed8dcab9e325d420e23564617fbbc15b8d47ab657094d83a995ae6c29cfd645e410e6c71208ebc155a7da80f27373d242509f27eed918da96f1087f31c2afffa085d80cd1fd038aed4d40799b17e71196994e594ffc565d1aeaec6b92def971e9e28522f14195f2c46f756bb3e59b119ac5dbb6e65d6a0b91a6ede787fec7704edce8af0f3283db6282e00bbe1b77084353cc0697da5cad40fddf099d321fdf13c2a83d2996d6de4bdfa2762601e82ab5208b92d08309cdac39c3e26b6b9815e896d5522133ffd34263a4387bb43ebc8e0e77acd7ac51061c3432be87ab95fba045c1a44caaeb157f878628a97150483c7a813d3e2fd8f4647a2ea6b82de3045465a8ef71862b406b9958e4dad8bda91fe30d242185931531b87f3b47e198e452a1866faab99e902ea5a6907cb969432f05080c0acf1b1c7a0ba3134bc6607932ad94121e9484527a3a060524baa2857be7827cad0364c3334773c654688fa5672db827ade5ba5a527722b08ad4598127b8e4f96b1ccb119bc6727191461359ca25b21c0bca86bdc5e753c6509e3a23786244aea98120f78c5b43f879bdaf2f59d22829e7e4ec6791fe40d245f881e33816800cff7b00c232b4bac74d5d7eab0c3e477555e9c20b71866dd90c4949dc5829a0c8392130888eebc4fa752abab419a6568afdf3404609cbe63d3fafdef4a557cab617e944e002d768f0123018bdd7b7d27b9bc145f13abdc53bc979e0de421033c69443c660fbe955c62bae22186b4761666e1348202a64b4893e92ff785ff877a426e99fa329b66c5057337013b5f6e7ac4da159664d792100832e8600223aa536e441bb2bc514cae0084784f05e579221ae0b4672e68a8d8bc1117d477ae1e40f30fb6863f001e34c80611a3d920d80f60256dd21b8754764e4041855aa694da1f524a51caecd678589ff9ddaa8516f15abac0c0d45ccd5df1061cf45c8a82b78263e4107517d6ad2b03f6a6ac5be38714e956ec431a71f6dace96f1970e7f52dd3bff553b15e6ef5d8d3b997b941d8affbbcfc920a6ff78316ed32fe07e6acee645aa18044e9d4836f5e8ef1b74e95f34dedc790c0a44d8ff1fb5acedd624aa2fea58ff12bfe9f31a7618ec875186355fa941608c3a022d07b696bb026e2f43913bf1964545d233c34d48ea0bd209dc44e11fe1f2bb0de16c2a87bac8e373325dc2696bde07b20da2e6733da66cb2917bd108226c51df8a4b8c6f1a3d46b7bb4a64220772b65c5d56c59f5cfe900508ec9ce8b960997926005eaa44a429a1a8bc61d73582d5c88932a457e4b033e7b6b367a771f57b5d28b3f2b8419c22bcd1c190f703f54b9932a479f06367381cfd95de098947f742ef5a709c19223aab7a48d1aa6098ba319b0b4bf43393e230401f75a7dba3b0425be9312f732fafdef82311bb489ee4d687141c61c8cd0301094d351c1327f84bb1a7cf142c708375b946e4daffa457b6c9bec70a6660c25d79f8cc1b0017899d8ab7bf03ee7f216e09c762a1ca5f4889b98fa351d85567d6cf12c5306232f3a9b62eb13c834f787d7b4ad93f201f694a0f146da66526460b2cfe72118d07ba6b7e9530e4675eb9a92be5a83707e101ae32453dbde5ceb35d2eac7ef624bdbdfef04686cd155c861a577776972b54d75e7e99376fd5a402947682d1f84c1d14f5c0bc2b045a086a97460adf1a63c342196dbd3c41ce270544c7c275f8156c9e244543bc5e78017579000d4c643899563c1ae57c246539fa4a1bafb6b3c5f7ed308fdb5cd15eda28bc4fa269c762b87a289afbd31eda08469129a2e3d28f955c42c78d2100ce534fca836d2145813afea576c59e75c078540a9a92adc12e1a265e918855c8a174e0252ab5f8114625fe78eb07f3588f728cac6dc5873bd2715ec0bb1852d33256b036bd14ccdd44b0d04be0bc17a0580828d7a24a17524b3a6442392eca1a844b5ae03b174761343ce1ad85e4ae1d7bb8d4eef8c52718c341e33ea5e8dbf8b053fdf6fe5a9e153b6a9ff1c24d8f70c8f756525f38876889d652c8baa920f228b91744f238ec3dd653df1298a73f58bac6bb0c0317030186b592cc64fa1c467f04eec09f3bf48d6adbc8eb7b1ae3195d08168e65acf5f5c4c2028c3c798fb0c89edba2f67d76ec469bb076733423318182b1ef7fa759dbccee15a2ec9db424aafcb52862baa9f7d7db3e9ee7417d9467f6604d505166ba9d787dedffe761a797fb13de536c4d0e69d85bb159dd8c895c93b52044905dd3c4f6685a522c3776782f4174df7f80cfcac747d5d4d77a1b904557b973e220d0053cfc91e3bfc03420db7de22e972516f692f12e010039308b5698c4e8f7224788cdca6c6534f88c8c9d16eea65b66f1780bcb9cbe593ab44e282acd19c39af0583f144620837a12b04174d19d2867b777f32061ed284f86287ec44433645a694580cf4db7d99b96a7646fd6e1bb5e1ae29853c501a2a8021522620d6d1e22194b28cac92201e80865d3126df5dc34115aac20711573d96564db335225d865c3096b9181329e488d39ec19f7fa5ec1bd9f0ef7736018df725bb0418cead8b2cf5465ff83e2a39fe5a5ed7f704a9cc5af43f3a488c36c0bc6d192452f7e0a2002f65718db8cb6aedb53855ff4d2ae67c09c3f1f0356354b117ecec06b27b3521065270a1f52f5d8486ed00feb26a537a20b03d27357823580bc74e60874cc6843211ce182b656d8dab56b8bb38ae118686d8736b574588d77ff094974b43d974962c6cc86b5012c7d1b0e6ee7d1d690dfa1288b05023823c21ae929271bb48aa2f1ec455980b732d3d1735eff6b7ad58e96273e453255e7174f8c4368ac18c4115cc5c6f9aded9b6856132950680b236d83db1a9d06b172b0ffcc3571a68d34cdf7ad74eb2a4676022e8c20258b02064c26d87e2a3861e6f353f4e81d1f0d50586321db6c6d7c6f5e874a80d0b8df661d9b2d517aaec5ccecfdbc7bd51d30929a71467fe102532cd455bd353bacf4f6669eb6cd8c0fbfe8306e0574cfac2c8a24aea1e1fba10a69e61e33e72325d1ab29eb93b5e160d08a403ecf2d504c9d5fd52a88e1d25ab939dd8bd0533974c1dac22b7cce4b6032d0256e6a9bd053d8b4dd6ed5d198d1bbdf05a7408e18f434fd5200baaca15e3f2cbcb272c7e2d9f2b6a61cadba7ec4c27c449d88cb47be40cb8145e31a4669b9f7be3a6a4847f69dc25e5c6093a37a2ae273a4666af9c4131a4c2d28628542e8a1d99080d04313f30b7bac98213b1080c4230264adc2aa1e44bcb1180f521379dc9e70e2c05399e56b4bdce62635f3c3f89e57b3dbd9d86d1832e0c6f6ed0f158e8480e947fdfeeaf8b14665f6837c86ce32c0102a71a14151ed3691c79554dbb881af990f5b0402b7ffc23037f01a484d222c780d6f2e6dda2923a6254214808cc5a65fc0d20bd8910f8feb179894f0c6203dbe3f67fd023ca3c8df11b54e56800b6fc10fc13394c4c11944485af20f0e23babe6f6e7ebba9411c7de927ce1ccd4dc467bba5dc464552790ac9ff9736779b9cb25dee8811a4c1a9ff1101f0fa5338c9739140142683fc9e2775c09d3284ad29b62dc3390f062fd58316180bf792e18cfdeb1f275b5c7d12f9af17967f55bce7ff68a229b4d2bc9427ad23734452ef1c2b8bb98724b146a8f6718009464dae11cfd23bde6f7401730b3b688beadb481aab1c87f29c8e7508b7b0a4caaa7dd497aa39bff912487c8c60f30d34ff47cdfef581c0fee60f599aab8cf1dbdadde70dcfbe9a6d1a4eded8406a6062014b911704952c6c7eb64d4aff0292ce73df0400273089b161f63c3ce155c66e51edb3eca9835ca89f9cbdfa48c014086b3ae5c8a50cacb63a0b6387e6c0581ada4f7c392f8e0b8b08d047f30e798a322802db81bb8ae177876020f63efc3ad02186b66aba129a7be7e6df4a8bc17bcb2ac7ef191cbd0601eee8bf696c62d6036d67f317b58681abedec64f5b71be7c3dafd4a220b16df5424ad6d37bf9b0d46b32c182fafd242a4d05aef5a5e9661f5ff98f077d7734a0a52a5f729b33ceb49f10488186a17c8c990cdc06216bea5837e6e8fbe85a4e09b0ce8ab9d6b4d038d7c841e6567fd5db2f49de514bdac138e09224a329cc82bdf43a2c0719a0d906bb8694ec9de23488fc8661bc50e766e345a9f9354f2ee6d8974b1a70441dc11a84cf85fed58152f64ba133c922374d24", 0x1000}, {&(0x7f0000000440)="dd4072b34ae0a8cc35fc1de24ffc0fcda7ed7f010bf7e39cd93bfd6ea9067fc6be4468eac79652203fe18508e0abeeda93b428d96ba1896f00c1dd7812b9651b464aa221f9cb89bbfae6", 0x4a}, {&(0x7f00000004c0)="24b3637b028edf0aac6b001f3b1c754e5019112b18724053d453f08e2deeeea049118c1a1ddb0e3259987267801bf3233be4ec8da8d9e6a5449a1f4ee7b50975c59385264091797608e4f442eaf83ca2ddeae3e3e462d40d7cfc09353882ebaad265feb7d99a4cc3bf789fba53c82d382e3396c1dce1a171285a1321cb1e12f57d0e83e75ce564f50908a5bb6e32a06e6b9568b1e9c19f1b7f3d9e926f07702490fa2ba4e718ad10a17f9ca6ec4e6a6cc2621800a27081e41bd26afad8520a9d184fa83a355d81a2abd54177edcb3b9ffdda80aaa2d726d3b0e5d121dfe2048ebec0913708962fb2b3fa27d0300bb1f01d9c2fdd06bcc4ef8b044c3e16", 0xfd}, {&(0x7f00000005c0)="3195a4e6de7f6dabf297685f86e80498b0d210edcbcd5942aa62a37e8c419e22f247dca423f92af1e1d9d22075c691112077caa7d3ccf387a85ca7d65643c8f53ca64b982ed180a423b18fc80bff71bc22438517f211f1986594db8b0817", 0x5e}], 0x9}, 0x840) getpid() getpid() getuid() getuid() sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) (async) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) (async) socket$unix(0x1, 0x5, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) sendmsg$unix(r0, &(0x7f0000000700)={&(0x7f0000000000)=@file={0x1, './file0aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa'}, 0x6e, &(0x7f0000000640)=[{&(0x7f0000000080)="e788da0e8e3dd3e1214ac8b8fcbc8e8ebcf3ad7f27912de41534fdd91f407accc469d2ae43724dd669090531270031367a9e69608640e01f2a88c17ea7825ca2a7465154956759ed879553a94e24eb7a37cb30abb6dacd5bb3653ec1165f743b895643197860000f2b1f80603b65f80124234a6e341fd099fb458b4e82e67914bdefd352b388aa18fa6ac3", 0x8b}, {&(0x7f0000000140)="430fbec6b6579e10aab2a393ac54d78a7516da14851ff9be9d75abebc0a8cddc2238f5f5ab0016817c042a1714b1cab8651735217eb2e1692c7b6da59570a288c1b78fc64c6614bef5fba93d066ace8e3757bc4b35b47280d9e53d3687247f582f992a520cf9b9ae3c724d8733127cb4db11b99a312872838eec892dbd7b955e97b596dc39e258e3", 0x88}, {&(0x7f0000000200)="55361c3ad3719e092ae24d35e52648f4ad3aea31e5235fe0ea3c8934b506e42142507f11be0489a05c16e644d340b8a592a35700e83e8b3dcde10d3e9fd475b6b4d8251ee711e8ffaaea5b05d8adbf1c8f7718e7988a4c178436a8bf9ddcc6ed47300ddaea407a5345034070d5deb2638a7c351581cd4add7672d68f9aaf1780298dc3d7d67d822757869da570cf687c522f1838e3074c3b6ddf69d3d25fdad50e96690cd2967604d6dcbaedf8c5e9be2a1d360c151781cad8d21deef236ac8c72d930b07766b7a1938501c0a5b2066a29d4c88ba413fd874f7676c68fcf98f2d944e1bf7fa62029dc8fef7c85", 0xed}, {&(0x7f0000000300)="e828d896f715c6d15809cd113f47a5ef8aae158e3739770396bd32d3b9b85adee260f09312fedf2203f6008415dc82e21a5113ea54d98e2daf62a05a9205fdbddd9eed4e8ca64ce78a8cad401297c5e14a8cac05df73251418dc704a1180f5c07f6d", 0x62}, {&(0x7f0000000380)="76cedcac2912ec32bdc2f31969975861654fb5e932791b0d7446d0f278580d0c20ba19bda58a8050f5083b240823b8712474902fcb083ee88837add4d2dce7c351b54e58023c7872cb82cdba144678ac1e91a37f8657cad4e18bf9566fd2cf04d4bb994e4510d9bd92c12e4a799f1d8529d5719e2a85a6980a3017509dd2b8f386247d58ea24126b585599167dac5200bbed120d6f703463047a83b38112bde1b79b16d4dda670f98afeda76", 0xac}, {&(0x7f00000008c0)="fc0f4ae8500aea363dbbb44a85fb519d50efc83e50fdd1f1c9b56197e54c2f7eca6abb2a58b769805efa19b9693fd16b19ed04e039a81b4e97045284f1f01aa2299e60e719b2ff31ceff21f29fc6220b9d940a024dfb3553147c6d4bf69ba8a7534b7ff8bef95ff8127ab1702d049a7c7cc04639f272672f21f2958820728a161196c1172fa3ea5dc1099b9bdf7c090bf7fc6242713c09a7125e3eeb86f9f1c3f9ac505f39b308152631e81145d2fbdf02f83516d92746f4a79bcfd32a634abd0f7f28009d8fb5e6b870f54baa7f8e21eacf074ce374b92056b08b59a41375a83676ae7b03f68f94e1d3d4a8c92fb68f5a4df65af1774343943dda891a1c8ee55458bc72f28fca9e36e24828f7ada607a87ed922ba6df9dc6cea472624b5abc80407905bc352bd77a3cf208b6cadd8d492c8f20ba1ce7df84ff7edf3fd4bfe50113916e1702825a12159663aadc98712b7f070b9f0e6c43e7974eeaba0b45459441498a0bf7fceef0c6e5a50ace72ebaa6326d83efbe7e74d4955dccc140eeee560aedf59b92ccf9341faf73260ec016bd5e7d4ba4dde3fdf586eae5278b82e87f1d107a1ee1871b28081ebe595adb9d3ac788fc894ba03ad2e925027adc2b5c4451ae1f4953435368ed6bc24c837056deca98e5a6270f05dcc0ab1a53ce3700455f603718ecb3a1a88d5e5f2c80fa39faebc4067e406f025f64a18ebc0fa763aeebf075e1484117241f3731f8edec92b90ff10ab703698730f71193a297cb2c317eef784d3f1492e79d0198c4f76d9e4e2885e846e86529cff869370918ffd88f80da007a1202e1fdca7aa0751008e008ad50ebbe16991e2b1f33dcd4a05a842102d8c0c4db1d02f92dbf73a7f152a5490c53e89d917097a56e053f9cdc66f925625dc53a583edd6fed4ec70b92ac08dc5aeb269eb7e063fe86a4a06441b0eab41e4597fea30fb2fc177889ac4a961f7d891f5e7554359b989653155a01022ee9b9bf46f6e843b3b5845952061558a71d8b8ba091b0826545fe768c34d69982ec9fe2c8899185d10f775b8a975dd92efca8a2d1732194c5ffe779e4591d44f3e54a4afeba865595344876e661b8eeaa5be401ca9d273de8b2f4f3629442aafc11e1108f996ce921f82643bc1331b8b6fc39cad3ed065f6a4372f6a6ab399fa415fca488dfd67c46fda3a03266c0a85a2685dbcc77bd55971fd85495bd03f55c38c41f427a88b5cf0174482a8a7ba80202fb25922ebdb88aed4154e32852b956acd968173b5b081367c5f972ed74b0f9c1176dfe93b7518ed45ce14c120ae3be42f23b5b4fc35902bd64c1526aeb5368da25d7367dc2e18927eac3193e01a3da341629dedd44434109f3e6fffc88bba961563caf05755d184813501b93f503afb20b8027325eaab784f5729c9af5ed57e06aea86a66d8e7b57f95d83161946584b0d22c1b2fb3dbcce1ea5d23f34bb7ebb6e47d44a5ac1287e594af1c38c1533fc0a034a1de07ba6d61db73fe5059988c7a9c7b6b6a972c6e2a441978a4b31fe155c70e2c0a0a808f8a2d22326a672998fd90323c94ed8dcab9e325d420e23564617fbbc15b8d47ab657094d83a995ae6c29cfd645e410e6c71208ebc155a7da80f27373d242509f27eed918da96f1087f31c2afffa085d80cd1fd038aed4d40799b17e71196994e594ffc565d1aeaec6b92def971e9e28522f14195f2c46f756bb3e59b119ac5dbb6e65d6a0b91a6ede787fec7704edce8af0f3283db6282e00bbe1b77084353cc0697da5cad40fddf099d321fdf13c2a83d2996d6de4bdfa2762601e82ab5208b92d08309cdac39c3e26b6b9815e896d5522133ffd34263a4387bb43ebc8e0e77acd7ac51061c3432be87ab95fba045c1a44caaeb157f878628a97150483c7a813d3e2fd8f4647a2ea6b82de3045465a8ef71862b406b9958e4dad8bda91fe30d242185931531b87f3b47e198e452a1866faab99e902ea5a6907cb969432f05080c0acf1b1c7a0ba3134bc6607932ad94121e9484527a3a060524baa2857be7827cad0364c3334773c654688fa5672db827ade5ba5a527722b08ad4598127b8e4f96b1ccb119bc6727191461359ca25b21c0bca86bdc5e753c6509e3a23786244aea98120f78c5b43f879bdaf2f59d22829e7e4ec6791fe40d245f881e33816800cff7b00c232b4bac74d5d7eab0c3e477555e9c20b71866dd90c4949dc5829a0c8392130888eebc4fa752abab419a6568afdf3404609cbe63d3fafdef4a557cab617e944e002d768f0123018bdd7b7d27b9bc145f13abdc53bc979e0de421033c69443c660fbe955c62bae22186b4761666e1348202a64b4893e92ff785ff877a426e99fa329b66c5057337013b5f6e7ac4da159664d792100832e8600223aa536e441bb2bc514cae0084784f05e579221ae0b4672e68a8d8bc1117d477ae1e40f30fb6863f001e34c80611a3d920d80f60256dd21b8754764e4041855aa694da1f524a51caecd678589ff9ddaa8516f15abac0c0d45ccd5df1061cf45c8a82b78263e4107517d6ad2b03f6a6ac5be38714e956ec431a71f6dace96f1970e7f52dd3bff553b15e6ef5d8d3b997b941d8affbbcfc920a6ff78316ed32fe07e6acee645aa18044e9d4836f5e8ef1b74e95f34dedc790c0a44d8ff1fb5acedd624aa2fea58ff12bfe9f31a7618ec875186355fa941608c3a022d07b696bb026e2f43913bf1964545d233c34d48ea0bd209dc44e11fe1f2bb0de16c2a87bac8e373325dc2696bde07b20da2e6733da66cb2917bd108226c51df8a4b8c6f1a3d46b7bb4a64220772b65c5d56c59f5cfe900508ec9ce8b960997926005eaa44a429a1a8bc61d73582d5c88932a457e4b033e7b6b367a771f57b5d28b3f2b8419c22bcd1c190f703f54b9932a479f06367381cfd95de098947f742ef5a709c19223aab7a48d1aa6098ba319b0b4bf43393e230401f75a7dba3b0425be9312f732fafdef82311bb489ee4d687141c61c8cd0301094d351c1327f84bb1a7cf142c708375b946e4daffa457b6c9bec70a6660c25d79f8cc1b0017899d8ab7bf03ee7f216e09c762a1ca5f4889b98fa351d85567d6cf12c5306232f3a9b62eb13c834f787d7b4ad93f201f694a0f146da66526460b2cfe72118d07ba6b7e9530e4675eb9a92be5a83707e101ae32453dbde5ceb35d2eac7ef624bdbdfef04686cd155c861a577776972b54d75e7e99376fd5a402947682d1f84c1d14f5c0bc2b045a086a97460adf1a63c342196dbd3c41ce270544c7c275f8156c9e244543bc5e78017579000d4c643899563c1ae57c246539fa4a1bafb6b3c5f7ed308fdb5cd15eda28bc4fa269c762b87a289afbd31eda08469129a2e3d28f955c42c78d2100ce534fca836d2145813afea576c59e75c078540a9a92adc12e1a265e918855c8a174e0252ab5f8114625fe78eb07f3588f728cac6dc5873bd2715ec0bb1852d33256b036bd14ccdd44b0d04be0bc17a0580828d7a24a17524b3a6442392eca1a844b5ae03b174761343ce1ad85e4ae1d7bb8d4eef8c52718c341e33ea5e8dbf8b053fdf6fe5a9e153b6a9ff1c24d8f70c8f756525f38876889d652c8baa920f228b91744f238ec3dd653df1298a73f58bac6bb0c0317030186b592cc64fa1c467f04eec09f3bf48d6adbc8eb7b1ae3195d08168e65acf5f5c4c2028c3c798fb0c89edba2f67d76ec469bb076733423318182b1ef7fa759dbccee15a2ec9db424aafcb52862baa9f7d7db3e9ee7417d9467f6604d505166ba9d787dedffe761a797fb13de536c4d0e69d85bb159dd8c895c93b52044905dd3c4f6685a522c3776782f4174df7f80cfcac747d5d4d77a1b904557b973e220d0053cfc91e3bfc03420db7de22e972516f692f12e010039308b5698c4e8f7224788cdca6c6534f88c8c9d16eea65b66f1780bcb9cbe593ab44e282acd19c39af0583f144620837a12b04174d19d2867b777f32061ed284f86287ec44433645a694580cf4db7d99b96a7646fd6e1bb5e1ae29853c501a2a8021522620d6d1e22194b28cac92201e80865d3126df5dc34115aac20711573d96564db335225d865c3096b9181329e488d39ec19f7fa5ec1bd9f0ef7736018df725bb0418cead8b2cf5465ff83e2a39fe5a5ed7f704a9cc5af43f3a488c36c0bc6d192452f7e0a2002f65718db8cb6aedb53855ff4d2ae67c09c3f1f0356354b117ecec06b27b3521065270a1f52f5d8486ed00feb26a537a20b03d27357823580bc74e60874cc6843211ce182b656d8dab56b8bb38ae118686d8736b574588d77ff094974b43d974962c6cc86b5012c7d1b0e6ee7d1d690dfa1288b05023823c21ae929271bb48aa2f1ec455980b732d3d1735eff6b7ad58e96273e453255e7174f8c4368ac18c4115cc5c6f9aded9b6856132950680b236d83db1a9d06b172b0ffcc3571a68d34cdf7ad74eb2a4676022e8c20258b02064c26d87e2a3861e6f353f4e81d1f0d50586321db6c6d7c6f5e874a80d0b8df661d9b2d517aaec5ccecfdbc7bd51d30929a71467fe102532cd455bd353bacf4f6669eb6cd8c0fbfe8306e0574cfac2c8a24aea1e1fba10a69e61e33e72325d1ab29eb93b5e160d08a403ecf2d504c9d5fd52a88e1d25ab939dd8bd0533974c1dac22b7cce4b6032d0256e6a9bd053d8b4dd6ed5d198d1bbdf05a7408e18f434fd5200baaca15e3f2cbcb272c7e2d9f2b6a61cadba7ec4c27c449d88cb47be40cb8145e31a4669b9f7be3a6a4847f69dc25e5c6093a37a2ae273a4666af9c4131a4c2d28628542e8a1d99080d04313f30b7bac98213b1080c4230264adc2aa1e44bcb1180f521379dc9e70e2c05399e56b4bdce62635f3c3f89e57b3dbd9d86d1832e0c6f6ed0f158e8480e947fdfeeaf8b14665f6837c86ce32c0102a71a14151ed3691c79554dbb881af990f5b0402b7ffc23037f01a484d222c780d6f2e6dda2923a6254214808cc5a65fc0d20bd8910f8feb179894f0c6203dbe3f67fd023ca3c8df11b54e56800b6fc10fc13394c4c11944485af20f0e23babe6f6e7ebba9411c7de927ce1ccd4dc467bba5dc464552790ac9ff9736779b9cb25dee8811a4c1a9ff1101f0fa5338c9739140142683fc9e2775c09d3284ad29b62dc3390f062fd58316180bf792e18cfdeb1f275b5c7d12f9af17967f55bce7ff68a229b4d2bc9427ad23734452ef1c2b8bb98724b146a8f6718009464dae11cfd23bde6f7401730b3b688beadb481aab1c87f29c8e7508b7b0a4caaa7dd497aa39bff912487c8c60f30d34ff47cdfef581c0fee60f599aab8cf1dbdadde70dcfbe9a6d1a4eded8406a6062014b911704952c6c7eb64d4aff0292ce73df0400273089b161f63c3ce155c66e51edb3eca9835ca89f9cbdfa48c014086b3ae5c8a50cacb63a0b6387e6c0581ada4f7c392f8e0b8b08d047f30e798a322802db81bb8ae177876020f63efc3ad02186b66aba129a7be7e6df4a8bc17bcb2ac7ef191cbd0601eee8bf696c62d6036d67f317b58681abedec64f5b71be7c3dafd4a220b16df5424ad6d37bf9b0d46b32c182fafd242a4d05aef5a5e9661f5ff98f077d7734a0a52a5f729b33ceb49f10488186a17c8c990cdc06216bea5837e6e8fbe85a4e09b0ce8ab9d6b4d038d7c841e6567fd5db2f49de514bdac138e09224a329cc82bdf43a2c0719a0d906bb8694ec9de23488fc8661bc50e766e345a9f9354f2ee6d8974b1a70441dc11a84cf85fed58152f64ba133c922374d24", 0x1000}, {&(0x7f0000000440)="dd4072b34ae0a8cc35fc1de24ffc0fcda7ed7f010bf7e39cd93bfd6ea9067fc6be4468eac79652203fe18508e0abeeda93b428d96ba1896f00c1dd7812b9651b464aa221f9cb89bbfae6", 0x4a}, {&(0x7f00000004c0)="24b3637b028edf0aac6b001f3b1c754e5019112b18724053d453f08e2deeeea049118c1a1ddb0e3259987267801bf3233be4ec8da8d9e6a5449a1f4ee7b50975c59385264091797608e4f442eaf83ca2ddeae3e3e462d40d7cfc09353882ebaad265feb7d99a4cc3bf789fba53c82d382e3396c1dce1a171285a1321cb1e12f57d0e83e75ce564f50908a5bb6e32a06e6b9568b1e9c19f1b7f3d9e926f07702490fa2ba4e718ad10a17f9ca6ec4e6a6cc2621800a27081e41bd26afad8520a9d184fa83a355d81a2abd54177edcb3b9ffdda80aaa2d726d3b0e5d121dfe2048ebec0913708962fb2b3fa27d0300bb1f01d9c2fdd06bcc4ef8b044c3e16", 0xfd}, {&(0x7f00000005c0)="3195a4e6de7f6dabf297685f86e80498b0d210edcbcd5942aa62a37e8c419e22f247dca423f92af1e1d9d22075c691112077caa7d3ccf387a85ca7d65643c8f53ca64b982ed180a423b18fc80bff71bc22438517f211f1986594db8b0817", 0x5e}], 0x9}, 0x840) (async) getpid() (async) getpid() (async) getuid() (async) getuid() (async) 00:28:43 executing program 2: sigaltstack(0x0, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:43 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1b00}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:43 executing program 2: sigaltstack(0x0, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:43 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1c00}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:43 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xcb160000) 00:28:44 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r1 = socket$unix(0x1, 0x5, 0x0) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = getpid() r4 = getpid() r5 = getuid() bind(0xffffffffffffffff, &(0x7f0000000000)=@generic={0x10, "8649ed818f7eec8299e10dea59eba6110e208a14a2cd2563e0f9f7bc9708a0e37e1c7f399d59e907410b5881d2b4996db8a4311afd451d76363092457b74a169d2036eb14b48b88ba7bcf1a8c44a8a7dfcda1cfce550ef256472e42db86d6f17932e3cfcc598dd7b3f75cb3716bfe5167219a45581662f14ee1db7098766"}, 0x80) r6 = syz_init_net_socket$bt_sco(0x1f, 0x5, 0x2) r7 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000000680)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x9, '\x00', 0x0, 0xffffffffffffffff, 0x1, 0x4, 0x3}, 0x48) r8 = bpf$MAP_CREATE_TAIL_CALL(0x0, &(0x7f0000000700)={0x3, 0x4, 0x4, 0xa, 0x0, 0xffffffffffffffff, 0x3, '\x00', 0x0, 0xffffffffffffffff, 0x3, 0x1, 0x4}, 0x48) r9 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000780)='./binderfs2/binder-control\x00', 0x1000, 0x0) r10 = geteuid() r11 = socket$inet6_mptcp(0xa, 0x1, 0x106) getsockopt$inet6_mptcp_buf(r11, 0x11c, 0x4, 0x0, 0x0) socketpair$unix(0x1, 0x5, 0x0, &(0x7f0000001d40)={0xffffffffffffffff, 0xffffffffffffffff}) r13 = syz_init_net_socket$x25(0x9, 0x5, 0x0) bind$x25(r13, &(0x7f00000007c0), 0x12) r14 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000001d80)='./binderfs2/binder-control\x00', 0x0, 0x0) r15 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000001dc0)='/proc/sys/net/ipv4/vs/sync_threshold\x00', 0x2, 0x0) r16 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000001e00)={0x2, 0x4, 0x8, 0x1, 0x80, 0xffffffffffffffff, 0x2, '\x00', 0x0, 0xffffffffffffffff, 0x2, 0x2, 0x5}, 0x48) r17 = openat$loop_ctrl(0xffffffffffffff9c, &(0x7f0000002200), 0x20c0, 0x0) r18 = fsopen(&(0x7f0000000140)='sysv\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r18, 0x6, 0x0, 0x0, 0x0) r19 = getpid() r20 = bpf$MAP_CREATE_RINGBUF(0x0, &(0x7f0000002240)={0x1b, 0x0, 0x0, 0x9, 0x0, 0x1, 0x0, '\x00', 0x0, 0xffffffffffffffff, 0x3, 0x5, 0x5}, 0x48) r21 = socket$inet_tcp(0x2, 0x1, 0x0) getsockopt$inet_pktinfo(r21, 0x0, 0x8, 0x0, &(0x7f0000000240)) sendmmsg$unix(0xffffffffffffffff, &(0x7f0000002400)=[{{&(0x7f0000000000)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000180)=[{&(0x7f0000000080)="76d750b80a323d71ee0d15005f90850a1f697b41ef3de6407fb317ee53664e39275c43bffd5a5fa70b79f513fe35215ede2c38735abd87423404486f80784bb02dfa24b6d49a8beb2f3df2d106b81b76ddd7dc5a18c7d2f62691c71e8cd75f225911a1220945c54937b8f5e165ef6fc7fece6a6beef12635bfa39a045f59b30f91bfcd6f75634216267c72e55892c94d53bfcc657d693700a87e7afd6dcdd85126473cdf7cfaeba4a2b906210d0ec0c0430218b4a57ea63d00fb0a7ed0ac379510185d69b3cf55815b58ed953a50ca4c2ddfac758c1cfc966f86be87a5d11b9d0366b317a87d66a98c8a8834d17da0", 0xef}], 0x1, &(0x7f0000000200)=[@cred={{0x1c, 0x1, 0x2, {0x0, r5, 0xffffffffffffffff}}}], 0x20, 0x4000000}}, {{&(0x7f0000000240)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000440)=[{&(0x7f00000002c0)="dfc42b6297fb512773cb3531917c42e904320d0a31cfad4bf9e1d67de68b3b88f5b575cac5255121e9740b40", 0x2c}, {&(0x7f0000000300)="e68f09fab3244eb76bad247e82934b8db9353d8c8e43b0e02cf2df60f8a41029796a5644625cddf913ff10469a7456b9585ee998d8a43d74e03dcee2fcb9a6c280005a66ceb6a842da3474239437ba31b80834db28281dd593a08836f7681cc18089a96433f00fb73cc174ba6ad248bf576ec4d92fd9113b949cc3995c567386f72e8716ff3fcb14a7fbbde3963e335d90b4ad6980738a55ff14c600a7e5405846a058f1692eac2efb6e812ab9fd4740e4a6b8c0048e94f2efff2245cf57540efe9c637fc9efab2fea4aca02195f3138dfd0ca46", 0xd4}, {&(0x7f0000000400)="6883716d23ca9fb47d919f572147c1d447cebb5f0e8196387342115fd9152a9a163dda077722", 0x26}], 0x3, &(0x7f00000004c0)=[@rights={{0x28, 0x1, 0x1, [r2, r0, r0, r0, 0xffffffffffffffff, r2]}}], 0x28, 0x884}}, {{&(0x7f0000000500)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000640)=[{&(0x7f0000000580)="916ac906b38d404e6b309858a0bc615d50124c00ef4f8103ae2ec6a247b5f434b6fa9ad719726187cac0615a1456fe2604bf4225d6c03ae68d2e687333396aeea32e8267cccbd7c90e540389e5985edef54e7cb0fbbc9c95915d2be37ee1582e503a0e3c87623a43ea4212f88a3cbaefbcb284436ea3c6e721a6b4cb835e412a840f3ba2380b24647ff8bc6c5e29fbf93777a9ae8eeb4d34764050051331f7842175d3d6e14732e33b2ce2a97c9252e720a23ce0fbeacd0a8b4c469a3c251290", 0xc0}], 0x1, &(0x7f00000008c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @rights={{0x1c, 0x1, 0x1, [r1, r2, r2]}}, @rights={{0x18, 0x1, 0x1, [r2, r2]}}, @rights={{0x24, 0x1, 0x1, [r2, r2, 0xffffffffffffffff, r2, r0]}}, @rights={{0x28, 0x1, 0x1, [r1, 0xffffffffffffffff, r2, r6, r2, r1]}}, @rights={{0x2c, 0x1, 0x1, [r7, r2, r2, r2, r8, r9, 0xffffffffffffffff]}}], 0xd8, 0x5}}, {{0x0, 0x0, &(0x7f0000001cc0)=[{&(0x7f00000007c0)="86b6d2b7190f23598a777315a6d0754f6478cb3120b54384db723cd9441b1d21655a567a80b744df157d9efd59bd7c0929fd555ef76863007b4369ae79bdc58b484fb2059cf0d5cecc8d174e9e43dc3a916199acb957d8a7b5afce", 0x5b}, {&(0x7f00000009c0)="ed03565fac98ca237cb909d0a4956c8f932f501d85a6f1f2e3117526d383c39dfcc52c597cccc420522413ce769c4d258c71a2e3624bbfbf7c4d06e3340b69befc92ae0670859df34dc593bfb8f67c5ae077a496d8d3b9c64ab576b2f862c496d13e468c55bf8fb2305cb342f8fb1ab7d9afbe2e481621d5c42d7d2c159a", 0x7e}, {&(0x7f0000000a40)="1cef4312c27b7aa3e1e9d7d1b2604572c696f47e102b0038718ebc463c2b16597b3ecd5ec7f6170f4e3a56fa11948736cc850494eb5a37480d1850273f257204803fbfaa3be1e24755979e969002992ddb3fbaa595819f454793bbc98428f7026447da7b10028b3d0075671d5902562c09d2c652270d3a835b269abaa519489dfb92727ab24c2d6a555733a724c7299067490ea5fb7139d0abea6aeef5554827e3d28b60", 0xa4}, {&(0x7f0000000b00)="eaee41a87464a455fe9e344118a9ce26aa390da45fc0c2404667f044d78fbdf8fc8e8d88814007a325df22d058ee113f9bf524124259880baeae382963db327ff674b77b5955dee06f74c093b3e53c9ca22a0bd6d1ad0b7d19204eacf405e5b44101ccf891011689c8f6d282578c6eb9c3be5731e466cc6c652120f4a62ddb441d074b920e157aaab18cce3724e35ad0247893018b09856c045edd60ee0637468705fd10fbb90976778ab10c8795f5f8d5a08d9b5afe18c283dbe1f1d8e7008e333dfc857437b2cee8266929cf1819dc8fe5c310a8db4e11b1b3bfa57e6c3630940dd299b061758f9701735ea49b6790b145c8930ceb40e44bf5c2a08926b3b619e628886847b2d25251e027ea58d3d2da6531aa56ee216ae046eee746b49aa4f576c724f3fd039f7bd8c889f190e795a0b13d5cd676bef5a77b3ecf6822118ef8a4c6c8a7aee91c3bfecbbecf54ce16906c9a8aff1df3dba69bcd3c6078068fc0b5f882721ed398efbc04b0f7b4c4620247e50003fbbf131b77fd4208f580bb7e567d0f2861c0ade309b8f9c85ff326f993b944f4d3ac4be9f81823f7b3958a2dd8bb8cfda6c15f1cc38d257710dcfc73043113abc823ae9366cec4ce031098dabf775f4ebb448a8819d05f0203e839ffc1f942c19f4e0c4a26f3c8f1c530282bca3bbd223d6285de37b85390d8a47f3eac5efad9d5d47910d7c11d99857daee549e45f240f856dbf8bb068350ac5f8fdada1f20d913eb536992b5b37ece311d350cd4f0459444e73b30d400e53ef1dff1c1181c90d46a09a0d4b433b3ebe51ebf7e5e5809d9f4e5cf089e3cbda5ce239e20eb4bdfdf610ff0d892c283e2fb804d1140feea373f17adbdde5e52e9fb0c9cd5a5742355f1affdf01db40b4acc1f08fd066400b9fe4f23e47563019341bb0a269bf69b06c6e5b1c331bd4fadeea9ab00f1399d3b31bc0a98ccc09334f37b539ae8c390823da806a6f812a5dc41dfd3ffcc214c252eca465e925513cbdd35731f15bc51b27a9aca80e768cdd4706404b5408583c49ff98185e17ab4b9654a0bfcbc6c7a8aa67e3bb98a75ac81f2b843bbd59b3e34c4b03c2e349b38f7e534e72e43f5d3dd10849a17c884f3699f935fa321529c05f27dca0c3fb7351c873f69661762a536af50232588cac4584832bd2b6c1e33f9cdea20e6befdea2ce7097b48bf1d0c181ea431f9013097e9076c8626b77fc70cf4961df357384fc11091e357c6bfd100d533868c2239980d2942790f24d5546d4daf2c768f26c274a4d732eeca6dcfe014660310db07942a82cbf776b116e52cca76d27da0370a3116f56353b0e318dae59b78de5ec4510c30fedadbb21ab0da01297841b6635efbfc765e24000bed3d56d63a2a8f11ad1f73c2ad44119bc86c90bb364a2d7d580034199fa8f48db048833f8b13c55ca41d5c42c5b4bc86c8807c62a71e10fb736c5922fb54e6822d791dfb2d186b24dc3285913c6d133faa2c85a2219007e8c386f99bb2f3cf4b577cca94ecf69b4c0a479f33d315483b38dd8492dba86e96550c4c23089cd8f97b7b8adb6632b05745c51b3b97b13b001552f4175f487c61ecf768e246a5479083b2605e28d3b6d5f680de55aec8581c4415f800646a8d077899dd2b13812bc5b9ce66ee83e0aa4302a79354458136d447b4d5e36ea14d88ad624e10ece2b280e6563ccb5f75403429f5b15416a442ab1a47e29349473cb1be4c18cdc4705fd560a279ff298605b1be2b0b0563649810c5db8c8acaaf8d80b4feca8de9394abc929f42427e78076d1b46a9df2a80e291c8e1f48dc847d458919788afddae885d08f6388309af30b6d7c54c3d22c78622a74c336459bf7d8ee0a977f8b0422974f224b740331908d5cd33eb0e50f379531ff7764c1ade714ff57f81ab208f71b2b5df4bc05c562e0b2847c48feede38073e402bb6f9aaddf94f3b0a0259ebb70edd25e79e2f27dd6fcb7c8ac8957807d39676202a6a688afbe28db037bbb62afd42f3761b35fd25516fc0cc8a567faf5fbf77a897c532de946fe9b10e9d04ae42ec566917dc7fd1e571b4ced437a452799ddaec5abea6bd8ff53a97a4bbed5423abff7b791f6b80d4bee79e81ccf955b432ae51a675b8bcc83652f24de43dc25800da418f9ba5a721ab37bac1f49a502fe96d1624c01eaf01ae224c0d32db73fb7cbe8d7d6c171c3564c296638b9fb939271a2fa0c4b0128fb2c43baa5122c373600e226d9ed8c12a9e646c153cc1dde1d182dffd8a508650d6015d8aa75581399ba327015fd03b156d51b2476fed164a333241f2de6d419177716abfe4c56055704c3b48df64cfaa2bd4cc366f7d1b612b62d8092e271090f6c5183dcc64b36f50511b0d10f5a4953381c1834cf3e153cad78641c394937a6dabf3d5272f792acb3a430eca859a24698b1548cfc814b3a5d07dc85048d1beeceaa715798f34d62bfe39152e757cb400c380366d03f9a84993ffefd3ac73096e8c02a212e8bf0f72a99cd5971ffbbf7fc9a747cf1b021a1cd6bdb020e93e5483f958a7b4b72b36761628659a0052d9076f8652b175869e336f41c6b498c8ca717d906ae3eb7529eba1095936da71c0bacdd9af0ee943894eec88ed1d2c69861958f2555e778e608123d55e7591cda1a78fe0881b89f74a2b9020ef493438ccbcf6019ff5e83a2f407bc54b920f241966b3be9d777af817a1b3b095085ce3a043a3cd7c7888e20d85e2a30b12370cddd4e1c3692145b7d366d6ad35db9245aa04b52cf212ceb6a7331697d2a7e19fbbb2223bd2b35f36a76af6f14ca9a87ee95f738d7f3476d6f1e98d6e9ce02db8ad3b67a33c8d8152cb5c5c18a37da9616f31c0fb4bfc88014c9bcfbf25ecf8fdd2ca3b166921c2486e871b31621865b46ba920c8445c7d5dee7075ee8a2ff2dea00952cd804c1730b9115260e71c479048039f0442301599328dbed86132357d77b2100f5d5449afb7c83d578f01e392a823ad0bbddcd1414463907c1482b44e04bdd42ae6ffaea767dd9748957d24936bd3e38d13638f756408a90d9d066579ad9f53599d2fbf216139702c8a0bb49ed6ed926e0fc04aee5815c4114f6dd9777214bfc8fa4e467ac38065e03d5b9887d2df227c6abea27f53768dd80d1606a7a30bc70c8cdf2908f9e97509763fe983173d2ce210a813b4b6f3a64fd0a16942aabe2ba9ecfc4fd80254c1869cc3f0ceaf6243a57f9140f0b32c67593999223e94e0f354484fddebdc14120172ca70736b8ece004d88aaed13a8da2a0bf7a28a1df6ce79881c6130f3ceee3e4ec301b1bb8ab068471f80fd6cb75a679f5c98fb910b93d5614639bdd2d68d9eac0db54edc03c4b21246135922516512ee9f0da31261caf4dac8deeb5e2b11409cfdb108d1fe068bdfae1ae51d7df60ce2a5cffdaacb5a464e88ea2db8c34a7fc225cf3feb1b2fbe4d88a4d90d0242c953648d7549a4b7416cf903d43d349eb12e6bc75f80db07052771ccb6fbe755b455cf193a3f52cfba87f4b5fc9ae3a77d11414ed364991d57027509584965aec0ab756402e9caebba599b902b566aa43b482d8519965e7ba82275ee5d6df4f7cca8c3c0de939125bcd6b559145789d99d34cbeedc335a0df5770706851ee2bbf4b9646a3c259dbaa0b39ccf42580aaa26c3a4669be981743dadfee3532a813d8b2e7c8baeed233dece5c8724c49b8cffde538bdf83d6206efd1f9577649a285b37ba50b9809172a2dd3e48e81d04d92c72b5443794c333a1f414826cd5bd419ab709fcc611224bb96d7e6ded3f67feedd292a3607f149cf9c7654221cab67ababb207a346e1fb2a8d1287285ca14a69c1db46f72598f6a0751e4e4a4837cda0a333e2a9dc1d54f5fdf4c6192494933ffb317a62850e22efd31e176a19f543f1348376bafc72b7f2aa86c1967a81b98238dc50144af08c38fb7aa479e8411a8bf3991195073e461775f678e0a1edc3e53528415ba3bdc60d807b8934ac8ad551e6a34b7391d80a669ac4a01ff06e93fc223492f77179527fdb48d7417645fe4a7ac83cf7114a153b6139fdd6f5b7ec07da5f2f3b3b0120c8ad8e26b92578b1120997077752d930b64f3fbfd5239a6ba4087dfc186716d71777b68ec1a22aba67aeceddde61600aacb22bb2892a8c4a12587e724187ebc28d02a658d43b8b40e77e39c225d713952f30e2b00204e3290657b967468f5aa6bd3cd325d04abe165d5860e8ea099763eb7b4dfd299873591e4da0f4fb7bda13e0d59c1fbb36a28aa01290f1680d4068692e6bf83522ed4fb3c36a52a6a617a27630232b9a937cd482e59d59d55defb7cda5f48e78a72cdfcbba5f607125d2355b3ee7b33fec9943884af4417b9b067d7095852175cf9022527327f22ee82ae551690fddf869660d54a7a798a5d14c550cc966d90297d0fa3992985d468d6fc5bd6dd7184b34faafa8b0ae2ffdcecc26ed9f62d74b00bf424e46d575d3baace7262d563634e41646506d7cd8284df4d177db0475c119acd4d12b533579c7f2e4485ab11f15a3abc9c530c4fa113eab19bf72cb28a52de45b4eccb6e7ba53391a84f12a3a265ba778ec81fcb50f5f9684d4912161a7b08840ed88acb452364cc1833cc364ba856c8cf49eed7d560da085bab8cc7ae3b1a69167d94cd6f969471004ef728e04e08a88eb65cb0d9998a1323f88b75b784bede6a65e2c848f7e44802cebe43ed6b074cd830f9d093c8e8ad833400500e7c48fc93eb9c556c689bde6408ebcff4b7b138166e5f9e2ee91a02d8fe82aa6488a12f83630f7bbd3a66d7550877dd0c143b26cedd2b43d5474d3873bb3d6c7ba3908dfd4d0f1f9c3ab94d9b759189773c5e60128f1924907bbb51381eadd6636a388126884f06459919074acb0a255ec1717f74fa9584a46c3dd2e4572489814cd8fb394145ff025226f73c73da9844e1ddef654d5d42e4a9393ff69b4a0ff6db0c75b715a7e152c2626adfc8b1fe303d26aefffac667b15c240f973aeef4c9f46cf8ed563013ef541d02d986101a0a66f029e1f919b72ae1d0e6650e79f8492f81990d094dd018cf44f5e7f02be97a626523b3f0737169e16d2300b6ad91dd74a09831b2404987506a26bbddf5fb0379a5f3fdbdfab2d6416f21704ac86f21228bac744c138a088f644165db313c539c55ea5ddc38add09b35e2b1b5515e687ee137e173d33255336ba2ec6236706b4537c811ad9127d94ca19056f3c7a9e6703f8488450f3b1313291167c81b8c5cf9ca0dcba002da64595f6ad3e1399b4fdcd014199e7f6aed7b6899a8c2c9a93cd0f7d386432345dcef1594d132a14e494fd602d4d2a4bb7e8f77aba2cbfead5e383f717b8ee32a66310f7ae890100e10cb0a74c1b42d1d04308a6894434cdf4f04b0094bfaa5d226d906b8fa05389a6ce9aba416ad68955d5e73783577b21b5e04a60f612dd59dc61137913ef906456482e774fc3dbbb994cead8656cdedf939343c6fe11f8916a96159c5d148a4a4371ee3bb70928a6a496bb26f706873a8bd2acb1a62d1fcec2392a3873af151e9c1127abf4f40a580a4b5e437b06ff93eca6fcbb968db4467fa3de1efd7ecb245edb441cbb4ad03fd90742a059941c2bb6097a956f2e8e33e252d15020fea728014fe284e71db1bc014ca30b5734bbe46af8b039ff144e155b23ab4dc54e62e454ccd5ef5209663c07f953cb10963bc188ba8b0e991eef7343639e0afc94fa2611796ba66facb9428c47f385a45f3b56ecf814a7f6e5739d28941036ec0b9cc", 0x1000}, {&(0x7f0000001b00)="ffea4b91faa8254cd0422cd49e7d7b4ad911f174f04d3e969d69138087a9f76a795e25090f275f", 0x27}, {&(0x7f0000001b40)="59f0f837838a31e66ec2a56d15b9dd45c044c2b6eda6", 0x16}, {&(0x7f0000001b80)="1232786ac764324ce162ae0b2747c4b58a49b54c1a51187ee961bcd83fc3d824b0a215bc67888a7765dacb4b27352eb0c3b20f6212d8e5ee5e01d6b26a08fc6ca824cbb764639dcccde8b6bf92a73f89af60d93eff969b5061b1e4deeb431ea20572eccd896e7bde4027c61c6ae9862baa89ec5c4aade26ac4aa27c491c3e513c47637891edda03f8b579a8a84966fa2b1bd3c0d202693a5d00ca289ca2c392ea1290a48604a4878e5dcff5be2a48606fda6f98124ea0c3b7f8e054ba096631dedc1264df2", 0xc5}, {&(0x7f0000001c80)="8962099d22fb862d6fd6d6973ae1a275cd22e7cc524cade4962f6edcb6414035122de5e237dcf5ec083475596fa365a7", 0x30}], 0x8, &(0x7f0000001e80)=[@rights={{0x28, 0x1, 0x1, [r2, r0, r2, r0, r1, r0]}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @rights={{0x24, 0x1, 0x1, [r0, r0, r0, r2, r1]}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, r10, 0xffffffffffffffff}}}, @rights={{0x30, 0x1, 0x1, [r0, r0, r11, r1, r2, r12, r13, r1]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xffffffffffffffff}}}, @rights={{0x2c, 0x1, 0x1, [r1, r0, r14, r1, r2, 0xffffffffffffffff, r2]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, r5, 0xee00}}}, @rights={{0x20, 0x1, 0x1, [r15, r1, r16, r0]}}], 0x170, 0x4000000}}, {{&(0x7f0000002000)=@abs={0x1, 0x0, 0x4e22}, 0x6e, &(0x7f00000021c0)=[{&(0x7f0000002080)="d7d7b9a6bd47a22a6161af0fa6101c1a79804ba49b11e35f34aa94f92990ceec01ef5c48d5d0a69996ff7a6a460bd98db69b6117248abbdbd02209c218e08448c5e77fd8df7fafb32f0429810b7550be7550ed7970da86f80982c601f8e95e655587efaeba64e9e2", 0x68}, {&(0x7f0000002100)="b2f97dc09b90d141535f9ac9d035343d527176653676c4d8fd8086add2f762a814d75f670d43b3ecbb8198fbd38618ed3778652819952126e1e620359e9b", 0x3e}, {&(0x7f0000002140)="136e38778c34c4758567b97dd90283afd67de25a112f3ec344a1f209207d8929394f2467e5e56ba3971838b476b2a03494fe667e98aa21d58fd0cdd8f3174f33759915d73b6c4870a60f16cb2d3b69cf0708ec8a9ce4543ee0b5acc38d9e741a6c8af9b937818e597c333b85a806ac36a7d445722e6ae67682", 0x79}], 0x3, &(0x7f00000022c0)=[@cred={{0x1c, 0x1, 0x2, {r4, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee00}}}, @rights={{0x2c, 0x1, 0x1, [r0, r2, r2, r17, r0, r1, r18]}}, @cred={{0x1c, 0x1, 0x2, {r19, r5, 0xee00}}}, @rights={{0x34, 0x1, 0x1, [r1, r2, r1, r0, r20, r1, r2, r2, 0xffffffffffffffff]}}, @rights={{0x14, 0x1, 0x1, [r21]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5}}}], 0x120}}], 0x5, 0x0) getuid() 00:28:44 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1d00}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xebffffff) 00:28:44 executing program 0: sigaltstack(&(0x7f0000000240)={0x0, 0x80000002}, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) (async) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) bind(r1, &(0x7f0000000040)=@caif=@util={0x25, "989a93b67dd8f49f40ffe5bb900c3661"}, 0x80) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f00000001c0)=0x0) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r7 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) getuid() (async) r8 = getuid() socket$inet_tcp(0x2, 0x1, 0x0) (async) r9 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r11 = geteuid() r12 = getpid() sendmmsg$unix(r5, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r6, r4, 0xffffffffffffffff, r4, r7]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r8, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r5, r9]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r10, r11}}}, @cred={{0x1c, 0x1, 0x2, {r12, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) ioctl$UDMABUF_CREATE(0xffffffffffffffff, 0x40187542, &(0x7f0000000200)={0xffffffffffffffff, 0x1, 0x8000, 0xfffff000}) (async) r13 = ioctl$UDMABUF_CREATE(0xffffffffffffffff, 0x40187542, &(0x7f0000000200)={0xffffffffffffffff, 0x1, 0x8000, 0xfffff000}) socket$inet6_udplite(0xa, 0x2, 0x88) (async) r14 = socket$inet6_udplite(0xa, 0x2, 0x88) ioctl$ifreq_SIOCGIFINDEX_vcan(r14, 0x8933, &(0x7f0000000000)={'vxcan1\x00'}) getresuid(&(0x7f0000000240)=0x0, &(0x7f0000000280), &(0x7f00000002c0)) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000300)=0x0) geteuid() (async) r17 = geteuid() r18 = openat$smackfs_ipv6host(0xffffffffffffff9c, &(0x7f0000000340), 0x2, 0x0) r19 = getpid() r20 = syz_open_pts(0xffffffffffffffff, 0x400040) getresuid(&(0x7f0000000d80), &(0x7f0000000dc0)=0x0, &(0x7f0000000e00)) getpid() (async) r22 = getpid() tgkill(r22, r22, 0x816) (async) tgkill(r22, r22, 0x816) r23 = getuid() sendmmsg$unix(r0, &(0x7f0000002400)=[{{&(0x7f00000000c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000180)=[{&(0x7f0000000140)="a2f0d12222dd29494720", 0xa}], 0x1, &(0x7f0000000380)=[@rights={{0x10}}, @cred={{0x1c, 0x1, 0x2, {r3, r11, 0xffffffffffffffff}}}, @rights={{0x24, 0x1, 0x1, [r2, r1, r1, r13, r1]}}, @rights={{0x24, 0x1, 0x1, [r2, r14, r0, r1, r0]}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, r15, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r16, r17, 0xffffffffffffffff}}}, @rights={{0x28, 0x1, 0x1, [r1, 0xffffffffffffffff, r2, r1, r18, r0]}}, @cred={{0x1c, 0x1, 0x2, {r19, 0xee01}}}], 0x108, 0x8c0}}, {{&(0x7f00000005c0)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000b40)=[{&(0x7f00000004c0)="774abe746686470cf2a5184e6d084ab4e6e0a7ef71f02477b3ce8cf639d4bec031680abb0c08e5c34ba3e3cb8d5e14", 0x2f}, {&(0x7f0000000540)="85ac0b4c78faf59edfaaa19ba7d874e4293614", 0x13}, {&(0x7f0000000640)="facb1498e4b77704757538110f0c36d751027c57214910bbccda159fda24e3c503a8df01c71de8f9d8", 0x29}, {&(0x7f0000000680)="0b4720db8d991fc867657defadaab95a9bc86160e23715f7a4a54e29fc84d0ec7820c70e5a8aa502d91bc4371175a51d3259839d7ad362c2e46b234feb41f8a1b0c05903fca67f3e515140c2e0bb7bf69844129b1b9df18dce96336d1810194c6fc606d8dc6509c45eefc82f64d93c993c66a71052b644d7f143fd53c89aa5b2340cc49cb44bca59e5d9ea8f8d46288db71668282b24af068e652fc0ccf9b7b1c68223b0c4d41f78b5fc7abd313cf47b6ddc0dca928620", 0xb7}, {&(0x7f0000000740)="1ab94c45d888547a7f1adb75f41d87a637f79ba33253ea1343eb645aa4f6e53306024580b39aa33f4d648154b1ebd345e4c6cd2131bb1a98fddd0422d16eaa7bdde1e4106aaf92303e304ae6d303c79bb60c5686a05bf104a05779bb251d61122270e0046483b2c10fdc5f07cc960d0f67c7e62111e035653b3140820b179abf768abeeba3a0c514ae9e22e17a24590e67209471", 0x94}, {&(0x7f00000008c0)="db1f1096417a77bec9b051e9383eb05e81a3120eb1c81a4346ccd490b58d813a6cb3456d6455195196662116a9ba9c62b014311cf64ce69d7d86ae64880069a52a7d021cbe356e1273edb81c5579c80a68c3669f314c9870d20848e4cb74da2839cbf3125be325f4174352eb2623f8f985a798cba7280b430847f1c53ce318635bf13bdf67974f01bd66f9d050bf1efd05298381178c72dda7627d4d03f41a21eb4c3f0a9aa21ef41dfacaa77511a3ac3e10b1ac91180a42683b0e49efcb76b877f389d32db16062dbde711fdb33aca86afe95f55965c6b908ea9655093b780070ea09a8739fb14e8414dd3d1f3220ed", 0xf0}, {&(0x7f00000009c0)="34bdd92d60ab7b4264d48b7ec5a9d9b24702e2424241073fcd1e29047010b7ada9b609e2dd3cb85dfca292da79bfe1854a24a86cbf9270c85961a106a1027b0ed012662f3416cd7797bbc6ecf5e54d8f057eb83051c1cc72f05d2ab02aff2e7aa99cfc87c05236cd74984c120c1c04df0391bf1a2ea33f2058fe18e4a02f11a7581c1f4115bc46a82c28c05f", 0x8c}, {&(0x7f0000000a80)="fa4f107f1a4e405b0f8c6aee172427f8a587aac17d28ec4cf440f9dd7734662d54a8311e1bd80c672f7604b5bb9c459b10132a12571b6edb2b079bf07e80b909a237ccb0e69f5c22885321781ff13892cfffc84a39079ad37914512e1ec0eb5562f09a4f4af9bf0b2f1609a4dfb74dd56c5e5a5f72f7334ae1ff6fda4db9ea4941ae052b84a83d8c5f52ce841eae04d924991a06aa", 0x95}, {&(0x7f0000000800)="5dc66e7a1a0810683168cfd46af4ffad98626025459b889c418c17bbb67c275a56e96a41d76bdc33bc91db2901efaaa08c633b5c8b00a48470707fa3113eb9e161e3e8851a8c4a0fc5bb2313bb", 0x4d}], 0x9, &(0x7f0000000c00), 0x0, 0x4}}, {{&(0x7f0000000c40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000d40)=[{&(0x7f0000000cc0)="5054747e0cae3571d87bde5d51e6c5220f73d1b2f09bf71c97d5efbf978cd49442a12a1563bc18d4f015a4694ab9c1e7e0f8cccd86f4c98eede0544bfaebdfddf46738ad6cf64f066f5fc1c7522b93e90427", 0x52}], 0x1, &(0x7f0000000e40)=[@rights={{0x24, 0x1, 0x1, [r20, r0, r0, r0, r1]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r21}}}], 0x48, 0x8081}}, {{&(0x7f0000000ec0)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000002300)=[{&(0x7f0000000f40)="f91b228507a65d8acd958d73de9502a29dd926b6869244b52cc9093a900598751b8b2e43a96967796b0989ad6aed0560efb92235da57cc7253c2e55345f8af7f0cedf62b43f7b97bc506f1f708cea1c1988dc26a75357a3739396b96d1c3d13178ec2155cc62449172dd6a88a5f3751917d2ba085d03d630a8cb49dd6447df1dc13dfe93711cd3b16044d92726e3190196ccce04e4d7d05f450ff8312d01ad4a12cf7f0d060f91ca8c3865c3b607edef128af28a9921a1e5c97a27354af6da52b34e5e027388fb1df81440394362d39ca74ba492513d17099898e539598cb25de8c7c00a6514a009195894c36b1d058d478e9341956bd3c283aeb0cb106c9ef8c73d367528728d769c84820259367ebf6817db0edb6a97f8da188f4c7e0eac0c467c90748bf3f09f159ec13d85720407506844198dea186da5efb2c739ca3fd4e40d68260efef0e678fda0dfa1a3bf1ef927f60693755f884f2ffaf0a5a13002e7de294b06b7c292f2efb592f027ac3c724f58fe8888f420bb74cc15283156b176a3d35f8dcd7356d3ba0902bfc52ac8a2dcca49a236d7f6c51c944acb46e9cc4a40607e34f6699d0a4970f3a9fdd793c96520d382851b3ffffe11e0a21358d4b5359a655c4b123cdce511fc250a8e67264791002c89ebb736cc0556deadac8d5e8e2b9c5f1fc30bd22786c58da4869d3cc6bb2843138197591fc8f484337030213093bb9369b5339b789d961af830d532a2b819e99aa9248de97bf096fd00532e943bc47db3e414040709ec88d742b8300b4694006afaf118692fb23b77c9141ad7f40db583bddf7e14dc961210195c1921a20bb4bc503b61e1835d490edcb8391cacfdadf1ff2d1ddda3d986a843cedb0e94b0dd4b6cef5cfc2a4698899dc8c7a1920167f0025f2a452137022e92456a99d8aeb65940c718b90d4ffb6e4537b599080e405cb29cfdb6149d186c07815a0daad6b4af2ea2d675df42b67b2e4f80eabaec36c6168e1d6a86b6d4172d6fa76f3df868327cbb0825c37644052750a525da021dff3a4cfb0ff32b39f7af425a2d931074307e091bcb59b3c63aef996dbcd3c238d5dd0dd0bee37efb722283bb13bb2cdc354f7eaec2ade78df44ae42491a52ca1e4b7a238ce22bf526a99a79fd3278c0195bc7dadea4063783b342a7ff016aca8a174134aad2d9ad0410510a3ae1989798846418e2edfed02b108e09a5097cb7d4132edbfe3e4dc93c8cd45d0142a886744e9da45ad2a22e832b4894a6c821fa3ceea828d1d39afb99c90ed310cba20ec7f5ea84edc0c5b0141027653d2bfe7aca63fb9c4b898aef2ba8d3e5eab9f9b83a0e402ccf8a6b93e496fef608358b08d774220928c08d06a63b7a19fd8e843752ae7d103ea7fafe82bc8a4137acfcac6dc8950751227833b6ecdca2d6f6f7d39f4f6ff699059bf7122c3fa82baabc806ddcd88528cbe02b09eceb8ff8d9055475415dd0be19f85925346d8b856791778ef722b64bfbfd3f11b5c712bb692d1b58d9ced6e3d24006cfef8f836ca7086abc5a18fa433afce08ffd971f03b185abf2938ce3312a1b1cd407126413669adf32ebce2b92bc5f9e533863b83526a80652766c0d73bd37ad9941bfcda0bd8a83b4c59eee14f99b2f92aeae4959100838a9dc002ecbcd0752782b1abc5251a70a4aa5fdcbc2b721e6b1b9e1b98f6de74a4fde2a09a4716e2bdcf72c65b0ba94d8d83ce02df70d12d6181174afef53336fcfa669e1cbdd3a159072c7e4232da5034778a1216242182dd7a71333dec28862306a3a10a836122f281a68363357dafd9eb91f3b57d2d722c79a41d072c9060783480a4fe32ce757edfee4ea2ba79c61c5de232479298749dbfc3ed47f54ee249f4aa7a51283603d8196c019706b8afd1f0da98fd62695ef81c405225cadae5b72acebc5afde67609f798462dc4fd512f9d37098d16919c24f0c32d3afc6346e242ad6aa4671352ee3017abcc287f0b641dc99e8ef9fa26e4733bbacf0197da3df7c907d822b212ee2e556ad4ac3da925ffb00aaf1957a0b62c92545adeaf93f8eaebed0a9c1875d073fc386c2fc15bb6d6eaa1837e24a9aa8a1542db86931191775d14af0992ba5f150b62b68014d604e67cf4bb566d8c3deaf45d8bfd52c7cf921fd8ee228022c4dff1c406fdd3d52d63bf685aa33254dc7fc0042f03d10150beed5eeb5c376cbb9b6905a160dd1ed8ba551331070238d935818d278bfdeff1042fe61cc741809f6663d66817bf14bfd3451f8c78ab7272a5e399710684f1a3c7ef43285622db81bc16c30cd192e1c284b2214819009715d7af97d3bd1a44a5457a43ea2d9b23f06ac87b83b413db35546e122a633a046e4ba143615efa3599000b38dcb3fa6724dc6c72854088439fde876e3a88f1bb11c89f45a5ed06fdd35ecde31d51ca25e12853d38bc38ba3c2b5cca418cf95fea956af7f8bd6af27831c97c2dd57f9fe3ec587ef4db32fcf5deb663f6449bc8ce44afaa1f3b7049f74c5e869b18ccb940fef96b983a56df6712e6fdc046141eb3e62fb8abae47177d20f035ab07085ef244f461b10d8505e841ec0cdcebd8998eed3e667fad496d9183e5971d004a85c4005b4c4bf02db32667b16450b3e491efb1099a0c2919de31b63744fa3ee9b9df3182946cfefb65ddd35bba4c156c84e5525d91ef09d320b164329b153178e7e77032552524d8436a18877c2d582a138e3b3581071d2965dd0dfcebf3a6067121eda4c39491a0dcb257399099b4511f383652f913c60253f6d6e9be16b21fffab2a90ecd0aa23e369a0b675afaafedee15a5558ecfef096de66f20710b2b22842ec0027ba80be114454cbb25bcd77edd727a7f56f0e58e6dc627f23e62175babba39b4cf4d3161299f24d09882082c82c1a144f098323866e15c1b79781dd7115f3a9b67b97b40284ebdb187d9d51a2f7a399221a2a3e9e2d6e23270d66078e8aef9369de80fab9bc0953215c83e91e64734623fc112c9abc4865405688b600dfeedd49f9e9666ca46dacbcff2d543781cb31d584aa0f0061bca93a700f8c284eabf2b7b7ecebb0ac66c575ab8565067b3f8914dde503c763bd16e94891703409e478ac02731e7563016ff5b38e5d3d9b80174bd3fcd0cb19214b7c15fbd2192af9a283c63cf42e7c42090e3fb6cb914d16141e28b24e6e47b154e528795d2dd59de21d942e2cb3fc154824ad5463090cfd00463daa2407b15508597d9b63fa2badeec049779292e4c4e952dc753d26c735a7f5abd532e34876c08ede94d2d0eaa1575cfb423d6fa1cae0034490e1c80680eba263d69a2c344e57f8edfd2d359f50128e0d5d17247cd3c2196b993ec5d33a99b2a54ecae1f60a9fee3a193d3a1d5ddeee6e42f093bd933625966aa779f70f9778e830f0e39d94a357c4078c467c863cf2d319aff7525dc9861658a91b157d026254f15a8d4481fdd285f60b3906d68bc514fb982a3f1511679d2d335fe5138e84d52c41402b9a565a249d483c6f2e14843adf62be9e64eb97818280b76b889b818d995db582bedd1e4fb8aea3118f7b493300f5b886cf2d724d428ceb591b2ef92acd93f9b1e9014bcab3c584e6b1c094d7de76d65eeae5a20ddeaa9c208ea702965bcb8979de730014fc64e841682278afcfc89c93feec7029cd73c43b300a345bebd91bc6a3c200c606b924a416a8a62321f7ee710ff0b3b655db046d3745819c6633e396ebf7b83ad153ea796e878e85bb6164d44304281ecce15944219ce86181d95eb97f5addb763ebcd6d90b60300ac4385df78ad10e515ba73fa15a8471d11a7cd557584c6fc1dd733faa8e6d0f34fbb0686a7b34a666f147fac5cfa77e641965472a46567f0433e133eb25d786e3b88f5bdd857458778c4ca07a0f3a92e49dfc47b47af5ba71daacf535e40436e6959fa1c482478a3a4bf19e29aa4492e36d859659da87613bd8696dfc1c7217c3932f70b21deb4da27f161150e0da491b8999007aa670b9defe7d595c3026d3f9a7a47db72dcb2b6d09da1fe55865b2aee3e004de8e404e986fb59522d4dae5288afd5311ec87f55291fc621a45ce6404729a93389697e5f2a4006b25bf79ef92ef928a8703ed42d78082b1011642353b8aea9dba4c1e2bce1adcc0a9b8e982aaef18b299982af7e992be24e593329e9aac877c437718b0e025b49d40c93a594244624a1179ef0502ae3aa6e845b8c2fce31a1111695fd63939921cd8b4b58cae6d553e7b8d9fbd297bd2512d63b45b96f2f268a51201b5d03cf41403493d6b2ce31322234cdb5a44028b04c65c09534b8e6ae7c0a2aba7ebeddb5f06e6fb944c661177e870545a3570d7baa6e469c4dbd895e81517c03521254508f39823873f47737f4f331e88e2ea7536e215490069d400cf6a8911c6b5573af3179aec96c20a6c4cf7b6b2157767254648116686be68251be326a93232abe0677e23934c4b72d8894292b674cf9c67085595e9deab4b862f71b7043c092504993429a13e31c5532a8c814fbb030a5e8fc14dfda8403bab82eeb9a58b627cdd008f6ff664bd1894987c4eb37e0e916b871199e300bd817dd0098dfc9d1b4155d18f612e137812315e4253ea4ca44fced536a63acea94458ed34296815fce552ac5c8d93be1fce4fe9f377215211787c956f87bb0e86a9b96a7a6fbb9d37b7423692336c3d4efcf7d463f549119832e7a12ead655faeb1ad3bbd3ff0588d16a4c3e07d6b4d7d6b33736e8a7038c16ec6c0182d2a07253c8e27f944ee8f638ee0ab1ae64b30960875c452b2144c11a7f6639d5c2cd00348a0c6277f18eeede0d52163ba282a19edd8d839cf734cb248bc3047596daed6e572522fb24c95734df5c4387c69e2854fc83e5d3ce37145e445eaf199dc4f2212c932b9cccfe82bb8a081639b3226323d8ebc6f75aff71c5edef8455ed4fad758abcc2e1333a5230efc90478b8975f04f0b05d3f7cfd41d9df357e7d197699e63afebab617fb8668010788d1f29e4bdc1f6f2855f2051a8bb0a37408f2c2e6aa74a92600c280177c515bd5ea9e1fd52e8f3cae36975be4c725e2a7c4bf56821690e093dc2cb824b2a4fc055e43a13cd03f029c85cfd19d46067d12694d9013f6fd4d9856f3f4074b582c90a559956eb385eb273d59f3292ef93a607243a97bf00f03667fd59ff3aed06fc4a99f0793614bf88377bd94e5fc59b7b06cc7f0402afd6edeffa049471f751b0c86be0acf64964c2864d3b2e7fe884d7d9827cd2b6861f32c76f0298c97fd75c84dd2f9f598e5ec1e96ce2474f50ffe892ca21f18e585972e319a8e1a3fcc65d1cd9aa206dafe1ee5fe58803428a165126306a35c1777e7dbdd833d8f9213727818910c82ef18e3b6cfb1d83c3c50fd1c58bc869206944e4ab2bea82cd86bb9aa736f4ea46ff3fc7e1138a3693e34d275642aa9e7ed0b8fd57e00791ffa243ed8cab5e6e7a52c8505c280e70d21900856690b3e35725f025523072f9ff260cc45fba52f9cf8f59c750a6c250b9f14bf151c56420ffa4a3d9212f73d7ead42626da73618994b6d5532bb00ff231228e35aaaff89f2ae4507cc727d4676403887bf8711afd70352245775f32c60238b869def27306e884705e0674dd9ebd5f939fbe25b213db2b4b72e5dbbb143fc0e375b8b1ae33e7fa558640025ad028247262a3bbf43bde6630934156655a2a602d0eb689888005e00110a0ac7afb51384e4fcd9012f08489b5dcbaa04a79bc3e0b28c15dd832daf82447083ce3483b6346736cf7b1fb", 0x1000}, {&(0x7f0000001f40)="b45a5a2307e1ac7e01f97c31cb42f2cb6eda15b9ad448f7494b2dfb79bed8ad68cf427b4242e", 0x26}, {&(0x7f0000001f80)="be2314fcd1293d74b2dce075f0e2918da58f7571a54d17768472821bb40a5e4c548ce8bda11922bad379a8", 0x2b}, {&(0x7f0000001fc0)="aeaa9b32c2b608f36f874397edbf4948fea371d0dc83f2a2bf1316f7335a1c80e2e419e4c2c6a21b0d18d8c441a8c7ccd5ae10d2725a6d440fa8b4b7a0a77c8501c2bf77495f3e4636e434a6a76f8d50059bc3f8bc7114cb62b2703edbf50145bd2d0ea92ae448a6828c011950b995c39f792f23c0f7ebb554208d46ebd7527b10f57a97357ec96aa2885edbdaf9e3d61028d14a63c976a4bc7d4bbd3914f9949b9006f0fd32c139b90eb3eb8433993b4cd93e8ee83f53012009cd023cc603a8099adaa39c9e7a23e02b536318fb", 0xce}, {&(0x7f00000020c0)="9ac51d1478126c0793642800ae4565197bb4cfec43ee5f65aa6fcf5fa33a00b456196df995daea32c3a16ff550a16fa5068348762ebde0e103cd5eaf3b1c63f504ec46d04187c1f0c7674a28ec618c368b4ecf3e56d5bbf31c3b256706731d064874cdc81aeb934060494abd40ced490356863cc41e9c60c604f77085998c777528723984903e2205813a77d9243b376df0f25815028ea1f8d637f300d17907a67dd074854c998224ab3d580b4bd1e3f98ea4a20312c0efbcc80bf0cee0f0ae5eee7ee9d1d023ac9a9fc3c7a97597775ce4d845d3ef5c117a15906d178d4dc0953", 0xe1}, {&(0x7f00000021c0)="e42d16c0f9100ebbac570b9eef9533cac7a68494625f335c662630b7c0236224a17e7c34", 0x24}, {&(0x7f0000002200)="5e6a3487aaaf2bd80d9902c749c134afa1bded518edad98d01c63b13a76ee753189b078073204eca69a144019e8c1ec576d6fc1f709e82b414b820884ebc18870cf4cb90f8a8e4a4e2becb5d4ee5ee1c6e57de5bb7f5ae63cc0f80132fa5f91c9b58e905aaabd8e93e6e680309c4d2cc66724b2d21e5c150216d5b0c80e0d9f7b01e28467113854adaa13c726963ef0166c6db4257e7b0eba874f3e8fd368597d94d8df975944ca0cdeb6fd7b1fb0ff47e488092cc4f6b5da48e753cb6af832167887411cda566314ddb64391e", 0xcd}], 0x7, &(0x7f0000002380)=[@rights={{0x1c, 0x1, 0x1, [r2, r0, r1]}}, @cred={{0x1c, 0x1, 0x2, {r22, r23, 0xffffffffffffffff}}}, @rights={{0x14, 0x1, 0x1, [r0]}}, @rights={{0x28, 0x1, 0x1, [r2, r1, r1, r2, r0, r1]}}], 0x80, 0x8015}}], 0x4, 0x1) (async) sendmmsg$unix(r0, &(0x7f0000002400)=[{{&(0x7f00000000c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000180)=[{&(0x7f0000000140)="a2f0d12222dd29494720", 0xa}], 0x1, &(0x7f0000000380)=[@rights={{0x10}}, @cred={{0x1c, 0x1, 0x2, {r3, r11, 0xffffffffffffffff}}}, @rights={{0x24, 0x1, 0x1, [r2, r1, r1, r13, r1]}}, @rights={{0x24, 0x1, 0x1, [r2, r14, r0, r1, r0]}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, r15, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r16, r17, 0xffffffffffffffff}}}, @rights={{0x28, 0x1, 0x1, [r1, 0xffffffffffffffff, r2, r1, r18, r0]}}, @cred={{0x1c, 0x1, 0x2, {r19, 0xee01}}}], 0x108, 0x8c0}}, {{&(0x7f00000005c0)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000b40)=[{&(0x7f00000004c0)="774abe746686470cf2a5184e6d084ab4e6e0a7ef71f02477b3ce8cf639d4bec031680abb0c08e5c34ba3e3cb8d5e14", 0x2f}, {&(0x7f0000000540)="85ac0b4c78faf59edfaaa19ba7d874e4293614", 0x13}, {&(0x7f0000000640)="facb1498e4b77704757538110f0c36d751027c57214910bbccda159fda24e3c503a8df01c71de8f9d8", 0x29}, {&(0x7f0000000680)="0b4720db8d991fc867657defadaab95a9bc86160e23715f7a4a54e29fc84d0ec7820c70e5a8aa502d91bc4371175a51d3259839d7ad362c2e46b234feb41f8a1b0c05903fca67f3e515140c2e0bb7bf69844129b1b9df18dce96336d1810194c6fc606d8dc6509c45eefc82f64d93c993c66a71052b644d7f143fd53c89aa5b2340cc49cb44bca59e5d9ea8f8d46288db71668282b24af068e652fc0ccf9b7b1c68223b0c4d41f78b5fc7abd313cf47b6ddc0dca928620", 0xb7}, {&(0x7f0000000740)="1ab94c45d888547a7f1adb75f41d87a637f79ba33253ea1343eb645aa4f6e53306024580b39aa33f4d648154b1ebd345e4c6cd2131bb1a98fddd0422d16eaa7bdde1e4106aaf92303e304ae6d303c79bb60c5686a05bf104a05779bb251d61122270e0046483b2c10fdc5f07cc960d0f67c7e62111e035653b3140820b179abf768abeeba3a0c514ae9e22e17a24590e67209471", 0x94}, {&(0x7f00000008c0)="db1f1096417a77bec9b051e9383eb05e81a3120eb1c81a4346ccd490b58d813a6cb3456d6455195196662116a9ba9c62b014311cf64ce69d7d86ae64880069a52a7d021cbe356e1273edb81c5579c80a68c3669f314c9870d20848e4cb74da2839cbf3125be325f4174352eb2623f8f985a798cba7280b430847f1c53ce318635bf13bdf67974f01bd66f9d050bf1efd05298381178c72dda7627d4d03f41a21eb4c3f0a9aa21ef41dfacaa77511a3ac3e10b1ac91180a42683b0e49efcb76b877f389d32db16062dbde711fdb33aca86afe95f55965c6b908ea9655093b780070ea09a8739fb14e8414dd3d1f3220ed", 0xf0}, {&(0x7f00000009c0)="34bdd92d60ab7b4264d48b7ec5a9d9b24702e2424241073fcd1e29047010b7ada9b609e2dd3cb85dfca292da79bfe1854a24a86cbf9270c85961a106a1027b0ed012662f3416cd7797bbc6ecf5e54d8f057eb83051c1cc72f05d2ab02aff2e7aa99cfc87c05236cd74984c120c1c04df0391bf1a2ea33f2058fe18e4a02f11a7581c1f4115bc46a82c28c05f", 0x8c}, {&(0x7f0000000a80)="fa4f107f1a4e405b0f8c6aee172427f8a587aac17d28ec4cf440f9dd7734662d54a8311e1bd80c672f7604b5bb9c459b10132a12571b6edb2b079bf07e80b909a237ccb0e69f5c22885321781ff13892cfffc84a39079ad37914512e1ec0eb5562f09a4f4af9bf0b2f1609a4dfb74dd56c5e5a5f72f7334ae1ff6fda4db9ea4941ae052b84a83d8c5f52ce841eae04d924991a06aa", 0x95}, {&(0x7f0000000800)="5dc66e7a1a0810683168cfd46af4ffad98626025459b889c418c17bbb67c275a56e96a41d76bdc33bc91db2901efaaa08c633b5c8b00a48470707fa3113eb9e161e3e8851a8c4a0fc5bb2313bb", 0x4d}], 0x9, &(0x7f0000000c00), 0x0, 0x4}}, {{&(0x7f0000000c40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000d40)=[{&(0x7f0000000cc0)="5054747e0cae3571d87bde5d51e6c5220f73d1b2f09bf71c97d5efbf978cd49442a12a1563bc18d4f015a4694ab9c1e7e0f8cccd86f4c98eede0544bfaebdfddf46738ad6cf64f066f5fc1c7522b93e90427", 0x52}], 0x1, &(0x7f0000000e40)=[@rights={{0x24, 0x1, 0x1, [r20, r0, r0, r0, r1]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r21}}}], 0x48, 0x8081}}, {{&(0x7f0000000ec0)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000002300)=[{&(0x7f0000000f40)="f91b228507a65d8acd958d73de9502a29dd926b6869244b52cc9093a900598751b8b2e43a96967796b0989ad6aed0560efb92235da57cc7253c2e55345f8af7f0cedf62b43f7b97bc506f1f708cea1c1988dc26a75357a3739396b96d1c3d13178ec2155cc62449172dd6a88a5f3751917d2ba085d03d630a8cb49dd6447df1dc13dfe93711cd3b16044d92726e3190196ccce04e4d7d05f450ff8312d01ad4a12cf7f0d060f91ca8c3865c3b607edef128af28a9921a1e5c97a27354af6da52b34e5e027388fb1df81440394362d39ca74ba492513d17099898e539598cb25de8c7c00a6514a009195894c36b1d058d478e9341956bd3c283aeb0cb106c9ef8c73d367528728d769c84820259367ebf6817db0edb6a97f8da188f4c7e0eac0c467c90748bf3f09f159ec13d85720407506844198dea186da5efb2c739ca3fd4e40d68260efef0e678fda0dfa1a3bf1ef927f60693755f884f2ffaf0a5a13002e7de294b06b7c292f2efb592f027ac3c724f58fe8888f420bb74cc15283156b176a3d35f8dcd7356d3ba0902bfc52ac8a2dcca49a236d7f6c51c944acb46e9cc4a40607e34f6699d0a4970f3a9fdd793c96520d382851b3ffffe11e0a21358d4b5359a655c4b123cdce511fc250a8e67264791002c89ebb736cc0556deadac8d5e8e2b9c5f1fc30bd22786c58da4869d3cc6bb2843138197591fc8f484337030213093bb9369b5339b789d961af830d532a2b819e99aa9248de97bf096fd00532e943bc47db3e414040709ec88d742b8300b4694006afaf118692fb23b77c9141ad7f40db583bddf7e14dc961210195c1921a20bb4bc503b61e1835d490edcb8391cacfdadf1ff2d1ddda3d986a843cedb0e94b0dd4b6cef5cfc2a4698899dc8c7a1920167f0025f2a452137022e92456a99d8aeb65940c718b90d4ffb6e4537b599080e405cb29cfdb6149d186c07815a0daad6b4af2ea2d675df42b67b2e4f80eabaec36c6168e1d6a86b6d4172d6fa76f3df868327cbb0825c37644052750a525da021dff3a4cfb0ff32b39f7af425a2d931074307e091bcb59b3c63aef996dbcd3c238d5dd0dd0bee37efb722283bb13bb2cdc354f7eaec2ade78df44ae42491a52ca1e4b7a238ce22bf526a99a79fd3278c0195bc7dadea4063783b342a7ff016aca8a174134aad2d9ad0410510a3ae1989798846418e2edfed02b108e09a5097cb7d4132edbfe3e4dc93c8cd45d0142a886744e9da45ad2a22e832b4894a6c821fa3ceea828d1d39afb99c90ed310cba20ec7f5ea84edc0c5b0141027653d2bfe7aca63fb9c4b898aef2ba8d3e5eab9f9b83a0e402ccf8a6b93e496fef608358b08d774220928c08d06a63b7a19fd8e843752ae7d103ea7fafe82bc8a4137acfcac6dc8950751227833b6ecdca2d6f6f7d39f4f6ff699059bf7122c3fa82baabc806ddcd88528cbe02b09eceb8ff8d9055475415dd0be19f85925346d8b856791778ef722b64bfbfd3f11b5c712bb692d1b58d9ced6e3d24006cfef8f836ca7086abc5a18fa433afce08ffd971f03b185abf2938ce3312a1b1cd407126413669adf32ebce2b92bc5f9e533863b83526a80652766c0d73bd37ad9941bfcda0bd8a83b4c59eee14f99b2f92aeae4959100838a9dc002ecbcd0752782b1abc5251a70a4aa5fdcbc2b721e6b1b9e1b98f6de74a4fde2a09a4716e2bdcf72c65b0ba94d8d83ce02df70d12d6181174afef53336fcfa669e1cbdd3a159072c7e4232da5034778a1216242182dd7a71333dec28862306a3a10a836122f281a68363357dafd9eb91f3b57d2d722c79a41d072c9060783480a4fe32ce757edfee4ea2ba79c61c5de232479298749dbfc3ed47f54ee249f4aa7a51283603d8196c019706b8afd1f0da98fd62695ef81c405225cadae5b72acebc5afde67609f798462dc4fd512f9d37098d16919c24f0c32d3afc6346e242ad6aa4671352ee3017abcc287f0b641dc99e8ef9fa26e4733bbacf0197da3df7c907d822b212ee2e556ad4ac3da925ffb00aaf1957a0b62c92545adeaf93f8eaebed0a9c1875d073fc386c2fc15bb6d6eaa1837e24a9aa8a1542db86931191775d14af0992ba5f150b62b68014d604e67cf4bb566d8c3deaf45d8bfd52c7cf921fd8ee228022c4dff1c406fdd3d52d63bf685aa33254dc7fc0042f03d10150beed5eeb5c376cbb9b6905a160dd1ed8ba551331070238d935818d278bfdeff1042fe61cc741809f6663d66817bf14bfd3451f8c78ab7272a5e399710684f1a3c7ef43285622db81bc16c30cd192e1c284b2214819009715d7af97d3bd1a44a5457a43ea2d9b23f06ac87b83b413db35546e122a633a046e4ba143615efa3599000b38dcb3fa6724dc6c72854088439fde876e3a88f1bb11c89f45a5ed06fdd35ecde31d51ca25e12853d38bc38ba3c2b5cca418cf95fea956af7f8bd6af27831c97c2dd57f9fe3ec587ef4db32fcf5deb663f6449bc8ce44afaa1f3b7049f74c5e869b18ccb940fef96b983a56df6712e6fdc046141eb3e62fb8abae47177d20f035ab07085ef244f461b10d8505e841ec0cdcebd8998eed3e667fad496d9183e5971d004a85c4005b4c4bf02db32667b16450b3e491efb1099a0c2919de31b63744fa3ee9b9df3182946cfefb65ddd35bba4c156c84e5525d91ef09d320b164329b153178e7e77032552524d8436a18877c2d582a138e3b3581071d2965dd0dfcebf3a6067121eda4c39491a0dcb257399099b4511f383652f913c60253f6d6e9be16b21fffab2a90ecd0aa23e369a0b675afaafedee15a5558ecfef096de66f20710b2b22842ec0027ba80be114454cbb25bcd77edd727a7f56f0e58e6dc627f23e62175babba39b4cf4d3161299f24d09882082c82c1a144f098323866e15c1b79781dd7115f3a9b67b97b40284ebdb187d9d51a2f7a399221a2a3e9e2d6e23270d66078e8aef9369de80fab9bc0953215c83e91e64734623fc112c9abc4865405688b600dfeedd49f9e9666ca46dacbcff2d543781cb31d584aa0f0061bca93a700f8c284eabf2b7b7ecebb0ac66c575ab8565067b3f8914dde503c763bd16e94891703409e478ac02731e7563016ff5b38e5d3d9b80174bd3fcd0cb19214b7c15fbd2192af9a283c63cf42e7c42090e3fb6cb914d16141e28b24e6e47b154e528795d2dd59de21d942e2cb3fc154824ad5463090cfd00463daa2407b15508597d9b63fa2badeec049779292e4c4e952dc753d26c735a7f5abd532e34876c08ede94d2d0eaa1575cfb423d6fa1cae0034490e1c80680eba263d69a2c344e57f8edfd2d359f50128e0d5d17247cd3c2196b993ec5d33a99b2a54ecae1f60a9fee3a193d3a1d5ddeee6e42f093bd933625966aa779f70f9778e830f0e39d94a357c4078c467c863cf2d319aff7525dc9861658a91b157d026254f15a8d4481fdd285f60b3906d68bc514fb982a3f1511679d2d335fe5138e84d52c41402b9a565a249d483c6f2e14843adf62be9e64eb97818280b76b889b818d995db582bedd1e4fb8aea3118f7b493300f5b886cf2d724d428ceb591b2ef92acd93f9b1e9014bcab3c584e6b1c094d7de76d65eeae5a20ddeaa9c208ea702965bcb8979de730014fc64e841682278afcfc89c93feec7029cd73c43b300a345bebd91bc6a3c200c606b924a416a8a62321f7ee710ff0b3b655db046d3745819c6633e396ebf7b83ad153ea796e878e85bb6164d44304281ecce15944219ce86181d95eb97f5addb763ebcd6d90b60300ac4385df78ad10e515ba73fa15a8471d11a7cd557584c6fc1dd733faa8e6d0f34fbb0686a7b34a666f147fac5cfa77e641965472a46567f0433e133eb25d786e3b88f5bdd857458778c4ca07a0f3a92e49dfc47b47af5ba71daacf535e40436e6959fa1c482478a3a4bf19e29aa4492e36d859659da87613bd8696dfc1c7217c3932f70b21deb4da27f161150e0da491b8999007aa670b9defe7d595c3026d3f9a7a47db72dcb2b6d09da1fe55865b2aee3e004de8e404e986fb59522d4dae5288afd5311ec87f55291fc621a45ce6404729a93389697e5f2a4006b25bf79ef92ef928a8703ed42d78082b1011642353b8aea9dba4c1e2bce1adcc0a9b8e982aaef18b299982af7e992be24e593329e9aac877c437718b0e025b49d40c93a594244624a1179ef0502ae3aa6e845b8c2fce31a1111695fd63939921cd8b4b58cae6d553e7b8d9fbd297bd2512d63b45b96f2f268a51201b5d03cf41403493d6b2ce31322234cdb5a44028b04c65c09534b8e6ae7c0a2aba7ebeddb5f06e6fb944c661177e870545a3570d7baa6e469c4dbd895e81517c03521254508f39823873f47737f4f331e88e2ea7536e215490069d400cf6a8911c6b5573af3179aec96c20a6c4cf7b6b2157767254648116686be68251be326a93232abe0677e23934c4b72d8894292b674cf9c67085595e9deab4b862f71b7043c092504993429a13e31c5532a8c814fbb030a5e8fc14dfda8403bab82eeb9a58b627cdd008f6ff664bd1894987c4eb37e0e916b871199e300bd817dd0098dfc9d1b4155d18f612e137812315e4253ea4ca44fced536a63acea94458ed34296815fce552ac5c8d93be1fce4fe9f377215211787c956f87bb0e86a9b96a7a6fbb9d37b7423692336c3d4efcf7d463f549119832e7a12ead655faeb1ad3bbd3ff0588d16a4c3e07d6b4d7d6b33736e8a7038c16ec6c0182d2a07253c8e27f944ee8f638ee0ab1ae64b30960875c452b2144c11a7f6639d5c2cd00348a0c6277f18eeede0d52163ba282a19edd8d839cf734cb248bc3047596daed6e572522fb24c95734df5c4387c69e2854fc83e5d3ce37145e445eaf199dc4f2212c932b9cccfe82bb8a081639b3226323d8ebc6f75aff71c5edef8455ed4fad758abcc2e1333a5230efc90478b8975f04f0b05d3f7cfd41d9df357e7d197699e63afebab617fb8668010788d1f29e4bdc1f6f2855f2051a8bb0a37408f2c2e6aa74a92600c280177c515bd5ea9e1fd52e8f3cae36975be4c725e2a7c4bf56821690e093dc2cb824b2a4fc055e43a13cd03f029c85cfd19d46067d12694d9013f6fd4d9856f3f4074b582c90a559956eb385eb273d59f3292ef93a607243a97bf00f03667fd59ff3aed06fc4a99f0793614bf88377bd94e5fc59b7b06cc7f0402afd6edeffa049471f751b0c86be0acf64964c2864d3b2e7fe884d7d9827cd2b6861f32c76f0298c97fd75c84dd2f9f598e5ec1e96ce2474f50ffe892ca21f18e585972e319a8e1a3fcc65d1cd9aa206dafe1ee5fe58803428a165126306a35c1777e7dbdd833d8f9213727818910c82ef18e3b6cfb1d83c3c50fd1c58bc869206944e4ab2bea82cd86bb9aa736f4ea46ff3fc7e1138a3693e34d275642aa9e7ed0b8fd57e00791ffa243ed8cab5e6e7a52c8505c280e70d21900856690b3e35725f025523072f9ff260cc45fba52f9cf8f59c750a6c250b9f14bf151c56420ffa4a3d9212f73d7ead42626da73618994b6d5532bb00ff231228e35aaaff89f2ae4507cc727d4676403887bf8711afd70352245775f32c60238b869def27306e884705e0674dd9ebd5f939fbe25b213db2b4b72e5dbbb143fc0e375b8b1ae33e7fa558640025ad028247262a3bbf43bde6630934156655a2a602d0eb689888005e00110a0ac7afb51384e4fcd9012f08489b5dcbaa04a79bc3e0b28c15dd832daf82447083ce3483b6346736cf7b1fb", 0x1000}, {&(0x7f0000001f40)="b45a5a2307e1ac7e01f97c31cb42f2cb6eda15b9ad448f7494b2dfb79bed8ad68cf427b4242e", 0x26}, {&(0x7f0000001f80)="be2314fcd1293d74b2dce075f0e2918da58f7571a54d17768472821bb40a5e4c548ce8bda11922bad379a8", 0x2b}, {&(0x7f0000001fc0)="aeaa9b32c2b608f36f874397edbf4948fea371d0dc83f2a2bf1316f7335a1c80e2e419e4c2c6a21b0d18d8c441a8c7ccd5ae10d2725a6d440fa8b4b7a0a77c8501c2bf77495f3e4636e434a6a76f8d50059bc3f8bc7114cb62b2703edbf50145bd2d0ea92ae448a6828c011950b995c39f792f23c0f7ebb554208d46ebd7527b10f57a97357ec96aa2885edbdaf9e3d61028d14a63c976a4bc7d4bbd3914f9949b9006f0fd32c139b90eb3eb8433993b4cd93e8ee83f53012009cd023cc603a8099adaa39c9e7a23e02b536318fb", 0xce}, {&(0x7f00000020c0)="9ac51d1478126c0793642800ae4565197bb4cfec43ee5f65aa6fcf5fa33a00b456196df995daea32c3a16ff550a16fa5068348762ebde0e103cd5eaf3b1c63f504ec46d04187c1f0c7674a28ec618c368b4ecf3e56d5bbf31c3b256706731d064874cdc81aeb934060494abd40ced490356863cc41e9c60c604f77085998c777528723984903e2205813a77d9243b376df0f25815028ea1f8d637f300d17907a67dd074854c998224ab3d580b4bd1e3f98ea4a20312c0efbcc80bf0cee0f0ae5eee7ee9d1d023ac9a9fc3c7a97597775ce4d845d3ef5c117a15906d178d4dc0953", 0xe1}, {&(0x7f00000021c0)="e42d16c0f9100ebbac570b9eef9533cac7a68494625f335c662630b7c0236224a17e7c34", 0x24}, {&(0x7f0000002200)="5e6a3487aaaf2bd80d9902c749c134afa1bded518edad98d01c63b13a76ee753189b078073204eca69a144019e8c1ec576d6fc1f709e82b414b820884ebc18870cf4cb90f8a8e4a4e2becb5d4ee5ee1c6e57de5bb7f5ae63cc0f80132fa5f91c9b58e905aaabd8e93e6e680309c4d2cc66724b2d21e5c150216d5b0c80e0d9f7b01e28467113854adaa13c726963ef0166c6db4257e7b0eba874f3e8fd368597d94d8df975944ca0cdeb6fd7b1fb0ff47e488092cc4f6b5da48e753cb6af832167887411cda566314ddb64391e", 0xcd}], 0x7, &(0x7f0000002380)=[@rights={{0x1c, 0x1, 0x1, [r2, r0, r1]}}, @cred={{0x1c, 0x1, 0x2, {r22, r23, 0xffffffffffffffff}}}, @rights={{0x14, 0x1, 0x1, [r0]}}, @rights={{0x28, 0x1, 0x1, [r2, r1, r1, r2, r0, r1]}}], 0x80, 0x8015}}], 0x4, 0x1) 00:28:44 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x5c00) getpid() getpid() getuid() getuid() 00:28:44 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x2000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 0: sigaltstack(&(0x7f0000000240)={0x0}, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 0: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xfdffffff) 00:28:44 executing program 4: sigaltstack(&(0x7f0000000240)={0x0}, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) (async) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) (async) r1 = socket$unix(0x1, 0x5, 0x0) (async) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = getpid() (async) r4 = getpid() r5 = getuid() bind(0xffffffffffffffff, &(0x7f0000000000)=@generic={0x10, "8649ed818f7eec8299e10dea59eba6110e208a14a2cd2563e0f9f7bc9708a0e37e1c7f399d59e907410b5881d2b4996db8a4311afd451d76363092457b74a169d2036eb14b48b88ba7bcf1a8c44a8a7dfcda1cfce550ef256472e42db86d6f17932e3cfcc598dd7b3f75cb3716bfe5167219a45581662f14ee1db7098766"}, 0x80) (async) r6 = syz_init_net_socket$bt_sco(0x1f, 0x5, 0x2) r7 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000000680)={0x2, 0x4, 0x8, 0x1, 0x80, 0x1, 0x9, '\x00', 0x0, 0xffffffffffffffff, 0x1, 0x4, 0x3}, 0x48) r8 = bpf$MAP_CREATE_TAIL_CALL(0x0, &(0x7f0000000700)={0x3, 0x4, 0x4, 0xa, 0x0, 0xffffffffffffffff, 0x3, '\x00', 0x0, 0xffffffffffffffff, 0x3, 0x1, 0x4}, 0x48) (async) r9 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000780)='./binderfs2/binder-control\x00', 0x1000, 0x0) (async) r10 = geteuid() r11 = socket$inet6_mptcp(0xa, 0x1, 0x106) getsockopt$inet6_mptcp_buf(r11, 0x11c, 0x4, 0x0, 0x0) (async) socketpair$unix(0x1, 0x5, 0x0, &(0x7f0000001d40)={0xffffffffffffffff, 0xffffffffffffffff}) r13 = syz_init_net_socket$x25(0x9, 0x5, 0x0) bind$x25(r13, &(0x7f00000007c0), 0x12) r14 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000001d80)='./binderfs2/binder-control\x00', 0x0, 0x0) r15 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000001dc0)='/proc/sys/net/ipv4/vs/sync_threshold\x00', 0x2, 0x0) (async) r16 = bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000001e00)={0x2, 0x4, 0x8, 0x1, 0x80, 0xffffffffffffffff, 0x2, '\x00', 0x0, 0xffffffffffffffff, 0x2, 0x2, 0x5}, 0x48) (async) r17 = openat$loop_ctrl(0xffffffffffffff9c, &(0x7f0000002200), 0x20c0, 0x0) r18 = fsopen(&(0x7f0000000140)='sysv\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r18, 0x6, 0x0, 0x0, 0x0) r19 = getpid() (async) r20 = bpf$MAP_CREATE_RINGBUF(0x0, &(0x7f0000002240)={0x1b, 0x0, 0x0, 0x9, 0x0, 0x1, 0x0, '\x00', 0x0, 0xffffffffffffffff, 0x3, 0x5, 0x5}, 0x48) r21 = socket$inet_tcp(0x2, 0x1, 0x0) getsockopt$inet_pktinfo(r21, 0x0, 0x8, 0x0, &(0x7f0000000240)) (async) sendmmsg$unix(0xffffffffffffffff, &(0x7f0000002400)=[{{&(0x7f0000000000)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000180)=[{&(0x7f0000000080)="76d750b80a323d71ee0d15005f90850a1f697b41ef3de6407fb317ee53664e39275c43bffd5a5fa70b79f513fe35215ede2c38735abd87423404486f80784bb02dfa24b6d49a8beb2f3df2d106b81b76ddd7dc5a18c7d2f62691c71e8cd75f225911a1220945c54937b8f5e165ef6fc7fece6a6beef12635bfa39a045f59b30f91bfcd6f75634216267c72e55892c94d53bfcc657d693700a87e7afd6dcdd85126473cdf7cfaeba4a2b906210d0ec0c0430218b4a57ea63d00fb0a7ed0ac379510185d69b3cf55815b58ed953a50ca4c2ddfac758c1cfc966f86be87a5d11b9d0366b317a87d66a98c8a8834d17da0", 0xef}], 0x1, &(0x7f0000000200)=[@cred={{0x1c, 0x1, 0x2, {0x0, r5, 0xffffffffffffffff}}}], 0x20, 0x4000000}}, {{&(0x7f0000000240)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000440)=[{&(0x7f00000002c0)="dfc42b6297fb512773cb3531917c42e904320d0a31cfad4bf9e1d67de68b3b88f5b575cac5255121e9740b40", 0x2c}, {&(0x7f0000000300)="e68f09fab3244eb76bad247e82934b8db9353d8c8e43b0e02cf2df60f8a41029796a5644625cddf913ff10469a7456b9585ee998d8a43d74e03dcee2fcb9a6c280005a66ceb6a842da3474239437ba31b80834db28281dd593a08836f7681cc18089a96433f00fb73cc174ba6ad248bf576ec4d92fd9113b949cc3995c567386f72e8716ff3fcb14a7fbbde3963e335d90b4ad6980738a55ff14c600a7e5405846a058f1692eac2efb6e812ab9fd4740e4a6b8c0048e94f2efff2245cf57540efe9c637fc9efab2fea4aca02195f3138dfd0ca46", 0xd4}, {&(0x7f0000000400)="6883716d23ca9fb47d919f572147c1d447cebb5f0e8196387342115fd9152a9a163dda077722", 0x26}], 0x3, &(0x7f00000004c0)=[@rights={{0x28, 0x1, 0x1, [r2, r0, r0, r0, 0xffffffffffffffff, r2]}}], 0x28, 0x884}}, {{&(0x7f0000000500)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000640)=[{&(0x7f0000000580)="916ac906b38d404e6b309858a0bc615d50124c00ef4f8103ae2ec6a247b5f434b6fa9ad719726187cac0615a1456fe2604bf4225d6c03ae68d2e687333396aeea32e8267cccbd7c90e540389e5985edef54e7cb0fbbc9c95915d2be37ee1582e503a0e3c87623a43ea4212f88a3cbaefbcb284436ea3c6e721a6b4cb835e412a840f3ba2380b24647ff8bc6c5e29fbf93777a9ae8eeb4d34764050051331f7842175d3d6e14732e33b2ce2a97c9252e720a23ce0fbeacd0a8b4c469a3c251290", 0xc0}], 0x1, &(0x7f00000008c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @rights={{0x1c, 0x1, 0x1, [r1, r2, r2]}}, @rights={{0x18, 0x1, 0x1, [r2, r2]}}, @rights={{0x24, 0x1, 0x1, [r2, r2, 0xffffffffffffffff, r2, r0]}}, @rights={{0x28, 0x1, 0x1, [r1, 0xffffffffffffffff, r2, r6, r2, r1]}}, @rights={{0x2c, 0x1, 0x1, [r7, r2, r2, r2, r8, r9, 0xffffffffffffffff]}}], 0xd8, 0x5}}, {{0x0, 0x0, &(0x7f0000001cc0)=[{&(0x7f00000007c0)="86b6d2b7190f23598a777315a6d0754f6478cb3120b54384db723cd9441b1d21655a567a80b744df157d9efd59bd7c0929fd555ef76863007b4369ae79bdc58b484fb2059cf0d5cecc8d174e9e43dc3a916199acb957d8a7b5afce", 0x5b}, {&(0x7f00000009c0)="ed03565fac98ca237cb909d0a4956c8f932f501d85a6f1f2e3117526d383c39dfcc52c597cccc420522413ce769c4d258c71a2e3624bbfbf7c4d06e3340b69befc92ae0670859df34dc593bfb8f67c5ae077a496d8d3b9c64ab576b2f862c496d13e468c55bf8fb2305cb342f8fb1ab7d9afbe2e481621d5c42d7d2c159a", 0x7e}, {&(0x7f0000000a40)="1cef4312c27b7aa3e1e9d7d1b2604572c696f47e102b0038718ebc463c2b16597b3ecd5ec7f6170f4e3a56fa11948736cc850494eb5a37480d1850273f257204803fbfaa3be1e24755979e969002992ddb3fbaa595819f454793bbc98428f7026447da7b10028b3d0075671d5902562c09d2c652270d3a835b269abaa519489dfb92727ab24c2d6a555733a724c7299067490ea5fb7139d0abea6aeef5554827e3d28b60", 0xa4}, {&(0x7f0000000b00)="eaee41a87464a455fe9e344118a9ce26aa390da45fc0c2404667f044d78fbdf8fc8e8d88814007a325df22d058ee113f9bf524124259880baeae382963db327ff674b77b5955dee06f74c093b3e53c9ca22a0bd6d1ad0b7d19204eacf405e5b44101ccf891011689c8f6d282578c6eb9c3be5731e466cc6c652120f4a62ddb441d074b920e157aaab18cce3724e35ad0247893018b09856c045edd60ee0637468705fd10fbb90976778ab10c8795f5f8d5a08d9b5afe18c283dbe1f1d8e7008e333dfc857437b2cee8266929cf1819dc8fe5c310a8db4e11b1b3bfa57e6c3630940dd299b061758f9701735ea49b6790b145c8930ceb40e44bf5c2a08926b3b619e628886847b2d25251e027ea58d3d2da6531aa56ee216ae046eee746b49aa4f576c724f3fd039f7bd8c889f190e795a0b13d5cd676bef5a77b3ecf6822118ef8a4c6c8a7aee91c3bfecbbecf54ce16906c9a8aff1df3dba69bcd3c6078068fc0b5f882721ed398efbc04b0f7b4c4620247e50003fbbf131b77fd4208f580bb7e567d0f2861c0ade309b8f9c85ff326f993b944f4d3ac4be9f81823f7b3958a2dd8bb8cfda6c15f1cc38d257710dcfc73043113abc823ae9366cec4ce031098dabf775f4ebb448a8819d05f0203e839ffc1f942c19f4e0c4a26f3c8f1c530282bca3bbd223d6285de37b85390d8a47f3eac5efad9d5d47910d7c11d99857daee549e45f240f856dbf8bb068350ac5f8fdada1f20d913eb536992b5b37ece311d350cd4f0459444e73b30d400e53ef1dff1c1181c90d46a09a0d4b433b3ebe51ebf7e5e5809d9f4e5cf089e3cbda5ce239e20eb4bdfdf610ff0d892c283e2fb804d1140feea373f17adbdde5e52e9fb0c9cd5a5742355f1affdf01db40b4acc1f08fd066400b9fe4f23e47563019341bb0a269bf69b06c6e5b1c331bd4fadeea9ab00f1399d3b31bc0a98ccc09334f37b539ae8c390823da806a6f812a5dc41dfd3ffcc214c252eca465e925513cbdd35731f15bc51b27a9aca80e768cdd4706404b5408583c49ff98185e17ab4b9654a0bfcbc6c7a8aa67e3bb98a75ac81f2b843bbd59b3e34c4b03c2e349b38f7e534e72e43f5d3dd10849a17c884f3699f935fa321529c05f27dca0c3fb7351c873f69661762a536af50232588cac4584832bd2b6c1e33f9cdea20e6befdea2ce7097b48bf1d0c181ea431f9013097e9076c8626b77fc70cf4961df357384fc11091e357c6bfd100d533868c2239980d2942790f24d5546d4daf2c768f26c274a4d732eeca6dcfe014660310db07942a82cbf776b116e52cca76d27da0370a3116f56353b0e318dae59b78de5ec4510c30fedadbb21ab0da01297841b6635efbfc765e24000bed3d56d63a2a8f11ad1f73c2ad44119bc86c90bb364a2d7d580034199fa8f48db048833f8b13c55ca41d5c42c5b4bc86c8807c62a71e10fb736c5922fb54e6822d791dfb2d186b24dc3285913c6d133faa2c85a2219007e8c386f99bb2f3cf4b577cca94ecf69b4c0a479f33d315483b38dd8492dba86e96550c4c23089cd8f97b7b8adb6632b05745c51b3b97b13b001552f4175f487c61ecf768e246a5479083b2605e28d3b6d5f680de55aec8581c4415f800646a8d077899dd2b13812bc5b9ce66ee83e0aa4302a79354458136d447b4d5e36ea14d88ad624e10ece2b280e6563ccb5f75403429f5b15416a442ab1a47e29349473cb1be4c18cdc4705fd560a279ff298605b1be2b0b0563649810c5db8c8acaaf8d80b4feca8de9394abc929f42427e78076d1b46a9df2a80e291c8e1f48dc847d458919788afddae885d08f6388309af30b6d7c54c3d22c78622a74c336459bf7d8ee0a977f8b0422974f224b740331908d5cd33eb0e50f379531ff7764c1ade714ff57f81ab208f71b2b5df4bc05c562e0b2847c48feede38073e402bb6f9aaddf94f3b0a0259ebb70edd25e79e2f27dd6fcb7c8ac8957807d39676202a6a688afbe28db037bbb62afd42f3761b35fd25516fc0cc8a567faf5fbf77a897c532de946fe9b10e9d04ae42ec566917dc7fd1e571b4ced437a452799ddaec5abea6bd8ff53a97a4bbed5423abff7b791f6b80d4bee79e81ccf955b432ae51a675b8bcc83652f24de43dc25800da418f9ba5a721ab37bac1f49a502fe96d1624c01eaf01ae224c0d32db73fb7cbe8d7d6c171c3564c296638b9fb939271a2fa0c4b0128fb2c43baa5122c373600e226d9ed8c12a9e646c153cc1dde1d182dffd8a508650d6015d8aa75581399ba327015fd03b156d51b2476fed164a333241f2de6d419177716abfe4c56055704c3b48df64cfaa2bd4cc366f7d1b612b62d8092e271090f6c5183dcc64b36f50511b0d10f5a4953381c1834cf3e153cad78641c394937a6dabf3d5272f792acb3a430eca859a24698b1548cfc814b3a5d07dc85048d1beeceaa715798f34d62bfe39152e757cb400c380366d03f9a84993ffefd3ac73096e8c02a212e8bf0f72a99cd5971ffbbf7fc9a747cf1b021a1cd6bdb020e93e5483f958a7b4b72b36761628659a0052d9076f8652b175869e336f41c6b498c8ca717d906ae3eb7529eba1095936da71c0bacdd9af0ee943894eec88ed1d2c69861958f2555e778e608123d55e7591cda1a78fe0881b89f74a2b9020ef493438ccbcf6019ff5e83a2f407bc54b920f241966b3be9d777af817a1b3b095085ce3a043a3cd7c7888e20d85e2a30b12370cddd4e1c3692145b7d366d6ad35db9245aa04b52cf212ceb6a7331697d2a7e19fbbb2223bd2b35f36a76af6f14ca9a87ee95f738d7f3476d6f1e98d6e9ce02db8ad3b67a33c8d8152cb5c5c18a37da9616f31c0fb4bfc88014c9bcfbf25ecf8fdd2ca3b166921c2486e871b31621865b46ba920c8445c7d5dee7075ee8a2ff2dea00952cd804c1730b9115260e71c479048039f0442301599328dbed86132357d77b2100f5d5449afb7c83d578f01e392a823ad0bbddcd1414463907c1482b44e04bdd42ae6ffaea767dd9748957d24936bd3e38d13638f756408a90d9d066579ad9f53599d2fbf216139702c8a0bb49ed6ed926e0fc04aee5815c4114f6dd9777214bfc8fa4e467ac38065e03d5b9887d2df227c6abea27f53768dd80d1606a7a30bc70c8cdf2908f9e97509763fe983173d2ce210a813b4b6f3a64fd0a16942aabe2ba9ecfc4fd80254c1869cc3f0ceaf6243a57f9140f0b32c67593999223e94e0f354484fddebdc14120172ca70736b8ece004d88aaed13a8da2a0bf7a28a1df6ce79881c6130f3ceee3e4ec301b1bb8ab068471f80fd6cb75a679f5c98fb910b93d5614639bdd2d68d9eac0db54edc03c4b21246135922516512ee9f0da31261caf4dac8deeb5e2b11409cfdb108d1fe068bdfae1ae51d7df60ce2a5cffdaacb5a464e88ea2db8c34a7fc225cf3feb1b2fbe4d88a4d90d0242c953648d7549a4b7416cf903d43d349eb12e6bc75f80db07052771ccb6fbe755b455cf193a3f52cfba87f4b5fc9ae3a77d11414ed364991d57027509584965aec0ab756402e9caebba599b902b566aa43b482d8519965e7ba82275ee5d6df4f7cca8c3c0de939125bcd6b559145789d99d34cbeedc335a0df5770706851ee2bbf4b9646a3c259dbaa0b39ccf42580aaa26c3a4669be981743dadfee3532a813d8b2e7c8baeed233dece5c8724c49b8cffde538bdf83d6206efd1f9577649a285b37ba50b9809172a2dd3e48e81d04d92c72b5443794c333a1f414826cd5bd419ab709fcc611224bb96d7e6ded3f67feedd292a3607f149cf9c7654221cab67ababb207a346e1fb2a8d1287285ca14a69c1db46f72598f6a0751e4e4a4837cda0a333e2a9dc1d54f5fdf4c6192494933ffb317a62850e22efd31e176a19f543f1348376bafc72b7f2aa86c1967a81b98238dc50144af08c38fb7aa479e8411a8bf3991195073e461775f678e0a1edc3e53528415ba3bdc60d807b8934ac8ad551e6a34b7391d80a669ac4a01ff06e93fc223492f77179527fdb48d7417645fe4a7ac83cf7114a153b6139fdd6f5b7ec07da5f2f3b3b0120c8ad8e26b92578b1120997077752d930b64f3fbfd5239a6ba4087dfc186716d71777b68ec1a22aba67aeceddde61600aacb22bb2892a8c4a12587e724187ebc28d02a658d43b8b40e77e39c225d713952f30e2b00204e3290657b967468f5aa6bd3cd325d04abe165d5860e8ea099763eb7b4dfd299873591e4da0f4fb7bda13e0d59c1fbb36a28aa01290f1680d4068692e6bf83522ed4fb3c36a52a6a617a27630232b9a937cd482e59d59d55defb7cda5f48e78a72cdfcbba5f607125d2355b3ee7b33fec9943884af4417b9b067d7095852175cf9022527327f22ee82ae551690fddf869660d54a7a798a5d14c550cc966d90297d0fa3992985d468d6fc5bd6dd7184b34faafa8b0ae2ffdcecc26ed9f62d74b00bf424e46d575d3baace7262d563634e41646506d7cd8284df4d177db0475c119acd4d12b533579c7f2e4485ab11f15a3abc9c530c4fa113eab19bf72cb28a52de45b4eccb6e7ba53391a84f12a3a265ba778ec81fcb50f5f9684d4912161a7b08840ed88acb452364cc1833cc364ba856c8cf49eed7d560da085bab8cc7ae3b1a69167d94cd6f969471004ef728e04e08a88eb65cb0d9998a1323f88b75b784bede6a65e2c848f7e44802cebe43ed6b074cd830f9d093c8e8ad833400500e7c48fc93eb9c556c689bde6408ebcff4b7b138166e5f9e2ee91a02d8fe82aa6488a12f83630f7bbd3a66d7550877dd0c143b26cedd2b43d5474d3873bb3d6c7ba3908dfd4d0f1f9c3ab94d9b759189773c5e60128f1924907bbb51381eadd6636a388126884f06459919074acb0a255ec1717f74fa9584a46c3dd2e4572489814cd8fb394145ff025226f73c73da9844e1ddef654d5d42e4a9393ff69b4a0ff6db0c75b715a7e152c2626adfc8b1fe303d26aefffac667b15c240f973aeef4c9f46cf8ed563013ef541d02d986101a0a66f029e1f919b72ae1d0e6650e79f8492f81990d094dd018cf44f5e7f02be97a626523b3f0737169e16d2300b6ad91dd74a09831b2404987506a26bbddf5fb0379a5f3fdbdfab2d6416f21704ac86f21228bac744c138a088f644165db313c539c55ea5ddc38add09b35e2b1b5515e687ee137e173d33255336ba2ec6236706b4537c811ad9127d94ca19056f3c7a9e6703f8488450f3b1313291167c81b8c5cf9ca0dcba002da64595f6ad3e1399b4fdcd014199e7f6aed7b6899a8c2c9a93cd0f7d386432345dcef1594d132a14e494fd602d4d2a4bb7e8f77aba2cbfead5e383f717b8ee32a66310f7ae890100e10cb0a74c1b42d1d04308a6894434cdf4f04b0094bfaa5d226d906b8fa05389a6ce9aba416ad68955d5e73783577b21b5e04a60f612dd59dc61137913ef906456482e774fc3dbbb994cead8656cdedf939343c6fe11f8916a96159c5d148a4a4371ee3bb70928a6a496bb26f706873a8bd2acb1a62d1fcec2392a3873af151e9c1127abf4f40a580a4b5e437b06ff93eca6fcbb968db4467fa3de1efd7ecb245edb441cbb4ad03fd90742a059941c2bb6097a956f2e8e33e252d15020fea728014fe284e71db1bc014ca30b5734bbe46af8b039ff144e155b23ab4dc54e62e454ccd5ef5209663c07f953cb10963bc188ba8b0e991eef7343639e0afc94fa2611796ba66facb9428c47f385a45f3b56ecf814a7f6e5739d28941036ec0b9cc", 0x1000}, {&(0x7f0000001b00)="ffea4b91faa8254cd0422cd49e7d7b4ad911f174f04d3e969d69138087a9f76a795e25090f275f", 0x27}, {&(0x7f0000001b40)="59f0f837838a31e66ec2a56d15b9dd45c044c2b6eda6", 0x16}, {&(0x7f0000001b80)="1232786ac764324ce162ae0b2747c4b58a49b54c1a51187ee961bcd83fc3d824b0a215bc67888a7765dacb4b27352eb0c3b20f6212d8e5ee5e01d6b26a08fc6ca824cbb764639dcccde8b6bf92a73f89af60d93eff969b5061b1e4deeb431ea20572eccd896e7bde4027c61c6ae9862baa89ec5c4aade26ac4aa27c491c3e513c47637891edda03f8b579a8a84966fa2b1bd3c0d202693a5d00ca289ca2c392ea1290a48604a4878e5dcff5be2a48606fda6f98124ea0c3b7f8e054ba096631dedc1264df2", 0xc5}, {&(0x7f0000001c80)="8962099d22fb862d6fd6d6973ae1a275cd22e7cc524cade4962f6edcb6414035122de5e237dcf5ec083475596fa365a7", 0x30}], 0x8, &(0x7f0000001e80)=[@rights={{0x28, 0x1, 0x1, [r2, r0, r2, r0, r1, r0]}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @rights={{0x24, 0x1, 0x1, [r0, r0, r0, r2, r1]}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, r10, 0xffffffffffffffff}}}, @rights={{0x30, 0x1, 0x1, [r0, r0, r11, r1, r2, r12, r13, r1]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xffffffffffffffff}}}, @rights={{0x2c, 0x1, 0x1, [r1, r0, r14, r1, r2, 0xffffffffffffffff, r2]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, r5, 0xee00}}}, @rights={{0x20, 0x1, 0x1, [r15, r1, r16, r0]}}], 0x170, 0x4000000}}, {{&(0x7f0000002000)=@abs={0x1, 0x0, 0x4e22}, 0x6e, &(0x7f00000021c0)=[{&(0x7f0000002080)="d7d7b9a6bd47a22a6161af0fa6101c1a79804ba49b11e35f34aa94f92990ceec01ef5c48d5d0a69996ff7a6a460bd98db69b6117248abbdbd02209c218e08448c5e77fd8df7fafb32f0429810b7550be7550ed7970da86f80982c601f8e95e655587efaeba64e9e2", 0x68}, {&(0x7f0000002100)="b2f97dc09b90d141535f9ac9d035343d527176653676c4d8fd8086add2f762a814d75f670d43b3ecbb8198fbd38618ed3778652819952126e1e620359e9b", 0x3e}, {&(0x7f0000002140)="136e38778c34c4758567b97dd90283afd67de25a112f3ec344a1f209207d8929394f2467e5e56ba3971838b476b2a03494fe667e98aa21d58fd0cdd8f3174f33759915d73b6c4870a60f16cb2d3b69cf0708ec8a9ce4543ee0b5acc38d9e741a6c8af9b937818e597c333b85a806ac36a7d445722e6ae67682", 0x79}], 0x3, &(0x7f00000022c0)=[@cred={{0x1c, 0x1, 0x2, {r4, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee00}}}, @rights={{0x2c, 0x1, 0x1, [r0, r2, r2, r17, r0, r1, r18]}}, @cred={{0x1c, 0x1, 0x2, {r19, r5, 0xee00}}}, @rights={{0x34, 0x1, 0x1, [r1, r2, r1, r0, r20, r1, r2, r2, 0xffffffffffffffff]}}, @rights={{0x14, 0x1, 0x1, [r21]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5}}}], 0x120}}], 0x5, 0x0) (async) getuid() 00:28:44 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x2000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 2: r0 = fsopen(&(0x7f0000000000)='nfs4\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:44 executing program 0: sigaltstack(&(0x7f0000000240)={0x0}, &(0x7f00000002c0)={&(0x7f0000000280)}) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 1: r0 = syz_init_net_socket$x25(0x9, 0x5, 0x0) ioctl$SIOCX25GCALLUSERDATA(r0, 0x8919, 0x0) bind(r0, &(0x7f0000000040)=@nl=@unspec, 0x80) socketpair(0x21, 0x1, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) accept4$x25(r1, &(0x7f00000000c0), &(0x7f0000000100)=0x12, 0x80800) sendmsg$unix(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) syz_init_net_socket$x25(0x9, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:44 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xffff1f00) 00:28:44 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x3000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 3: r0 = fsopen(&(0x7f0000000000)='nfs4\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:44 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) getpid() getpid() getuid() getuid() ioctl$BINDER_CTL_ADD(r0, 0xc1086201, &(0x7f0000000000)={'binder0\x00'}) 00:28:44 executing program 2: sigaltstack(&(0x7f0000000240)={0x0, 0x80000002}, 0x0) sigaltstack(&(0x7f0000000280)={&(0x7f0000000300)=""/168, 0x3, 0xa8}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x4000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 4: r0 = fsopen(&(0x7f0000000000)='nfs4\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:44 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xffffff7f) 00:28:44 executing program 2: r0 = syz_init_net_socket$x25(0x9, 0x5, 0x0) ioctl$SIOCX25GCALLUSERDATA(r0, 0x8919, 0x0) bind(r0, &(0x7f0000000040)=@nl=@unspec, 0x80) socketpair(0x21, 0x1, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) accept4$x25(r1, &(0x7f00000000c0), &(0x7f0000000100)=0x12, 0x80800) sendmsg$unix(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) syz_init_net_socket$x25(0x9, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) syz_init_net_socket$x25(0x9, 0x5, 0x0) (async) ioctl$SIOCX25GCALLUSERDATA(r0, 0x8919, 0x0) (async) bind(r0, &(0x7f0000000040)=@nl=@unspec, 0x80) (async) socketpair(0x21, 0x1, 0x0, &(0x7f0000000000)) (async) accept4$x25(r1, &(0x7f00000000c0), &(0x7f0000000100)=0x12, 0x80800) (async) sendmsg$unix(r2, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) (async) syz_init_net_socket$x25(0x9, 0x5, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) 00:28:44 executing program 3: sigaltstack(&(0x7f0000000240)={0x0, 0x80000002}, 0x0) sigaltstack(0x0, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x6000) getpid() getpid() getuid() getuid() 00:28:44 executing program 0: r0 = fsopen(&(0x7f0000000000)='nfs4\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:44 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x5000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 1: sigaltstack(&(0x7f0000000240)={0x0, 0x80000002}, 0x0) sigaltstack(0x0, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) getpid() getpid() getuid() getuid() ioctl$BINDER_CTL_ADD(r0, 0xc1086201, &(0x7f0000000000)={'binder0\x00'}) sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) (async) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) (async) socket$unix(0x1, 0x5, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) getpid() (async) getpid() (async) getuid() (async) getuid() (async) ioctl$BINDER_CTL_ADD(r0, 0xc1086201, &(0x7f0000000000)={'binder0\x00'}) (async) 00:28:44 executing program 1: r0 = socket$nl_generic(0x10, 0x3, 0x10) r1 = syz_genetlink_get_family_id$ethtool(&(0x7f0000000080), 0xffffffffffffffff) ioctl$ifreq_SIOCGIFINDEX_vcan(0xffffffffffffffff, 0x8933, &(0x7f00000000c0)={'vcan0\x00', 0x0}) r3 = socket$can_j1939(0x1d, 0x2, 0x7) ioctl$ifreq_SIOCGIFINDEX_vcan(r3, 0x8933, &(0x7f0000000000)={'vxcan1\x00', 0x0}) bind$can_j1939(r3, &(0x7f0000000080)={0x1d, r4}, 0x18) bind$can_j1939(r3, &(0x7f00000002c0)={0x1d, r4, 0x2, {0x2, 0xff, 0x1}, 0x1}, 0x18) ioctl$sock_ipv6_tunnel_SIOCCHGTUNNEL(0xffffffffffffffff, 0x89f3, &(0x7f0000000180)={'syztnl1\x00', &(0x7f0000000100)={'ip6tnl0\x00', 0x0, 0x29, 0x9, 0x87, 0x4, 0x40, @loopback, @remote, 0x8000, 0x80, 0x101, 0x100}}) sendmsg$ETHTOOL_MSG_LINKSTATE_GET(r0, &(0x7f0000000280)={&(0x7f0000000040)={0x10, 0x0, 0x0, 0x8000000}, 0xc, &(0x7f0000000240)={&(0x7f00000001c0)={0x70, r1, 0x100, 0x70bd28, 0x25dfdbfe, {}, [@HEADER={0x5c, 0x1, 0x0, 0x1, [@ETHTOOL_A_HEADER_DEV_NAME={0x14, 0x2, 'veth1_to_batadv\x00'}, @ETHTOOL_A_HEADER_DEV_NAME={0x14, 0x2, 'caif0\x00'}, @ETHTOOL_A_HEADER_DEV_INDEX={0x8}, @ETHTOOL_A_HEADER_DEV_INDEX={0x8, 0x1, r2}, @ETHTOOL_A_HEADER_FLAGS={0x8, 0x3, 0x6}, @ETHTOOL_A_HEADER_DEV_INDEX={0x8, 0x1, r4}, @ETHTOOL_A_HEADER_FLAGS={0x8, 0x3, 0x2}, @ETHTOOL_A_HEADER_DEV_INDEX={0x8, 0x1, r5}]}]}, 0x70}, 0x1, 0x0, 0x0, 0x20000000}, 0x40) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r6, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:44 executing program 0: sigaltstack(&(0x7f0000000240)={0x0, 0x80000002}, 0x0) sigaltstack(0x0, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 3: fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:44 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x6000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 0: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xffffff97) 00:28:44 executing program 2: sigaltstack(&(0x7f0000000240)={0x0, 0x80000002}, 0x0) sigaltstack(&(0x7f0000000280)={0x0, 0x3}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r0 = socket$unix(0x1, 0x5, 0x0) connect$unix(r0, &(0x7f0000000000)=@abs={0x0, 0x0, 0x4e23}, 0x6e) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) getpid() getpid() getuid() fsmount(0xffffffffffffffff, 0x1, 0x14) getuid() socketpair$unix(0x1, 0x5, 0x0, &(0x7f0000000080)={0xffffffffffffffff}) recvmsg$unix(r1, &(0x7f0000000280)={&(0x7f00000000c0), 0x6e, &(0x7f0000000240)=[{&(0x7f0000000140)=""/226, 0xe2}], 0x1}, 0x10101) 00:28:44 executing program 2: fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:44 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x6800) getpid() getpid() getuid() getuid() 00:28:44 executing program 0: sigaltstack(&(0x7f0000000240)={0x0, 0x80000002}, 0x0) sigaltstack(&(0x7f0000000280)={0x0}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 4: r0 = socket$nl_generic(0x10, 0x3, 0x10) r1 = syz_genetlink_get_family_id$ethtool(&(0x7f0000000080), 0xffffffffffffffff) (async) ioctl$ifreq_SIOCGIFINDEX_vcan(0xffffffffffffffff, 0x8933, &(0x7f00000000c0)={'vcan0\x00', 0x0}) (async) r3 = socket$can_j1939(0x1d, 0x2, 0x7) ioctl$ifreq_SIOCGIFINDEX_vcan(r3, 0x8933, &(0x7f0000000000)={'vxcan1\x00', 0x0}) bind$can_j1939(r3, &(0x7f0000000080)={0x1d, r4}, 0x18) bind$can_j1939(r3, &(0x7f00000002c0)={0x1d, r4, 0x2, {0x2, 0xff, 0x1}, 0x1}, 0x18) (async) ioctl$sock_ipv6_tunnel_SIOCCHGTUNNEL(0xffffffffffffffff, 0x89f3, &(0x7f0000000180)={'syztnl1\x00', &(0x7f0000000100)={'ip6tnl0\x00', 0x0, 0x29, 0x9, 0x87, 0x4, 0x40, @loopback, @remote, 0x8000, 0x80, 0x101, 0x100}}) sendmsg$ETHTOOL_MSG_LINKSTATE_GET(r0, &(0x7f0000000280)={&(0x7f0000000040)={0x10, 0x0, 0x0, 0x8000000}, 0xc, &(0x7f0000000240)={&(0x7f00000001c0)={0x70, r1, 0x100, 0x70bd28, 0x25dfdbfe, {}, [@HEADER={0x5c, 0x1, 0x0, 0x1, [@ETHTOOL_A_HEADER_DEV_NAME={0x14, 0x2, 'veth1_to_batadv\x00'}, @ETHTOOL_A_HEADER_DEV_NAME={0x14, 0x2, 'caif0\x00'}, @ETHTOOL_A_HEADER_DEV_INDEX={0x8}, @ETHTOOL_A_HEADER_DEV_INDEX={0x8, 0x1, r2}, @ETHTOOL_A_HEADER_FLAGS={0x8, 0x3, 0x6}, @ETHTOOL_A_HEADER_DEV_INDEX={0x8, 0x1, r4}, @ETHTOOL_A_HEADER_FLAGS={0x8, 0x3, 0x2}, @ETHTOOL_A_HEADER_DEV_INDEX={0x8, 0x1, r5}]}]}, 0x70}, 0x1, 0x0, 0x0, 0x20000000}, 0x40) (async) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r6, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:44 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x7000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 0: fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:44 executing program 3: sigaltstack(&(0x7f0000000240)={0x0, 0x80000002}, 0x0) sigaltstack(&(0x7f0000000280)={0x0}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xffffffeb) 00:28:44 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x8000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 0: socketpair(0x21, 0x3, 0xfffffffd, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:44 executing program 1: r0 = fsopen(0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:44 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) (async) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) (async) r0 = socket$unix(0x1, 0x5, 0x0) connect$unix(r0, &(0x7f0000000000)=@abs={0x0, 0x0, 0x4e23}, 0x6e) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) getpid() getpid() (async) getuid() (async) fsmount(0xffffffffffffffff, 0x1, 0x14) getuid() socketpair$unix(0x1, 0x5, 0x0, &(0x7f0000000080)={0xffffffffffffffff}) recvmsg$unix(r1, &(0x7f0000000280)={&(0x7f00000000c0), 0x6e, &(0x7f0000000240)=[{&(0x7f0000000140)=""/226, 0xe2}], 0x1}, 0x10101) 00:28:44 executing program 4: sigaltstack(&(0x7f0000000240)={0x0, 0x80000002}, 0x0) sigaltstack(&(0x7f0000000280)={0x0}, &(0x7f0000000440)={&(0x7f00000003c0)=""/88, 0x0, 0x58}) 00:28:44 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x9000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 2: sigaltstack(&(0x7f0000000240)={0x0, 0x80000002}, 0x0) sigaltstack(&(0x7f0000000280)={0x0, 0x3}, 0x0) 00:28:44 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xfffffffd) 00:28:44 executing program 4: r0 = fsopen(0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:44 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0xa000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:44 executing program 2: socketpair(0x21, 0x3, 0xfffffffd, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) socketpair(0x21, 0x3, 0xfffffffd, &(0x7f0000000000)) (async) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) 00:28:44 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x6c00) getpid() getpid() getuid() getuid() 00:28:44 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r1 = socket$unix(0x1, 0x5, 0x0) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = getpid() getpid() getuid() r4 = getuid() r5 = socket$unix(0x1, 0x1, 0x0) bind$unix(r5, &(0x7f0000000200)=@abs={0x1}, 0x6e) sendmsg$unix(r1, &(0x7f0000000180)={&(0x7f0000000000)=@abs={0x1, 0x0, 0x4e21}, 0x6e, &(0x7f00000000c0)=[{&(0x7f0000000080)="9e94da045afe476ca7", 0x9}], 0x1, &(0x7f0000000100)=[@rights={{0x1c, 0x1, 0x1, [0xffffffffffffffff, r2, r5]}}, @cred={{0x1c, 0x1, 0x2, {r3, r4}}}, @rights={{0x1c, 0x1, 0x1, [r0, r1, r1]}}], 0x60, 0xc000}, 0x4000005) 00:28:45 executing program 1: r0 = fsopen(0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:45 executing program 0: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x100000000000) 00:28:45 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0xb000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:45 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) ioctl$SIOCX25SCAUSEDIAG(r0, 0x89ec, &(0x7f0000000040)={0xb1, 0xf}) ioctl$SIOCX25SCAUSEDIAG(r1, 0x89ec, &(0x7f0000000180)={0x0, 0x5}) ioctl$sock_x25_SIOCADDRT(r0, 0x890b, &(0x7f0000000080)={@remote={'\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc', 0x2}, 0xa, 'wg1\x00'}) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:45 executing program 4: fsopen(&(0x7f0000000000)='nfs4\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:45 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r1 = socket$unix(0x1, 0x5, 0x0) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = getpid() getpid() getuid() r4 = getuid() r5 = socket$unix(0x1, 0x1, 0x0) bind$unix(r5, &(0x7f0000000200)=@abs={0x1}, 0x6e) sendmsg$unix(r1, &(0x7f0000000180)={&(0x7f0000000000)=@abs={0x1, 0x0, 0x4e21}, 0x6e, &(0x7f00000000c0)=[{&(0x7f0000000080)="9e94da045afe476ca7", 0x9}], 0x1, &(0x7f0000000100)=[@rights={{0x1c, 0x1, 0x1, [0xffffffffffffffff, r2, r5]}}, @cred={{0x1c, 0x1, 0x2, {r3, r4}}}, @rights={{0x1c, 0x1, 0x1, [r0, r1, r1]}}], 0x60, 0xc000}, 0x4000005) sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) (async) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) (async) socket$unix(0x1, 0x5, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) getpid() (async) getpid() (async) getuid() (async) getuid() (async) socket$unix(0x1, 0x1, 0x0) (async) bind$unix(r5, &(0x7f0000000200)=@abs={0x1}, 0x6e) (async) sendmsg$unix(r1, &(0x7f0000000180)={&(0x7f0000000000)=@abs={0x1, 0x0, 0x4e21}, 0x6e, &(0x7f00000000c0)=[{&(0x7f0000000080)="9e94da045afe476ca7", 0x9}], 0x1, &(0x7f0000000100)=[@rights={{0x1c, 0x1, 0x1, [0xffffffffffffffff, r2, r5]}}, @cred={{0x1c, 0x1, 0x2, {r3, r4}}}, @rights={{0x1c, 0x1, 0x1, [r0, r1, r1]}}], 0x60, 0xc000}, 0x4000005) (async) 00:28:45 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0xc000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:45 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x7400) getpid() getpid() getuid() getuid() 00:28:45 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xffffaee5ce68) 00:28:45 executing program 3: fsopen(&(0x7f0000000000)='nfs4\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) [ 1725.200976][T20329] Bluetooth: hci3: command 0x0405 tx timeout 00:28:45 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0xd000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:45 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) ioctl$SIOCX25SCAUSEDIAG(r0, 0x89ec, &(0x7f0000000040)={0xb1, 0xf}) ioctl$SIOCX25SCAUSEDIAG(r1, 0x89ec, &(0x7f0000000180)={0x0, 0x5}) ioctl$sock_x25_SIOCADDRT(r0, 0x890b, &(0x7f0000000080)={@remote={'\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc', 0x2}, 0xa, 'wg1\x00'}) (async) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:45 executing program 1: fsopen(&(0x7f0000000000)='nfs4\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:45 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0xe000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:45 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x1000000000000) 00:28:45 executing program 4: r0 = fsopen(&(0x7f0000000000)='nfs4\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:45 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0xf000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:45 executing program 2: bind$x25(0xffffffffffffffff, &(0x7f0000000040)={0x9, @remote={'\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc', 0x3}}, 0x12) r0 = syz_init_net_socket$x25(0x9, 0x5, 0x0) ioctl$SIOCX25GCALLUSERDATA(r0, 0x8906, 0x0) sendto$x25(r0, &(0x7f00000001c0)="168aa50caad928bfd87dee54fa71f63aba336a55", 0x14, 0x80, &(0x7f0000000200), 0x12) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) socketpair(0x18, 0x2, 0xffff0001, &(0x7f0000000080)={0xffffffffffffffff}) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) sendto$x25(r2, &(0x7f00000000c0)="5873c311103a2ce79f75cb2265713c5cb4f26a5151ac28a09840a6bcfa3759ef81071fcd66cab425a2075367045f3a108cea11735146b40c4a8c0071d30d5d8960d21a35ea22e019c484267ae9c6ba5a58a8df779a95d0287f04243e34d8f4fc3a360a7bf5e5cea804ee9596237c9374c5cc7ec27df6ea3ea1a9ad869e277241711c40cd092f9449d8a46086623297c5f8b73fe5b917069d5942b06e5c5f1c2c4901b63cc2eb8eed", 0xa8, 0x1, &(0x7f0000000180)={0x9, @remote={'\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc', 0x3}}, 0x12) 00:28:45 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r0 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000000)='/proc/sys/net/ipv4/vs/secure_tcp\x00', 0x2, 0x0) accept4$unix(r0, 0x0, 0x0, 0xc0000) r1 = openat$ipvs(0xffffffffffffff9c, &(0x7f00000000c0)='/proc/sys/net/ipv4/vs/sync_refresh_period\x00', 0x2, 0x0) write$smackfs_ipv6host(r1, 0x0, 0x0) r2 = bpf$MAP_CREATE_TAIL_CALL(0x0, &(0x7f0000000280)={0x3, 0x4, 0x4, 0xa, 0x0, 0x1, 0xf61f, '\x00', 0x0, 0xffffffffffffffff, 0x5, 0x4, 0x4}, 0x48) r3 = bpf$MAP_CREATE_RINGBUF(0x0, &(0x7f0000000300)={0x1b, 0x0, 0x0, 0x9, 0x0, 0xffffffffffffffff, 0x84e1, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x2, 0x4}, 0x48) r4 = bpf$MAP_CREATE_RINGBUF(0x0, &(0x7f0000000380)={0x1b, 0x0, 0x0, 0x2, 0x0, 0xffffffffffffffff, 0x4, '\x00', 0x0, 0xffffffffffffffff, 0x4, 0x4, 0x1}, 0x48) r5 = bpf$MAP_CREATE(0x0, &(0x7f0000000400)=@bloom_filter={0x1e, 0x4, 0x3, 0x8, 0x400, 0x1, 0x80000001, '\x00', 0x0, 0xffffffffffffffff, 0x3, 0x2, 0x5, 0x1}, 0x48) bpf$BPF_PROG_WITH_BTFID_LOAD(0x5, &(0x7f00000004c0)=@bpf_tracing={0x1a, 0x1d, &(0x7f0000000000)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0x4, 0x0, 0x0, 0x0, 0x2}, {{0x18, 0x1, 0x1, 0x0, r0}}, {}, [@initr0={0x18, 0x0, 0x0, 0x0, 0x6, 0x0, 0x0, 0x0, 0x4}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, r1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x400}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x1}}, @exit, @cb_func={0x18, 0xa, 0x4, 0x0, 0xfffffffffffffffe}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x85}}}, &(0x7f0000000100)='syzkaller\x00', 0x1, 0x8d, &(0x7f0000000140)=""/141, 0x41100, 0x53, '\x00', 0x0, 0x19, 0xffffffffffffffff, 0x8, &(0x7f0000000200)={0x6, 0x1}, 0x8, 0x10, &(0x7f0000000240)={0x2, 0x0, 0x8f, 0x2574}, 0x10, 0x11660, 0xffffffffffffffff, 0x0, &(0x7f0000000480)=[r2, r3, r4, r5], 0x0, 0x10, 0xb3d}, 0x90) getpid() getpid() getuid() getuid() 00:28:45 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x7800) getpid() getpid() getuid() getuid() 00:28:45 executing program 3: fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) r0 = fsopen(&(0x7f0000000000)='binfmt_misc\x00', 0x0) fsmount(r0, 0x0, 0x72) r1 = accept$alg(0xffffffffffffffff, 0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000040)='\'\x00', 0x0, r1) 00:28:45 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x10000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:45 executing program 1: bind$x25(0xffffffffffffffff, &(0x7f0000000040)={0x9, @remote={'\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc', 0x3}}, 0x12) r0 = syz_init_net_socket$x25(0x9, 0x5, 0x0) ioctl$SIOCX25GCALLUSERDATA(r0, 0x8906, 0x0) sendto$x25(r0, &(0x7f00000001c0)="168aa50caad928bfd87dee54fa71f63aba336a55", 0x14, 0x80, &(0x7f0000000200), 0x12) (async) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) (async) socketpair(0x18, 0x2, 0xffff0001, &(0x7f0000000080)={0xffffffffffffffff}) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) sendto$x25(r2, &(0x7f00000000c0)="5873c311103a2ce79f75cb2265713c5cb4f26a5151ac28a09840a6bcfa3759ef81071fcd66cab425a2075367045f3a108cea11735146b40c4a8c0071d30d5d8960d21a35ea22e019c484267ae9c6ba5a58a8df779a95d0287f04243e34d8f4fc3a360a7bf5e5cea804ee9596237c9374c5cc7ec27df6ea3ea1a9ad869e277241711c40cd092f9449d8a46086623297c5f8b73fe5b917069d5942b06e5c5f1c2c4901b63cc2eb8eed", 0xa8, 0x1, &(0x7f0000000180)={0x9, @remote={'\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc\xcc', 0x3}}, 0x12) 00:28:45 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000000000) 00:28:45 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x7a00) getpid() getpid() getuid() getuid() 00:28:45 executing program 2: fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) (async) r0 = fsopen(&(0x7f0000000000)='binfmt_misc\x00', 0x0) fsmount(r0, 0x0, 0x72) (async) r1 = accept$alg(0xffffffffffffffff, 0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) (async) fsconfig$FSCONFIG_SET_FD(r0, 0x5, &(0x7f0000000040)='\'\x00', 0x0, r1) 00:28:45 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x11000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:45 executing program 1: fsopen(&(0x7f0000000000)='ceph\x00', 0x0) r0 = fsopen(&(0x7f0000000040)='virtiofs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:45 executing program 0: socketpair(0x27, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) socket$unix(0x1, 0x1, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:45 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) (async) sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) (async) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) openat$ipvs(0xffffffffffffff9c, &(0x7f0000000000)='/proc/sys/net/ipv4/vs/secure_tcp\x00', 0x2, 0x0) (async) r0 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000000)='/proc/sys/net/ipv4/vs/secure_tcp\x00', 0x2, 0x0) accept4$unix(r0, 0x0, 0x0, 0xc0000) r1 = openat$ipvs(0xffffffffffffff9c, &(0x7f00000000c0)='/proc/sys/net/ipv4/vs/sync_refresh_period\x00', 0x2, 0x0) write$smackfs_ipv6host(r1, 0x0, 0x0) r2 = bpf$MAP_CREATE_TAIL_CALL(0x0, &(0x7f0000000280)={0x3, 0x4, 0x4, 0xa, 0x0, 0x1, 0xf61f, '\x00', 0x0, 0xffffffffffffffff, 0x5, 0x4, 0x4}, 0x48) r3 = bpf$MAP_CREATE_RINGBUF(0x0, &(0x7f0000000300)={0x1b, 0x0, 0x0, 0x9, 0x0, 0xffffffffffffffff, 0x84e1, '\x00', 0x0, 0xffffffffffffffff, 0x0, 0x2, 0x4}, 0x48) r4 = bpf$MAP_CREATE_RINGBUF(0x0, &(0x7f0000000380)={0x1b, 0x0, 0x0, 0x2, 0x0, 0xffffffffffffffff, 0x4, '\x00', 0x0, 0xffffffffffffffff, 0x4, 0x4, 0x1}, 0x48) r5 = bpf$MAP_CREATE(0x0, &(0x7f0000000400)=@bloom_filter={0x1e, 0x4, 0x3, 0x8, 0x400, 0x1, 0x80000001, '\x00', 0x0, 0xffffffffffffffff, 0x3, 0x2, 0x5, 0x1}, 0x48) bpf$BPF_PROG_WITH_BTFID_LOAD(0x5, &(0x7f00000004c0)=@bpf_tracing={0x1a, 0x1d, &(0x7f0000000000)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0x4, 0x0, 0x0, 0x0, 0x2}, {{0x18, 0x1, 0x1, 0x0, r0}}, {}, [@initr0={0x18, 0x0, 0x0, 0x0, 0x6, 0x0, 0x0, 0x0, 0x4}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, r1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x400}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x1}}, @exit, @cb_func={0x18, 0xa, 0x4, 0x0, 0xfffffffffffffffe}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x85}}}, &(0x7f0000000100)='syzkaller\x00', 0x1, 0x8d, &(0x7f0000000140)=""/141, 0x41100, 0x53, '\x00', 0x0, 0x19, 0xffffffffffffffff, 0x8, &(0x7f0000000200)={0x6, 0x1}, 0x8, 0x10, &(0x7f0000000240)={0x2, 0x0, 0x8f, 0x2574}, 0x10, 0x11660, 0xffffffffffffffff, 0x0, &(0x7f0000000480)=[r2, r3, r4, r5], 0x0, 0x10, 0xb3d}, 0x90) (async) bpf$BPF_PROG_WITH_BTFID_LOAD(0x5, &(0x7f00000004c0)=@bpf_tracing={0x1a, 0x1d, &(0x7f0000000000)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0x4, 0x0, 0x0, 0x0, 0x2}, {{0x18, 0x1, 0x1, 0x0, r0}}, {}, [@initr0={0x18, 0x0, 0x0, 0x0, 0x6, 0x0, 0x0, 0x0, 0x4}, @ringbuf_output={{0x18, 0x1, 0x1, 0x0, r1}, {0x7, 0x0, 0xb, 0x8, 0x0, 0x0, 0x400}, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x4, 0x0, 0x0, 0x1}}, @exit, @cb_func={0x18, 0xa, 0x4, 0x0, 0xfffffffffffffffe}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x2}, {0x85, 0x0, 0x0, 0x85}}}, &(0x7f0000000100)='syzkaller\x00', 0x1, 0x8d, &(0x7f0000000140)=""/141, 0x41100, 0x53, '\x00', 0x0, 0x19, 0xffffffffffffffff, 0x8, &(0x7f0000000200)={0x6, 0x1}, 0x8, 0x10, &(0x7f0000000240)={0x2, 0x0, 0x8f, 0x2574}, 0x10, 0x11660, 0xffffffffffffffff, 0x0, &(0x7f0000000480)=[r2, r3, r4, r5], 0x0, 0x10, 0xb3d}, 0x90) getpid() getpid() (async) getpid() getuid() getuid() [ 1725.279150][ T5343] workqueue: Failed to create a rescuer kthread for wq "nfc3_nci_cmd_wq": -EINTR [ 1725.596113][ T5431] virtio-fs: tag <(null)> not found 00:28:45 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x12000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:45 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x7d01) getpid() getpid() getuid() getuid() 00:28:45 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x100000000000000) 00:28:45 executing program 0: fsopen(&(0x7f0000000000)='ceph\x00', 0x0) r0 = fsopen(&(0x7f0000000040)='virtiofs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:45 executing program 2: socketpair(0x27, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) socket$unix(0x1, 0x1, 0x0) (async, rerun: 32) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (rerun: 32) [ 1725.677439][ T5445] virtio-fs: tag <(null)> not found 00:28:45 executing program 3: fsopen(&(0x7f0000000000)='ceph\x00', 0x0) r0 = fsopen(&(0x7f0000000040)='virtiofs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) fsopen(&(0x7f0000000000)='ceph\x00', 0x0) (async) fsopen(&(0x7f0000000040)='virtiofs\x00', 0x0) (async) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) (async) 00:28:45 executing program 0: r0 = syz_init_net_socket$bt_sco(0x1f, 0x5, 0x2) connect$bt_sco(r0, &(0x7f0000000000)={0x1f, @fixed={'\xaa\xaa\xaa\xaa\xaa', 0x11}}, 0x8) (async) connect$bt_sco(r0, &(0x7f0000000040)={0x1f, @none}, 0x8) bpf$MAP_CREATE(0x0, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) 00:28:45 executing program 4: fsopen(&(0x7f0000000000)='ceph\x00', 0x0) r0 = fsopen(&(0x7f0000000040)='virtiofs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) [ 1725.744769][ T5457] virtio-fs: tag <(null)> not found 00:28:45 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x13000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) [ 1725.791728][ T5470] virtio-fs: tag <(null)> not found 00:28:45 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x108000000000000) 00:28:45 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x7f02) getpid() getpid() getuid() getuid() 00:28:45 executing program 1: fsopen(&(0x7f0000000000)='ceph\x00', 0x0) r0 = fsopen(&(0x7f0000000040)='virtiofs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) [ 1725.839349][ T5478] virtio-fs: tag <(null)> not found 00:28:45 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x16000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:45 executing program 4: fsopen(&(0x7f0000000000)='ceph\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:45 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r1 = socket$unix(0x1, 0x5, 0x0) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = getpid() r4 = getpid() sendto$unix(r1, &(0x7f0000000000)="ea560b454b9b9006ae3c5a5319201cca294cbe0d6d0d6d4d504442d5f170985bbe41bd691ae5b37bdef99ae2553bc8ad9c93b51d26241ee676a4ce69dc4ae66f88375c2b7fc19ee745837058ad9244f20ae8012ac623e7120e8e4542b6753a22", 0x60, 0x20000090, &(0x7f0000000080)=@abs={0x1, 0x0, 0x4e22}, 0x6e) r5 = getuid() bind(r1, &(0x7f0000000100)=@vsock={0x28, 0x0, 0x2711, @my=0x0}, 0x80) bind(r1, &(0x7f0000000180)=@pppol2tpv3={0x18, 0x1, {0x0, r1, {0x2, 0x4e22, @multicast1}, 0x3, 0x1, 0x4, 0x2}}, 0x80) getresuid(&(0x7f0000000440)=0x0, &(0x7f0000000480), &(0x7f00000004c0)) getresuid(&(0x7f0000000500)=0x0, &(0x7f0000000540), &(0x7f0000000580)) ioctl$TIOCGSID(r1, 0x5429, &(0x7f0000002c40)=0x0) getresuid(&(0x7f0000002c80), &(0x7f0000002cc0), &(0x7f0000002d00)=0x0) r10 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000002d40)='./binderfs2/binder-control\x00', 0x802, 0x0) r11 = socket$nl_generic(0x10, 0x3, 0x10) ioctl$ifreq_SIOCGIFINDEX_vcan(r11, 0x8921, &(0x7f00000001c0)={'vxcan0\x00'}) r12 = ioctl$UDMABUF_CREATE(0xffffffffffffffff, 0x40187542, &(0x7f0000002d80)={0xffffffffffffffff, 0x1, 0x4000}) socketpair$unix(0x1, 0x1, 0x0, &(0x7f0000002dc0)={0xffffffffffffffff, 0xffffffffffffffff}) r14 = memfd_create(&(0x7f00000044c0)='/dev/virtual_nci\x00', 0x0) r15 = getpid() socketpair$unix(0x1, 0x2, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) accept$unix(r16, 0x0, 0x0) r17 = openat$loop_ctrl(0xffffffffffffff9c, &(0x7f00000001c0), 0x0, 0x0) ioctl$LOOP_CTL_REMOVE(r17, 0x4c81, 0x0) socketpair$unix(0x1, 0x5, 0x0, &(0x7f0000000000)={0xffffffffffffffff}) sendto$unix(r18, &(0x7f00000000c0)=';', 0x1, 0x0, &(0x7f00000001c0)=@file={0x0, './file0\x00'}, 0x6e) r19 = fsopen(&(0x7f0000007e40)='cgroup\x00', 0x1) sendmmsg$unix(r1, &(0x7f000000a040)=[{{&(0x7f0000000200)=@file={0xc563a0953a94baf3, './file0\x00'}, 0x6e, &(0x7f0000000400)=[{&(0x7f00000008c0)="26df6347310b2ce65912c6171ec0949f16097b0ea5a1bb75508e545f2350baff8a9e88c623f80fb50d9d23a97eeeca24ba8c256f3414f3de03b745de410f5e856172998b1318bfc12b0eed3806fe4abf662dd7575092e018629c901d2209927ebf953de877a7191044421708704820a0d28a40d00a94d1a9861992e84748d09140ddc41c1ed5305e5302421ba6461020a85d847d219f17e4fb1fb6e51986b475f53278227dce83eb443b1e48e80507ddd9919296d6c26843b417488df70ec501432155c2d1fdf91c0fdbcc69d8235b1446c4fa27995e6e68e96f3fe27efb4aaad3c7359dcbd04a78c210b10c9099c551ee715dfb5afc842eef0bbb509ce39b26120da98996de3cd3d17d19c1221a82684b2dba475eb1a89d1ff7122e6fb52f13e09fb5b7c03f10ee16d56ebad4571041fbb22c6a9afe0a1e8e35438b9bf3ab2c1b28de96b34492de0acd00ec696c969d9207efcb3896c5a8563d1ba2246c217d9f0c492634ecd78cc5bde70e82c3e26f6d208094f3bdf3b5a14af05362c46141bc9cb38d8f1701de16d36ad62a1d39412a9c9c06da88f01512cfe13cb5899c4ce955f918323ec53c03e2965ab48263ff5b65bbaf7ea84d44892c80a1b940b39922cb9c7a9800888c2ac625d6d23ac43b3c655e46bfcdb2fd033cd2ed803d9f4a0d264b5cc6fa02ad652232b5744566f5d5fb7fc7b9f0570f1457a781bbe77e3b70c0e2649e43df265cfbd5a6c38927f627481a343cbf4ff168adae746636658932ce2f0388d996ac2a9bbe0bf6bc7581fa624237f8ff556083e7200c26ab8cf24fa6c1e363618614f716d9a2e1af85982cc414e4ce48d5a7bac6dcce4441febc7cf54f20d6657d1fba1d543a63b90c38871b1beec8ef62281a191c0b5e3cc36c0df6e7a38d03730053c4f46f422de91ccde20ce8a282f605965bb1ace689a29ab4c327af1f99491b24ff08e73bff49027ac1ad4aab135727d893ffb86346002295994d335798b99187a034f1ead9e18524639ca60f69a90ea6bdae6d1fa3cc051900817d1cef2c06ff45577933915d47a892a36cfbdedffdcd1b084b3c29ba5b284d1efc953cdb9c4a329f9087c2c557cd0256b8260786933ea588cdb5898a6a187d0bec3135db5fcc35178697c14bab4d2cf6c861f6427f1c697d3d074167494830ff0e074277177135f4b1a5ebfd7a63ada11cdfe7907620b26676867dd597a6d54eabc30a1ba724d488ffe8c7a949ea373dadd27e039c17cd4cc3ef92ccecd73dd3db2dd9de79fe89d4f1626e7dbe1f3a6081a197220235974f8f63ee590445ac78bf2c346001d397481b197c2634a2a38e5f46e441a0583fa12d545c7de80fc3e8e137839f1ce5c0a2817b46b66e43152e41f04a4e5adc78f497b54c1b9e21d8f02bbd1148c0e21194d88a59e4f6a2b5065c050c053ed6e0810c03cfdd469ca1abd6abc11ae2df6475bcce9f2916c0e4fd9da5e28165c8cbd459d03a5be723697deeb9af93b15a637c5b0779c4c7cf76c3cab9e333df9b9f2297c4370c59d29889bcc363a4a7a24e1ff1836d373c45c7f7b9fe5f528b6e18b7af9924276d3b4612405fbf0eed72f5592affad74fce0c1098a6b12fb8455d348ee364e50dc18193ac587682b2d37d5d8427f6513c74e1974cbda833962611f95fb9cef61cd65e11ec3e328ac19180f3a313ee88685c95990470702db1bc7b0668d5c1839a22af1c365c5c303fcbdad7904e6609505042ab2b0d199abb93601cb00be3eb3c4652bb8bb44634ba11ed0e736c49f68e9802ec093aeb417df1c93bd7a622cf499b9c96525f54f27293edfc8bc0c4cd258084aec56ea02e8f276de1dd903cfc2beaf0e1754365e740805fdbcbe33b2bf804bbf0f8c5862be396e08d09bfed47fddc751f3dfbae6ea3224f3185629217eb5ec7505859aa26072c3aa9a5b1b3dc95f8964c48cf6dafc9cbe0ea83e715e7bf56f92ff3847e6524c92fac30eb48a178c79ebad3bc04772c8b96c5ae0b082f99d090ee29cbcc9e84897b413df84e6af5cf74e277c0e519baa9755d03be6996ac57a07fefeab749fcccc7e85c0ae137f2c5b1705535223a900f04fb04f54066123d9679f96e4c85f1bc909706997b34c6c42d705e089af12a9430bf32e6f3cfdd3fec6ad5a53666580386e604a241ea4ad95f9ededc70fb5ea7664aed7f78861963c4829ab0a36f08068df0c1c198a35e21bf02b23d9891ecdb3185f9820b93d05d3de4f089f85b95080fa94720a4e50e1593bebcd9ea123a04b02e15bf4e3ca2193ceadf9ecbf25c013a6ff1c2607681b56974cb046b58540d6396669fa4454894486016a6f783fc46793687979cd469e5eed81b8e8da97acb00f55cb887851e7023db6671ff3260d67dc041e17bddccb3692cba50061718b97b7a6964dbb23f523ef0826dc60dd09c34f05695f0e07a50c8c475ebdb96111c226a9856d231060ed65a270664da7fed3393ee69da5e27a3c142c26ff10f696f0a24cf0cfaa7b9b86b7a3e063c31273c8d8068f1ed10963b41c899082ff807e9072a09e33aba36d072540f4efa32b7c9bddf20aef302b8b8800aeee6f0e33ad2de2e4f37cae8f390ccda0fcbd854354cf7e191dd7d37da3318c99936cea27d820683058117337d10f19ce9a81b5803f33087d4387e87c84ebf8f4a24814e963a0db630c68f95f716543ee36b22fa8fc1ef3126938bc0203975bdb38c9e38045c124519b2c394e915e6e7ada1ba71406b580aa549f08561938f61c704d09264a0080322d409672c8413b17664861d334353112c1d4161bac59e87bff00d8fd9c186c8e76c8db0377380bc58f441006f40c85345e0f85af8a4b9ba062b7d816616b2d19ef699643eb891eb1e9f0e7588e4b0481d0d77e0773e9ae33b132fe6eaac99914ed298897846ab21b350422cb4cfd16334fb2b3d8ff850d79921899a31bcc4b2c4710515405f2279692b6ce29fdb79e4c7ba7c91ad21112ae2aaccf2f865fc2ad8959d5a77c69dee6ecfad26a604680b35577b621cdcbfd0796a937a0f7a4e2e075cbe87fb8cc14a4ffd938945243acf77c969edb7a619166c8c0d0fd5b147d8c322376c2d6e2e7637abd8233013f56965ce97d1ba7cd4a3c187b3f475f0896db9f376a2ae9e7e6ef093072ca39f1009d5934a9b1c9608eeb5e1bcea1aa6d9c9a0d05289f5f335b52ea773f9d652f81e607440e7a6264b9c4ae69dbbdb6036251bdd7606db2c70202bf55dde67e3e6f4c30ac4562797e55c4b14a6173ef45f19ace8cfa742b62f2d40f120f0831f9ce0274fb312ea311b210d8ef66338e72a69ba172c313c665277ee0e0bf9877540b32c9ecf3cb91191f4e95a2ac651a9f048f1b262264a189b4f211edbe43f75f6552cb5a63395b99f36e69eb7b50fbd3e71fe6011b2328f20bf2f63ad5b43bd496eccfa6f67407b97a02130d0de6c234279f742b26d7e848018e234d0425ccd92e9b6406a119f3627b7bd064e45a546a6fafa8cf214e2bc09ce4becefc7e7b6886da6387b4ee15e2dc89381d19d66a15540c0552931a170c4fc59e8dc4cabf9e8006d3c9082c37ba0e931ccef7855f89801c124192a0d8d1de93e0557ee17237aeb6f0c52d7aa6a8e4724fc97b28e7ff719c8cc61e0ea4b596bb0d121d96a82157b26ee4a7a7cd13ddf1d5e810ab219e8745ca50ffa866afd74155e9f55b4014a704ab27214d5f61068935b8f72103a56efa9b602fdf55afca083e7dfb3ab01e21d4669a2d954e4fbc66250f5b680d9074527c53f1e7a30eb428738cb7c84d110636d299da1aa8366c9e0a90def0808600ddfd20fafb662f4f0db9e2360a9e6080a3473d46f8e9fbf62c3bf37c8b848f87b7f53aae1d82d1486776a2cf44a1d41dc5265ef1ee1928c7338ef56ff3235d48fba64c076dba2986db6371b16b19160cd1a3a32e0e7ba97317856e8a1fde3d2ab32a498cde0e1f1e88519e01f9c4b51fdd068f5c396e220df70f1f9ddeea09134dee5aef17618c98eda3c293b519e89ccad03578e76f32a1f4d7dbb3c80345bc64b8978406da2866c868158575daa2858f89d9ca620d207385ab31a6ac5fa26564db0c94d4146164c6df139f756072f19fa7e1c3a95daa747c52a0f0eb80a2b9f060610e58f84023cc7834188d62db93e84ead8195f9561f8d17dabf463bf4ada7a895ada07fd0b44c7fc92856458cb9cacaf1eb8b15857de5b0654851aaa2f09cc969a2e7819ded0819f93774eaeae2521258730ab77540b253f1f2ca1788f404c641821ae1251abe6e63a9aac8ba4a18a29dc00e26352c9ae8705874a041b681fcc97b16567d3796648a9457f90dd98fb9cb627f40899aeeed36f955633464c8ad28de88fa0a07ccc50b9e194771e476d5a34905fa3e16d00abd99feca6ef5a21d644f90f4bef73f819ca57366834c8f2ec0c3bcc0e9bc8297609ae5fe5bbe97d11c7aa233ecb454db2a1319e30e58562a09863a7df8d4b48f5b3c8674d10e7d65c5f0675ef366818e7036990cd85527d50708dc6cd632f9b488e0392cea205e384e56e844d09774760d36005ae2f959cecae20c028659ea0576eb2bfbb486552bf9d0413be08ebc63aba0f96fcb765636f7fff0363e031217756244055e3d5ea5044016186b9eb43e0bb93fdc1ebec69da15b2b504ec3244e4c542a31c89748389c99273e7643e57c1f7bf6d421e4fdbe42c164114da9728ae11518e780c4cec968e9e55189f11090bc7a9b2a81f98b269e16f8b1c13e95c234ffc276906d8ac3dc10032f81197942bc29b215d73f98454b72cb3967d57dae0bbb84b1903676d8ba9ed1dcdf720d21985a5e9a1e38c5272d55e8c541e4c3747808fd23f75c47b11a21e8c715e677762310566f119bdfb3dead609900e73d5c423839dfc93c0dbe5a0de304d29f9b0e3afebbdbb179aa3702b06409b85643a51cb995a2bf47d566629ef76d674c39fda4eb4b9dec532d36eea32eebb05ce05d5497219b29bbd7b8454ff7315a0a7c473002c51f99478c0c27993a6d38f8b68efe51f5338ac6675c4b42dbfb049a42ee76a29ab7ea93613a2428aeadbc5cf34bc5702598c5bdad67602a4302ff55c36cc2560003c9d2d6273c72a15b913c4501838792231f85c0cca8b8f81ecb80acf96bdfc77cc1a80f2783eb37395b1b02e99a30c9f3cb583c61c162c94871fe3a084263b5918a1ff0423b39df5b76ce5f431aeebd92ffe6e5ed06b64ff3cd0ea9cf12914f6f076e31eafa1f1b26dde6f0340f21dac464b3b943036fa11f913f69da02b69fbd2b586bce8217ceccc05536fca81ffc179a494ac6b1a5e450bd0d316db8359da4fa1672deec9d1bf0411bdb943a2b601cac6ab8f3275501225131eb5193eea3f9f05090dfc021bccd6453be0428692eaf40f780da6210de0a05ac6ec3c5feabd72ce394313e57456155eea0b9daa19d4e3e3fbe5b0b65648f3b78c09fecbee0f00f1bf4a56751618106126d39cf065b5f412878da68538a0cf9457a50bbc363b1e2e5236128d30846e2f0cef94d961190130ae0acc1ab29ae7c419821659005ebd8c9167f23267316f2e4881c1d5b4a355659e9582c208b684c865e1cbeb5193e8c8a0d8404fe67113163d0decfd408e0dc91310efafe4f9c3f5628ac5ec236b104fcf904b5cea909f40fed759586672ff4a131256d53f3ae3ece76ce5606ba1b505f70781c9a6b26348181a976b658940041c78cffd56431d6b992fea4909a71b88b1d97a188ef010a8f8e2e8acdcd6d36fe0feaeda4268e6cff61f802e345e029ea2e838bc71996", 0x1000}, {&(0x7f0000000280)="8b7c5aa06b1f0149f7d035c4acae8eec44a4253f8aa16c76adfe395114ec1cd615360d5909ace6d43af55e83bc8b825a9d3b9426f4a6cd34a074a2c0c205afd479a1b0f6ee7749416ea304ce0de125df459b", 0x52}, {&(0x7f0000000300)="23754504d47058e50a708c5d928a28caa8863177a96628736cb5514a721e5e14082c8125cacd7b1ca49288e2a351d938c44d57b53cfc1c3392de5c3cbe9650ca88ecb9b5fa3237aab74140c2dc28b60209415eba667e171b072f172fd84a18f24fe5e36c5f4caf697c9e1190bd1499b902364c762e6c8d3835de51933bc6266002322bc1ac1bfb4117c6eea2207e55ff2572ce210743a78141736990e5298bf8c5a1cc9201bc70e6e3a3fb0a18fffbdf2974b1e0297017d7a391b50bc11aff9b8191e03694eb2506c7892227700542ebf45a5f9e02005ca245d86d6903da06495a3dc6955bb12ce70f4a8782076a4467eb27a897b052adc56f747399", 0xfc}], 0x3, &(0x7f00000005c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r5}}}, @cred={{0x1c, 0x1, 0x2, {r4, r6, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r7, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}], 0x100, 0x4000}}, {{&(0x7f00000006c0)=@abs={0x1, 0x0, 0x4e23}, 0x6e, &(0x7f0000000800)=[{&(0x7f00000018c0)="84e7419de7807afe986786d254a3807962232a70ee8f75d457990481b7564587feecf8d7d55e2fa705364d05d7c6adf0b1449042a26bd9c0ae420277e5597cfe15c63ccdb11efb11d6ae76cf854029798d1237d06221f55d501aab6e566157835e22f8a6cc877614db7eefa8e2095cfbfc6f3030145738958da325bef7aab490c1d5f022a233d991b2dfaf699b42a9ea2ebcef84abbb2cbfb69ae8f2cd58b7daa131cfdd1cbe3bedc4efe548ecb05db6d67619d30b98045f07aa23e9fb137ff01e5176fa50b8dfcbc1f5ae3c525aba864a8a876214db4a05b3a44be0c4a3841c24483b6dffaa0a36e54e1035b23040cc7dcd0dccf4d20e87d6fe706480ca5da9df715cfe8e37cedd19b09ba40406052a402aa0ca9542ffa99df2b61fc1e7b309e3d80a94f64efd69f35ad37a0fc665745f1d9d59e404b28a7ee300060de4f6678dfc1c0aea60e716838703107e3c24a9642fa6ace21ad8c4980793377040aaf3c81267c1bc86eefba67b226a08afb5592e89ccdcc785dc2e4a526d9ee7d610fea7ff5119908b87b41f9005779665218fdedd34cc7079e746a61cb7add96b9098f31255ff8711208ed51b733a5b1d5a29f422368ce6da55c38ec755a09772f2978db8c04228fba8d88b9a4118aca83707c0d792fa0155c52b015b0e23342d1b9c2456b1dcadf5a423187008aa43ab7d83bb9d41f3ee747f5ca9efa1dddbc4302ba8e9d82df09848cd78d9d2c2b7a2851177c2d458dd038f5dad0b6c63023afac0428f9638bfb1257febfae9a1a03ab122480c671748897131bdd0d3ef834ecb040078c967c67a14ee5f61a4f64f2922d120f773ae3fd8dd32acde92db5eda6ea891a6cd8d2c5c19702c1fdae8daf1ba0c16f3835651325a4594aafcdac04e36b4452761a13af62de2659757456594c7c2433741f3846a4c35f62207ebd1126436551baae5ea6c8bbdf51cec3ced03e51eaa2be3b7207c6e71a8808aca9cb3a1b84cedc903135dee31427e6f38a9f36dee06bde9d49a6091357bb15839c36039f8ddb0fa342edae7770ab0c5aa66de174a5813db5d97d9b55bbe0063d603fb851fb48b7b5db12961444b8f378d8e387a06a1c1bde3c6f5f850779cc2ced53793a462edbc2511bdce50535b47bfbebdc0444bcaf1aa130ab7d92437dbad442456d56a837966c253bf5780cd068620c31b49d10e45bd39d402eaedcfc524c0add302437df601e60db5f7e3488296e26305aed3317063d02c05b763e33e9bcf8dfb9ee520887aa44e9c4ea068240ca025c296d1e38146a6ed81abc6659d64295cfc55be5f422b546b81d0bdfa328627f6bcc866fbc88c00c33024c589535c7d97f6ae3111ec9f3e0f7814eccaeb927a520cd04119c99189f08c3983af1c560fbaf215ee57d67a602115b1c58128ee9ee4f017d6117d6bae15218d94c25d9e15f6760b5cae13626eb8688a76313bbef45d247a41eb17623707edf00ba12fce0080d007bf6204e5df6ecd83eab8b56e760d3e537bf5e8bd7bd6e8d0ab617f7cecda59e49bfca69e9fb82dcf492de21e59c2a81272ab7667db42af684621abe24a50eec332a93c365682a17ac3bf917dc746be4ac418d5d404756b62024fab85d0dc24483e69692f14f5abf1b24c85ded680af43aa4c38d0031680591f7458a75a6b2ac30f9a4a47c0badbe00c6a5ced8d4c2d1c4c320e8104319e2dae4fd59161959f4d6e3489f5e91a335d44a0712e4e37a40da1b4673ed199550dbee9a4ace0b00652f8a3bb0488b6d2f2d0dbf581ebd77c354d304b5be4ad560cd3cd6cfa01021b453381f668d73ad317a980fe4ba2fb1520363f6534dfdd29100a01fcd871961571670aef2ac026392f6e7bdad9c26e6fd7c5a8379f4671a3493443f179cfaf061c04752a1798624ef381700ce55a0424b2f8765a3e785513ecd0ad04e51adfe15c9dee0928913d87fbf87daf391bc5579085f08d022b4ed21917858d67b45aea2038803ee691d3c5bfbe4a0e6a4290216f1230137c3acbcfd0298dfaca032c551ac624b3fb24c26d002725fa796b75b4fcd15458841bc2a5c8b0d28f1c9a614e71994234f48fba29817531e5b7f56644727ede975921efd19741823b40bc49d68a2e2a2e0c8c6de7f676e3f545d2a347d73579e961f1096198576e3f1edde72c3b948f5c9fc7657d1406065394f0bdca8241fb0ef72b085e98a0fddd7ce299a9463223dfa34357f116dbf693762a7e66578a319b781b4c5fbcdf8ad1e994b4835d2f9c2298c05c2269cb5d31905dc0c7dc2d82e0d92db9f0e3b5de0f21dcd11080600bd8c8f6e0df8a1e0b6bf8423e592348b55041307e2d939d284bafbaebcd16f38ad56d71f8d2bdd2ccf40475f3ad6615213eb9bd7aa471213dc7fb449c75bc6cb9cfb334689a18fd0bac548cab3ec8c63d8c8e3dd42bb421d20845ca5a39e7a60d8bae06bd3d6888424f566358ab128a6e17396afa1c2bd121b6b6941aaf83dbe80eac4fcde184e4c66112c16c8d80d666fb32230cc7066ace8a4fb0fc67567280c0a91cccea6e03b45ccd6b0b04796ea31160d165932b63dc97b4298f0a39b6caf0f7115f65a238112c2bb387a64bfb9a61b03c9965b0ff80a2742730d05ffa4535232eed3e3098d40cbdc110ead18d7bf707451768f636375733927905fa5e24eab71300d09a63b42ab21cf2f146c0c540b7a520f0ea8b1ae4dd8f57b0c12d37df106769ae9c0e77010d1d2ac392bf2b291cde15420600c445c4b47b0fb2dc4038d71c0b6e25e22d232100694f6d33c6f8470848546bc601f0c6336e08dcb6cf661eccc5cf2432d00c230a21eabed73310b6fdd9d654ee16a904b04e79e493c855a80c2fbadb75e08cf946b58ac1b3ef6548771e9d0a11df91653e168753a58a611ff5910cee2c6447f5c565f838abca04dc6cc1959a4192d0f62eba7d9dbca297a01afae6095c49948d1b6e0289184e2311e6e3f6a7257452e9758911d47529a6ee1e238fdb93f4c574176e0805d8e3a7272d9086969fd7f3eaa5ed4c7f61e91abe6dc22b6ffc89cb48f15dfb47f56c1b15c07199709741300d4bc638246a9f69840f9736cce01f910e5bf6ae3af3c9affa5dc7268d26e06196c2a187f22fe2c3ea78017e0d643254a3847b583d34bc1965900cee2403f14e00f310eab0790b8fef28e8d2727631fa32058c9150910711969df9586d4c7d7fc60fd738de49ada181afcf96a921e31b1ef85eb48f05d2821323b0a1db183ddbd196bb72f77b82a29ab012379b15be331d107ddf4d9132199ff320452ec134a9114ffe4c9e4636790b99aa291ddc91e8344fdacd070cd3db029b3725b82b13ff28a7caf1dd7227fa72dfc867db6d86861355020ae83b705cc9fbe62af9aa1e4476707ec57a86998dc3f418ff894bfba45ff6a967ba22724aeb2401ed8f4c943ef5d8fc96b211e6632c752514d282c46ce494f11134becc50a51d7548354c8293b83f2b55a88029eddee89cf8be62339d23a7ac72156e26ea0902e1f7b445bc65e61c8a76b6611f742c6661a652d82da36bf2e457aa76b9c3340f19d25e8d5a0e8989b54ebf87711879004b9566e54e8b9df0d269f70a72ca8096f1d6966d66a7cf9a081090cf3ea99e1f91d683083c5edb3ae529a5bd73caabe2057dfd9d0584668b195a961df750e564c34477ffcd58b4333abc24d257ed76e470ff183386959982c9be61d0617ce83255129b98b70c46de328ca5ccd76f72917cd279bafeecc529dc55b0729d90515a08d5575ef21409db69b771448a75fa8489b2401c1bf2d3197346235dba1264e60b0a4ad8e0225963da86af0035f74a106ccd3947d4ade26ea5c6865e65ba6fc9ad51dc2a11f7aa2747312d8c27501b6e6f339796561e4210f32bd3ff88e557cb2d51f43d1104564318d16c227c35349effd23e93093bef312b09c14adf3ed3cc45e5b27d1e696a8038ef4188376c754027d9bc4746e44d3e78c598d9cd5c428215ff3ac8c10a2c4bf245454de8a43aab0eb5d975573c51ec2496a2b14d62f293b578f42f983789f31d6aeb9e2e8eeddfbd3091f57f94678b6b1906d173d99c467a1920fb4ea58ef73758c33352759f99d3302ad87388a301512908286f712084c818e737ed1cfd42092ae7e28b3c8371a036968b080464072808fea7bd950fa4aea78fd1055042ade67b52f739bcad9b0cc3355dd709c92455b7f8d6faa8d61278473bc0cde703e4788e01f7ba2eb555f3a7fbfa6e876829ecc1b7c619ea45336ef97e933ab514ec77ae4d4649740959b9bd9d897b06dd53dacbe35286a9aa5ab2d2550200edcfe1435fc78d9a855c48147a313511d7a895d5e16f6b7615551239e53bf0e5698bfab18e33bfc3d65e247ac944f62ea827b7c1a8a664c2150a9c51608e0719c22e3643cf0c5c7582710c6613054f3e22c13e42dcf15db6a78e39752415df811c843af56223789bf4708e1c108787a21e3f9608c72702ddf194306f573966374fc3882af7e77cf21a542502c5b6edbc54a8169da047de3371a146b6183563e4bc379ec44cda559534797fa4c6db16679ed20b6b474bb47322ee03ada85bed6a1de7fbb0fa84d0456d3953e4f10239ac9311e03a29009e880b89bd2a4a05ed0f7862d1e05e06423c136ecfe6af159a850ece88329b0c0758ae5b156303957dabd56ab7d989cb184ad0b91627ec1929b0e25b75a7c9bd2c480a4e0999f1a5422cbbfef0c6036d96bb9284c445d07b920e277a7f25858dbca51ba28c766e0755fb85ec40882ecb2399f7b30a6e90f4974208f45a400419eb9a484eb7f1189124b8a8a2f92e846528ca2be8b512326b5cc3ce5bfdff8e3cf1d5de76858d57d94992cf6a32dc4e1510a6189ca2e649cc90a3893be81ecda834dee5ae099df7f44e244548c580893e756c20505e77470db4f84b48a3b8ca1e4e2e317af499c5c12abbf69609632f7b9e2912bcc133d881ce760d1f3925b922f34266f1016ccc29ef56aad61b0a210b8cdb0b7d2af4c0d60ef273808bd9de06c4d84bb81c7675811beac7bbb503859245c4054967d899606d8674f9bf6fd15fc6c120e3315e62ace711b3c6d403a9c1785bfd5f42b095eb13ca94e5ebb27eca868101c5f2bdc464b6472314f22d73cbd731af745b4a13f7835cd8d8c60b6c8051cd445166ddc392f48412ea37b90792b470f7916331234ea2fb613d78db9defdb3e1f3e219b4e2250c64fc90c8b334e00c83403f04ed24d08f1b9862cdb82b2d725b1b4ee027451e553d1c07eb72b291db650f2141da3463270e6848aa172e6f41dfb59c5838c278ed17c853d06db43ba439ec09705204c6349fa8c9fd18a31598fbd432b186d3fb91e39954a300c35150e968bf4c722b9e544ffc073bb52ff3efd7c5234422a24bbcb2a1245d8b4349bf851cb5ce2528be8b9bdd227fd4e390567b209f70f3386ca8a6e70aabaf3a1a533fa74c823609c597a239301fd03484983e5bb898db683fb8c2d66c549b8c0069aefd0aba1d35ee9c415ab468e3a7b23858cc4d85fdf91a258c68fb150d785ca22875a7952f6ed43d7e08791017734dabebf002b54d6667a5ef951047b4834aa6ec2e384f217e0317840ceef9219eff31b6ac1795e5528488ddef90f5aa6a64cced8bb237ef70b94ee472c0083d117c4660507e32dafc7ffda13e0b20e079bb02005b16cf8eee8ea5d2c7a85cc42cd1899b9244eeede6adf33b32d69ede79bd8712c602975e5fde2425922b417d84014fbbca029ebbb70de2b8cc441fcf9c76c5f4edbc97c4fde1c0613cdca0ff603", 0x1000}, {&(0x7f0000000740)="466421a47d47df50c6365679ed631823743dd09002fdd7a9a79563dde7e30b25fce0944d50f303513e9d22098f2355010fb04a64c4756e0cb840a44f7c9851520c472122ee49b180cbafcb207f3f29cff39c71c04c18175286fe82af9d46eb06b9b9dcffddfef69eb48a3bc3ef8fd81e01b1c1", 0x73}, {&(0x7f00000007c0)="0f43e5038345cfb5ab33b5a14856cd27558c33fc625df021a38fc60b8d56d9852065a5047bdaf87390203f98e711ec90", 0x30}], 0x3, &(0x7f00000028c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee00}}}], 0x20, 0x804}}, {{&(0x7f0000002900)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000002c00)=[{&(0x7f0000002980)="df08eba0bfc089261b639ec7cb50b403d8d9e041f6f764edd1bdcd32eee1c4af1666b3e3eee568b9145e55a3deb0998c51d3a653b146028c0b4b856e08c6426cef98c188d27d215b4f8caa104920366df4d7a298f2f57a971ac3034e3394f2c4c4f9ef22488da0a83551da6f1f90a0805190f5d9ea75ea1dafef99ad0ba80683a7c8fe0526e7f133c703d0f88ad78688fd41f22455d1405fcedc19795858228a9fb18e830c737594a78e68443b775d75092ebc434b69a57cdf", 0xb9}, {&(0x7f0000002a40)="3f923d00d3c136dd8176ad686725ea1ff9314cb016f5dc00a41bfaeb9a918fd454c3702e6a815d3df04447bd24bc106df6b0bcb21d060484a8509bb4e219639b4db7ef0815cb1fa80e5eff0e1608c84984743a92c839af92ed8d9945499a5aa95bb998e1045babf7065dfc6cd1fe6d9431c9f5a561256d4f940eea19250c4d3e8e3bdc13111d16736c4a0026dbf92aef4d3b796fd5dd7f5c6044aa8185c572a7f38c7a58220cb0", 0xa7}, {&(0x7f0000002b00)="b21ac1057f742d5e21b705c1f3b78fa7db7c71b1a1b98c25a46d4b6187f83cb7e93182d09316f4ebf8cdbafe41eb693e0b65709893d2f562812e1e611ed90860f73c40632a6a6c5caa6028c2bb1ea762748444f2f22dbaa98416b5ae7d39dc9dbcb2c3007a63f0f11208b09ee2b61cab45a104d9e537371084d2af99a579469021b3a476906dd34b6124b9a5444ccdc962486ee5fdfb9d05f4a2b6a04bd43a538451562b6f31c3a244230578290fc9b1edc619a45fa1cea5d9df974313a7067daed8de30ac463a019218c60d71317fc25664a6d87abf5792e18061b1225028646d1166e0278cedc8621519ca3759", 0xee}], 0x3, &(0x7f0000002e00)=[@cred={{0x1c, 0x1, 0x2, {r8, r9, 0xee01}}}, @rights={{0x14, 0x1, 0x1, [r1]}}, @rights={{0x1c, 0x1, 0x1, [r2, r2, r10]}}, @rights={{0x34, 0x1, 0x1, [r1, r0, r11, r12, r2, r1, r0, r13, r0]}}], 0x90, 0x4008081}}, {{&(0x7f0000002ec0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000004440)=[{&(0x7f0000002f40)="c151b349f05d56f9f4c372dd0fed519c7ecc625e78e802d29227a974cfc9a58c7df6d2299d8a347b0eaa458ce9482b65943feee3e3d055726444119273bf6aa56a6c91ef7b466e6718b1a1e8e752c199b72cc2928651bbc1427ed15baa56cf9d8a82c0600887cb0c2026e6303be70d8381fffc2633b5452a1278006b7b21e3ff97775439985d16f87e0c57498ede83120a9785c4ff9c9c982ddf0f34305632b6d0cf7a60b24f6a65dee0655c6ae3d952b68439d5491fe2a6cc9c480c365fe31a6a9d5be34de0c1a5eb525af7b0a4a7b3fabb876cb0ce771831e6e9899c117f8027c95e2f3337aa5ea3a1b2fa5aed6736b7d7aa3811e0f310489013", 0xfb}, {&(0x7f0000003040)="c092ad957bd69724d3d352a6d42ee4303c6ec7f01aa6cab1503840b654382e922052c23717d1b59a2704db44c442d19a5cd507fb5df80dafb93d4c89a48b31b5c75ae790190b8caa5ce50713bb86f2db4778e42939c378091f5bd0d8606b0cdaefacb3bf7776bc625567029119a4b221768d2b7fd68bc3d00a9f19", 0x7b}, {&(0x7f00000030c0)="cc9163f5439c8454d4a7b6211cef87f1d67d99b5888988258fc5d2d2aafa73dcc6849fc4a36898952d7c1182c051f749f4a03ae4e1489f2b3b8af996f7975d99e9d15577228b8012ed02259e3c16ea2da7ba4aceacfc6cb97b321365009efe14d29d8d4a8399f5caa243cb5b1f334ec443e7f7da553064eb2443816eeaa8423ee05d5ce6900ae0019df448e1ae6f54c18b9360f73096d4aa7057b21857679d0728087854fad17bd16757a601d265111799f581ab7bb68b416d26d82e98d4cc96b4fb0ea3a6c338b35e7e4b958c06ad9f1261f8131842a56528be17b675c4e741abced708ca97732920ce03141e42b56f595e776e79151def5fbe34e78fac06e86e58aef484fea97316f74b19495e6c493bf7f7f73532806725291a1b652e82fab1412204211978046cfe2cea844be25b2f871e56f165fb1ee7e93e540091ada2a0c43dc466a2a8a8b3079719bcddc79981f654d6761fb1b3c8332d37173d81664f8bbae24501360517912a917257e78df356af97e2095df5673edbf83a2f687f1b759dc7aad090163b8fa85f4f5004ad22b9b0126bc9e029217f9b157df67b8b216eaeca948b79228b32fd9d6ab25e4dacd0bb7210813d2ca0de70d7e8bceeae291875a875d9e5bf7918847bcf4acf26841b31c67a50fb4994528e9325a0b0621ad3f4a5a894183d8afc92695c49f3cc71e83d9ef9e43d3eda710643ca737be30279b139a590e6d38a86c30f2c53eed782ca2eb5f0e1f7b623cf3598a3edea9dfe0352de884f3d0d12820f222f21b474d6f8b9062e5adcf00d8b5afcc2e4b6ef351e34d11668dcb8803c5040e5193660c3175169443c2d6b78bea6b0f91cd1942be4eb9873fa74865b7e544a279e2e58e3e940426bce26c6cae964d48be534ae11921b52007a9e8b7178c19b4a32785c1f55597f104afccca13426a0f7adeaed7f7830d29eebdcce2c461ed82b682423466b3c5ac60145f5549e98fd3b69c631eb391b3f0b064bdf6f6b398f0b3680319520f3cbd8dac50a02fa876f241289af83e002c1d15b52983edc343c037a613bf67acb243c7c3f89943d14e9d0fa511cd4ea3c376f75b6a4958173d9a8dc8b14d883252cef31d154f962b55ce4f8e9a2d2916dd478299456346e65c19f518cfdfef6166b202eb566cb293884e616625f43ad01839ea284fdbe109f793f78540f4dac8cade633583b76b524dfedeb02a4b8710bd568edb04463c4d7f00f65650c8b23dc21caf3629ab0769aba78a999df9905ed625be3d293c5f6d4e8206e4163352fe90cb8fb998674b0cdc12df8d419205c1abf21b8f2afd3384ee37dbf476be4d34ea7f1509fa401597e3515adce73af004d825bfd065c877f2f0a5523f7b9879ad1fe4869fcb31e4b6474be73b69ba633e89d9b69f1009aab59d6c028c0ac6933f36d8bb1857318daf1b1327f0582aaa44560b0b69ecabc92da7e9b2f48fdec905c4db9ada1568305595b95978ab42d4d2c9b06d5f04b6c7ede84a7b31e6dbe0458e824747329a8c3028a614fb3a7d1f29bbfff396b3f625d9295ced9e3d447804cfa68dbe05f6ec5536aecba5b5689fee6323ecd06e2c23ac55032dbc3856fe626fcb65fa4d414985c7bdefa026410de0265700801741cbf98c9db2c1e70d8520c1cbd38d3df1f1ff397ccd65d76826a00ff7594b319c3c5f0b3db6358983072c65efc4f236b6d7ae847004f537350d06cc2a265d7fb2355a3a6d44305b15cef42b85082c3d145eb53bbf08ad2084ca3edd179e7f46e2ce0e7afa737d0d5f100b2f8e897359234c04b5084aa163f249a0692596780b1ba36b21d1a57a24bede6fe0c61f5abf369549016e86260333708cf6853058d11c2bc95649216c81fd520180475cbe0f345f4ee02973163a24ec0c52943760db105dd000504c28ba9229c17d83c3ec3007c7647638f1ff5b8be84b7bc6f1f8076e1e19ffc4678cf33daaf2ad563102100457796a9491c7d9617dcbb9f254bb5d942795afbcc6bf7005a26492393256ecd9d55447ee03d8c611dd9d6aabe423592efbe98320f20e6aa230f18e632cecf2acfe7a4de9fde0710093c472e4cad938730456af8ac665bde962df136b604008771bd101b506d239eb74a529e9b231688c71ad4732931ee083835afa6cd126d691ef9a114a77b50f1310bc0d4b1db4609ca83a3c90fcb15fa137341f8e2f80b810a396cc7cae8ca524c3a695a0968513372995ee53ae01ce8ce080618d14fb8ab5c273153d748b26f52bebeb46e67cb7a9e9fcaae2a57ddcdd6422d51dbc8ea268cf78e33850a45354d67fcc78edb087f9257f5e293d71ce2491bd1706a1175037c3a3d3493d02cfc81064b30ab545b8e7e23fb3d5d67893b9a6fc5df1499755e9d4503bb4b7f12c9f3f6e6485cff65d12a30ea1258c311d3e16ad43f869fa2a457432985ed6a7770a9094156b339b091649438d7b9626376f73038fe8df181a81f8dc049f2aeca6cbd9892623fb7781c470986eb476edefcf1c3660fd66e0235a7bbb1a14a87a870723935243a812c302e0fac7d78ff148063f76abc24d546b484b3caa318623f66cb7f36e35ebaf42ed7addf410bcf9926418ff3b9ddcd3a971ec6554e700de8688bbcda7fffc6bee48b44d6d88ba71da36f05b7a64ba4f9003beaf04f5841960267dcd487b85b7aa55d3e15a98a2856321b4c745ebd3ace72d3226ede44f40a54eb18832575f72237c0119ecf4a5bf0021d9a0bc1fb8c47a0826c25926fc10ad8fc60990956224a7a3b9d4b087ed62935660bce228620ffe248729479314e7cd0589927445e07168385f63f70fdbc9e5120d3dfa566d539f608fb3e81f336f97a53b3d05e4a43cac42179888db4eaaf249206bb2688ae8416718b886629d3d0fcd5c76e25510b01f2f440eec0c13132cb7f8a1f67f20aabfb18d3ed4b5b7f101dbcaa26b66ab85c77faefac77097441d593bdb6ece85e4265ebe7c930af2f6e5170942d8eddf45502b1f33a7e805e15219e3b62a7926c2fdcdd2d587e54db984fae2fd9688292adc483deb3001b6e3caa76a8d4b77fc4dd84dec906df91557aa36d85a77dc70f6d410b3208a9be60dcd95cf23055752ea37f5ef2657617fdcb43c0dc4e7928c706e6404d1b50e94c8ee11a2ab174d24048d772cad0d013c30b769d62d25dfd8f9f913a3d5527504e31c5eb290a9155d30a92bdc32bd3402ab91f576b048bf9c163b81904399be97d7293e91992e4f7bfa167092535d40cc7b3a31a784f1825ed7faf465b9ee1293aa7b266aa07f4c4f3af59361dc5db2b39a2dbd5c1e1d3a1a39c108f50aba6aebef6084ef5a9b3a7eb607c04ba389e25d4e0fe6e261ff6a3607a896f72d3119e8d2c5f14d7a304fddf29fdbd6394850628ce875ad3609c28be9d5eeb8aabd07d685cfc4718ecfd3dc84732784e82cc625ad4072aac8c337174f3b532be0b12af2b7bbc32613f844a01a68c07608178b6a7c968953672301b33122fc335cbf01e35a03a4878839a1d21c742d0309c7392804ff5beeeb16ad04ce5a51587735bf86af627c311c19097cc0273b4a91c991b5ba21cdd6f4f8089770bcba35c92762f312cd6b9622d00174062e5253f97a85cf3556ff48222e806423add05542db3aa36fc19aee99bd53824d0c821dad5cf30a50aff4b8eee65fbda77354f768dc398c98ac2afb46647a9db99a50cced19edda7394abb1e54f4f772b081f58d47ec88d5f87305294ca8a42876ec15f8d8fce56e89022682fcf510fed21d3eab2456b638f44e127a3b53a7d4bab531d8d607508e8618a66dcf57b59940412124c58c40a42f68c4e0987491b3d9de9cfbeeba411d7cec5060ef1c78a004587f42df01d7cbc867a4f7de049c4b0ff05d1cb31fcbb7844f6a22e0052dcbe64b208430b7e1cf2bc59e93a6312eadd0c46f1b6704b0720d64ac6a8e899f050d0da6bd52dafaf834ff0f0ec4c4ae2cf66abd08d3aa73fa74edb208dc55c71e27a232408f81f5ebd57099a6f7f5294f19ada8d6244177b5ad997c81469865d8b60a057602251e616bf629561f4339066c35f97b49ceed51e9e22eb33f5f090c94b3a94569ba1c503b23634e034bb399dab9a62b6154b33c7813eb5442729524d7bd47b3d902378c20f0ac2d77c1e9c057083b0487508f5814c90ff44edcf643e54a05dd7e84c64796d40e50207743151544fc59c2eb6021a39d8ee4d73bd50b3d4951700d8d476d7c90c3f464edfba29ec1bc14c7e62aa553233a049406c33aaafa7b08ed318827b3cc6857201ecd74fd5e99a3509311b6d655712bc618a5216203837a2d17302c2e3fc5ebefd98b4691156e27360981c461c935813ee0f7c15cb99f88f8f1e01435bc6535a3f9778f60f7d3cfb4220144e0935f4a10c7410ff42f351b311dab6ddc70aa3e11de65674b6bb321101e81c07e486578d6a9ba44d210918bf178ff1927a3455bb357a06cf1b9f241eec6ae72bf2fa586370a8eee76df0ed4de0b8ab05626fd872a7b9666ecf9981c8309a8428c6c7e9a93739b60f5a77fc8c71f191a1bc37e763cf7b890e8f118d77b5c2cff9152f3624418e9f3970bd143454bc05994a8c7db8c7c644cf75241296cf2fcd06fa313c0e974ebcebfdc6de06a7d1e1ef4669d75573bdb7df0dbd351103dbf0e405a3937afd5b70db51e825f309138411f9b36099f57cc7e00ffff71346f34ac197fa0b407677a8f9e11574e290dd44a2b05c11485db2c07b1b9d653a8c4a86563a0df453ad0c45ddb807b038696f12c0a7f26057b3f7512d3634a9d20e28fb9d288da39c7b0e37375f12806fc0b16b5bf8d57293d85d5b1af9019643665737406d9ac21b07b75c80f24fa82c2e4b283e259d59d7b6059a89c1eb6331556c13749d3a843208f534d71f2709a42a16b1338b6c29ddecc790895ee23c897b47003fc4976bed645034c439237829198ac6722f5a0f6aef8b1c4ac3d363398f988386d80fd0a4d22a03b093f061b2ba37f1dc01cd18b81d33443d7fbe1f431fad632e8045aec524e7fc0c8f5b952d6d68556e34faa5aaff2ebcd52b178454a90b2dbdfad800d45756309c7526f0cff7cddb13ddc7304a2030a18505b251e3453b3286d6609561b4916d6022cc2c38431ec5c31cbf7ff630845c2bc7424cc97984dc8827baaec1d2161d71557dd78b4500ebef89774c91352f67589e4195f0f7ecf777ff50e960a3546326b4e1911561779a8316c0f4f1b280ad5bfe2f3afa4c429921641a3e665bc704dd7d13ae750c1c49ea8f411b92285f79fb8241480b8c362e95c9d2f994799114c18277a154197cd97a642a4291576b2f50636d426483f0a6c96c87cd7003ce727a4bb11153f23a9de5f6a072785e4888d34a15c8925d3e60c8a3fb765fe1c332aeb56c3a8134b14f2f6667a96e30bde68b2e0cc34ae02d617a9d6b1ef7299e582d924f5be910bea4784a7890d89a9c15b5b097bdcfffb4d4d55519d5c798ed1caf5be3c909c5372a78f0bdc6191f6c606b2d57dd30044402cb66a03a6d51b16fbe13fd4545dcfda6c64b8854a43e99f62d715c912e8e9fd273e994c7b318de19586e032436ce4c9fb7c41099472363df7eb41a1c1d6ff7cdc4810b14920df86e5563be67a4595f133123c471f26dacf43b1226e5ac9e07a9627822ed86a05f96d91e12b108ccf792177b3c76c1e1c4ac0dc318cb17ab40bd471526355a58a3ec11a01b8960afac2edc56c66f7385ae483465aee276b005af3b207ae27b6d831f97fa43b9f2a90e61ec24cdd95317825c9190e1f5b5c834e113b6c", 0x1000}, {&(0x7f00000040c0)="e75822fbe9625ecd5b69f80d49cd60e8235d92c010578eb7194457c781b1d62f2ae5ca9a6224a7807e190e6927dc1e2faed1a8c5d6b26d8b73865aaceaf82faf11b78ffc6b2754a9869b237c61c4b299c0e968ad59841fe510aab91ae4c6cbbfa6f367b44621c88ec59b3b04b19eee7e3ba632f0acff67971f0c3761869724dd890c3f24ae0426f3c25bada8400f155265a2425a7a077d13925e73aebc28654087b2e49b8e9a22abd89f732d9b4c412b617a8251aa403a8f5c46f917854ad2a2644fb50ff543e2c8f45d3c4080a27e7eaf60fea40eb05d5a338dd92f", 0xdc}, {&(0x7f00000041c0)="b4a8a8bc5480413ee82f9e03e7c3c1d0adae77ea1e9425f4638c7a43bcf1e6242e43c43ef0942360f22a1bf554e0963ac78edd5e59cf0fab9dd7c700c18921ce9dcf4b014dd22596b161f8a4c1cfa836bbb8b7196cf7eb497b668aa34d208cdb3b63d9bcfe897fc5b89f667cc12681db3a8e79696fc89fd6567f9d44fa2d51374b7e43c28ca217152acf572f60a3f14f7581014f71302320c4db13f08ca10e177c50069cf02c3d9cbb09c79c423ee7291ba48fbe14a281056b33ef1117abe7e431686083812fcf654cedfe", 0xcb}, {&(0x7f00000042c0)="066c4fe0ea89fa64a1830ba9ffdd0c79177ecf0cf6016284edad1e5857c33a2019bd08df5c3af7b1b719e145d0a54df2cac25e04", 0x34}, {&(0x7f0000004300)="8b89f5f34b669f16768879a584ec59dec3db0b1e63c6c685169c2bd5a61b08cc49a5b1c3dcaef1f6845e04407903a0707c469d15fdff8c48b7", 0x39}, {&(0x7f0000004340)="0650f75f11651aa660c58c6d278159288253915b3bf991960c54aaab6ff65f0fecb237bea89616d962844df02ab6be53612df2ca1de4db4906b1a1e9e6765dc292b17452d2276d31c078aa8ec78e395aa81004d2e331cf2f6f261fb0127a217a5a4b9cb195d7a39fdee97c1869c03ec64ef8cb73a39ac810426473863a6d973ad2338b293cf62b8b06566b73dc5d80a51bd10ddd064eb476f7a718a7f2ffc61342489fd01b2d2f48a6e4048eaa6896ba763fe7ae62ec84cd2b8fb94ad245f746141e80e9fb", 0xc5}], 0x8, &(0x7f0000004500)=[@rights={{0x28, 0x1, 0x1, [r0, r0, r2, r0, r2, r14]}}, @cred={{0x1c, 0x1, 0x2, {r15, r5, 0xee00}}}, @rights={{0x14, 0x1, 0x1, [r16]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee01}}}], 0x80, 0x20004080}}, {{&(0x7f0000004580)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000056c0)=[{&(0x7f0000004600)="1b4117e9ccd919767ab836f65396ae33c1e0385605bcdad46d5b9815d5d9caf2d266d3ea1c74533b89cc9ac60309e585c6e7a0507edc190e49f6a5a5466b306404714a92f6b8d7334d7715ea0148e52d0301fa01dd2d24d4", 0x58}, {&(0x7f0000004680)="e6491f54ab4f517320840bf4ebd4a671dd885bbf0358c6528ae3a6dc6a8f25961218e0a8f2a4d30ede01b01698de4ce1e5c00a4d74050b2468d4f4dd2c340ade1775239cd304174ab769780be54949b7891d9b955b0cceba38b78523d95e5c73a2f676aac2207481aea5bc5b80e02077baee1064ce925b5c2da9bd909a393f044c60b61efe55cd8baf75906be93fbfbbb9b0c9f4a37f9da79a878e583eb5677e850d50e147de0e213419df215611f65ca95eb5ff58c6e5a6d675f5f94ddfb79a2eb912be6c2d5a6cf01a00eae0e51e304b99785f7b135e3569dcf41b5f679c81f06a52cac7e30cb66ea0dc5ea68f819d407713f3e0f85e58ba8996be080224395584e32ea9416edb78d142bb7c3c766cec59bb4e5c1c711a374cc5f1a51868f4f4ba535790fbde0b7d44141a0c94a0cda32c2f1c82cd58e9db4d4c01359d73ca269481e6ed056c7b7a1166c6384de2cf9cdb08797b0d922be283b907e5c63418ae4baafb73fc04a11a44eeec0580bc6910866ffffb66072afc4bcb9c0008009913c0c23b66654d760a369c6667eeeb0ed8b839f2a142cfb5203160afc068090b5906496e8db3f5ba90e525533c62025ab6026c443390155f7b06eb8eefc006af7701a1dc49b65e15c22e47f4be9a5b98bed4cd8be19bfcab9aa1c6999577aa23fbedc0d011c00ac8e95e08ba5c707450205edaa8ddb0ef2b582994dd9eea08a88d576a6b0f210f472494a20c38c0391a8cec3316f08a7ac4e134ac1628d7d48b41c624ae20c2ef5c37aca0017a71b32ff8fd9fd81e779b2e895a5e2263023a46140e5c9421f318ebe094247f4eadb35ecb6f7aafeaefff6d59a7353ac173d2c818bc2fb4c9cd81b3b0ed77a4db8c3711a0090aad543a4f8e85c121587c2b17ce81f12e61a3cf74e52bad62637b4ce9e55a1e0d3640728ab54bfe14c99a71cc0284fb540df6b88c871e983c45f9091f9329f59ca47a5b6d2d721509feb198d5a4512dad2f65a812c30438626b6e0f38c09e8766706fa5128cbe80bee12efe9fa8f874f2cce1ae27ac202b0dafd9553364d3c5e26450af67a54f78b64670c4eef95e703f2187ecd22344384aeaef9447b1064357c231f4c8e6bc1a8df0ebef5a68d49faa9f444af76adee21f8cfb46082025ba25a0c27ef859382bc8006817f3a3994bc867bc2b2e0acf66b904b2e49f7bbe27023c8c7f4bd27476522c8272e609e3ae9d23f058ca783f9b63b27737e0bf22d30a8bd03ae8b80075de1dbd44812ccbf3619ba93345ccd399db2ded1cf9bafbb00eb02efcb8503377af9701ee16107adc04fa3cadee04c6b2c54c167c4169dd8c75d01b89e46e0ad79f220b2d21ceded3f10fa17efd887260020e7e6d01e62ac02ca9956cd20f0fcfd462746ea94c8061c3cdd6bc1c1cc25ae625d0190d77603267930eb06447de81dacd18e2618d6a1c3370174569f88a74ccf3a9721ea5fccab573b29d4b4d2a2fc639abf544beb3a81fa1798f098a14030f4c8a4787193ca12b8b3b40f5910ba57bb404a669040ee58b3660cc47e1b740e62470dd67130273f4a517ab568733644457efba8d2dd32892115e9bc6706617828bedbc8d73f516c58eae6a1dacadebb18b8a4bc3228546822b4ebf971f64a077598e71737a296bcc7f124561cf2731bc3f2f28630dc957452fa346e898e558f7fe45b54de17ff48befea3c1b47acfced86c0e94559250d403e9d9c3c945c827c7baf433fc1d7e9814649f00c824993b97e3a963c83d85921d0780a52fdbd72334faf5d53036a8d9704406071681bd72381979d1c9172b43f5e4b9b3782df1bc6d3d638768a8307c1adcc8573dc0e93727d8c7f176fd2a9f91a1bc56150858d9b3449e3b5e34519198e468c1bc80bd3e53d4a29ef808e2e6e0c3a62ed16bc19545409380a3dc9f04adbe6c871965a6e8756e85ed6c58441957b92e9cf9abf7d4e33ec366c93702bebc3062782838af4076e7a48246112a0b509d8a5d4fc8ecc78add2926ce8dcaae744a94618d04a2ca1df7082150cc5f1d288aa414f3136e1356e23cfa32234cb94d1aeee195c0e37857a712bbe30901d187c47b4b45ae30fae64bea0a8672d3099117985543c0a3b9f3abf306f0ab539842a798fbddf1b25c1a69b33513834e6094fcd315a86756c09343ec56ecedf6f87d4b12e64381ae1233afed806fa1ee7855ae7300a425a7f3dd200f806545d7416622bc90de53c10a2f1eb8e66422fff528e87eed5701f5af00fa2f230354409e6ef4b62a9de22f2f8b65657a050b2f5c17dbc9ea9cb9796e66d376ac57559fc61622050c0fc0b266130c90703d29039b54819dd36faca85b7fa2930f41f3c90cc65530fb19686994091b5026161033d2cf14ace36e0a10a817c533c967ace67e14810f128d010305111bcd895177fb997e93301c882957a382429b1fc774bc7ddf36d745e807391c37236fe9fb44656abc6e664616146949a39b081c32daeaf5be7d010a5b9dd81942eac33aa997681b99770db5fe411d864f1cfff401a907915ad96b5ca3fde6c6b4fe98ca7edc26b5e7110cc7611350767e42343322b075f2622f624c566c4d7fbec163fbade3ddfbb6052384c5679f58fdc8728113e9b50cfca98d9ec92e252aa9f6df6f69ac2dc0d264fbccac4b54673905372e91305f1530d18f42189e65e69bda6adf5487c1aaa2de7932cce8f7ea4edafe076d67c8a54d00d7a2dbc8765b900dddf133a39561f12f16fd7bd696cd68abba582ef8f7249665337e5f3f9f65f8bb45d79606b1942c7108288833f98873970939d1a58153529d3039b8700e4ae245b95456da2368be31aeff7e9747aff3dbb5b60b155fc5f846bbb61ef2ebe416dcb1c6a9fd896f4052ee9e0f0e67e6105b0c13f708a480964d8bf0c95c5464688c0732175e5244bc39baf11d97763e7dc7ba8bc380adbe7a626c4275920e46c2684ccab605ca0709fa84b61cfa16312ba1d864a6f18c20a34f00f18687c3b9c2b0fd9604624a5a09cfcf35f61feec87e964dbc4d60a4ab0be55b4f5e4a5d9817269596c08fb8a4c3b3e23367c0923316ab5ac0252bd937cd5f7f47968df3270907b1622a6d6ce19dbbfe2ea4d1677755cf30e1458431f456a749a4dfaaf62222de25cf534e551ef0605fd027d8ae06497f57ad3d35ccc7b6318d17e93aaa0c9504174662e530e5d4e070afd660db7ded8554fb581fe6e9cdef2cc867e82b50cb43f1cbc64ee81edaa52a29c02faaf3b95aa62963907b9de04f657baf12037901edd46d8bf2922323d65a142a210fe6b7039bdea6c3dc318b838fa30856a3ad61ddb1a5d038eb6a1ce0a88266deebb1d1ca847fa5910b78841b04e043f513692beedc81fc0f1d1597a104ff42940e02c6b2b50f3f0bf3a523652c6333ece693c70750cc8864168bcc08bcbc52aec0dbda5b59c85746821ad425f0389fa53041048aa708b403278f3cf0e434bf35e6f0ef8b88d54bdf7a599b2b5ab74cd9fc116e252eaa8d734504695273f5f034123f375a0cc70fa25385f08fa2687e9dcf39e0a44eac24fed06fa62f744a00e124e8e7931b84ae3b8a8de6bc834c685cbc7b8bcb76989eed71d808d4eca8aefeb593fc85ac20c4bd737e841e2379696f9593aeff223c301524a81ef877edaf98aa87100b27b6031d45f4c054100294286bf23c6057e2a8da3c8a633a830e4a4ff697891783bcbb2ffee3a0d896b77542b3028e93812e7d2f004bcb62b903eebbeb23b427f2d38aa09fb35323a1de4ba708769ecde4cc5d9150afeeee6639faaa2b546b7183144d362290ee385b224e60f5834427160fce114d8c0558ec1d32b5e66bf2ecef4e0a3023dade4df26abc72efb9a1ed119d37962974295a5150c73ba27aa7bbd17c0b322e50f9b4a1c2405dcc2955cce0dd46b8572749d2b30a3900d0d5e53b7473d80e26adcf15671ad7f4082df0a8f535e21b2d2839e4db381ddfd4124accaccc1cd0e1d88f5279bc53a520e8622244ff6f19dc63ded3ee5b520765df03069c5f61fcecdad99c00512849a53f8b27518a9b43f32ad53103f6636fa65977816bb0029fb3fa799002eaf66590c1eb7ae34a0ba33ec988cb0266e720367c5d24cce4bee8cce08b20de5825f224ed1198858ed62cd3c399ad47cbff5b8dbf7a3aac9b9029c2ec336e916e80ae479d321d17535d28904f7196bab1f13a12c74dc896bba0c989d1a71a493002461f024adf6eeb5d2755cb7bd33d34feca85b66c66ec40993b3b5ce955cd853dbddf6971f88869e021bdea59a671047101299674bf854813c8d32b5d3a0b52bf7951d5508c4428770f8d63aeb35a2650cef4e7a9898e31156c74b843949c281219f1423769673a4f61e55b3af35869f065d458326ef954a1c2afccf3814c5ce91141467351b2c15e426c82fa5de611324d328cc296ac0fc625c1e9d7688d6c432feb2214fe2a0aad8467ccdf5638a1d9ac099d57979bb8077f11b2d7bc446bc67cb3d57b8f60fd09d380ebd8f31e7df0c07934f8394e98ac2c119417a9bfdbe46b4dca2fc21d4565ccb42616b67cdbbe8745c0b4c5d9c2834c1c3128185e3ff4b43ca15acef713eaa1055534602e27ebed79d33a4db2f99bfcdaa8ff1da188ebf7312474512349d1f9db1a3b5f8fb83f69024adb051cc5a002b2ac12a4d8cf6566947dbd9acf53596154e88439473e08a5d554c24fda4e68aada6ba1911e16994078a6d4bb16b5c73ed09e12a4e55363b48498fe18d5726c6db2df1fdcc598c7bc9f3a25233c300cf8e9e9afb68b756db9e72681e7c59acf74f03ba86178be153a6b2325b7dc9e1ae34e0bceb290ab4c274225eeb1ba36db9f5ce72f49129c470a846bcfac870ac1e3e2f096c60796633c4bb6ec07eed2f9e6be6a87e95436c79600046c49c3e9dcaf46f6d52b90234b8ab4c55b9b822fe529c72db6299163315f59214b980fd565b215950c2a7f45502b3089d627b3b5fcb791908612c8441a00da8b7bcfdc89166fb0645636d9a0d479b6ab5f10fcc373a4d61e88ccbe105686f63457f26869fe30423432cd56c9f9883de2365456601cd3330bf2cd79598427a82034312b2765be76563eeed60ce12ef6e50db0fbbfe0f29bc9f3d03a1301432e06917682b5a42a580c8580d146a9e15313e9aa6b5008d0143723bb094afebf9f47b2d6004c9a5ce89eef0287c98d0a2204aca3f8e1fa8f9b81906adab2b53d69799f55fc09a8b0ea87ea77b303022b52c6dab8f138cd71a0d681d312e0a7b482cf189ed93d156669a8c86dfa8673fa12c2ab18498a0fae991e6f5287e35326acd9b375962fa44ce69683fecb7b1f4afca0b8e75017548774fd51409635b20400ac194714bd998da37a939480b9148345c2b4ba5888a53803d81e0602ce0c2233e67d7c66629027dafcc4f5987ca79c57aafa21b4ebb60a716a4c90fd5b8e262998b12d6841112238676c10e55536fe0ae27de72ac71440576563e067b20412437b46210adda9cc1cc7197398698c138e090bbe4c22c866c66ab15e4046bd3e6f10aea90db4b16db898607af911e0c41451978d1468e774b88a200d292109f5a106cfc369b020db2dbe3675be3e6749d6579274c858675b9ffa64b0007acf916587fa16f90e21b75b409d1d58809a9edb04a6de29a8b21592121a312c1fc0df1d488a77a6546d5daa67c10ffdd40a955c9b8b7dbc71bf013405efa69b127593a6bc4d4ff211fb37880bb358e709720dcaeca3f81a0e9447f82b63e7cb584f1497a21c78a43527fc3ed8f3be395e48c0f78f71ff9c4b8c5", 0x1000}, {&(0x7f0000005680)="65658d9cbf0cd93430a072a9dcbaf7d94b", 0x11}], 0x3, &(0x7f0000005700), 0x0, 0x40000}}, {{&(0x7f0000005740)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000006a40)=[{&(0x7f00000057c0)="a704126d4b4ad9c4d814c3bbabd928c32f47e828fda9583d793fee07257d5fb74283a630", 0x24}, {&(0x7f0000005800)="8182839ceeb209564f5191a97f469ab55e9c47d3998f44882c2b48a175dcaf769061a0a4000daaeb7bb835ab2bca5119e6098cf81955067de253a6c72cf68157b45f6aa6c12a54f32556701de45ee8efb9658473029e68deab72ac4059aa48114a1faa919f2e3c9fc7369a0e408ef3ac2abb2b311641efbe187875532a9992e8da2a99d3dd85a81754e88bf83ac1e19eed6399", 0x93}, {&(0x7f00000058c0)="cd38b8d253f650df0b0996e7e4f4261e732ba65f2a3d3c10a9e2b7ba18c74f59abaf5182f309862fde98fd1d16814c8a1868a93215bb212b8f0ce3af19851875ac946c5127cf29c1d56531add7992bbd42e310ac315793aacc2b3af0b9d7e6fa13b0ef29584a65d7caf9b1c543b9b382d5207b05fde34d474e79a3563a66ff873d1c2dd362a4cd06bd5cc4b3efb6b347b623fbd89d8ca7ec33ed7d0b81e1c937c10dfb2992d26df73ce9536ca14d30f9e6c162a47f2d3492a1963cb8ca3e8cb2343f8301a286f96ee5ef9f4a8391b548b8e1aff201238be24c48e3dc304aecaf9a39d66338a8c01c715b080ea2a0ce7cba160b9dfeea0c293fdde9e2d792ea35044986e4cba76eefcab273756d43d990973b1b535e74507100af1c57e4f2e6fba71ba60ca60017e28db9e295fbc11d1912638a189ee41dc78ed801af4570bcb84fed67cb52c834a1e6c49b53463962c740eb150c3b6b4fac7813e2e632c9e2a5de0a6d53ffb01a8039080ecfd546e65ff133a9b4e11dbfbee8305e3f6202c68db606c0edafd99e45209936082ea652ff82436a6423edcff8ee5fa114aa50ce4bb2dbcaffeb0ad7f2c6f7c692cf3d8da260853c4312bbd016507962925ba973ad589d03e64bb0d6e4262318c055b9f09609849114ab21cd9424438be61e1d20fad1e40cf89476ff3b59dcaa83813157969b16f10b0f6472113557659506bbf5d4fa9d6ab2652ae18e81f33d548899bd3e282a9ff602af0556df8317f8e970e2426e42adc793b71e98824acfc7e88320239b6e51401ae68e0b8fb7740dc612020cd72adb006bc87dafb068039a8322f664cfa3d5f3c48c8a153fd428343f25863000f9be69038b4b5bb60efbc6da59e755d1856998903bc1c363f1c22697259f920f42b59dbd684d9e56e64d75c0ae8cdff66038d5a9d0052d520349508e461cfe93ef9592334dafa55d1f0d79add50e102ec388fc577eb4a41d00e88069b537a0534fae958d77929a10023ea54c0f2b48cdd3a301ca99f7245dca79532654402eb23758f3b978278af0e8f77b9dd264d61dccbcaa843b633329d6e82b644a6819a4974084de7c5caefa88c407e336b4db965d41eadf99bd8324beac0792154cf27eab2c417dfb1deb467ece9b899eb89a231d35aeeb1f996be166270011d433ba5c71d28ae9da02006797c616ae948ebd23e2c8d9eb78f2bc8d994ada2d6d58fbad5688a09a337b54c7b349957818b51d6b1b4c1cdd53fc4b0ad265dfb8c1548e7078911098a81e4f7d4686bde8b53eb2f59a5a5482db30db76579592e57c66dbcc5e2b8a9edd72ecf1ed579089f00e5deaa9fcd0fe83fa27b9400540428f7aed1a9b1687ad72feedac2cd0bec6501528d3fd0490b24d357bca6e41a0a7139ce24e8ec152da68df18be94d2a7fa7954c3721d9b7ce1887115ea92fda1cc6b1d17b1e8cca7b45c2e292aef68b868f2edc89ee1da523fae93ac285c1b1fb056e40d799a438d9ab67de4a63d7e02c083ac55770fc0d7634a90412fd164ecb8aa98897e3015c925eb5af3488be9a831696bced14e7de1ee5bec3857cc2cccd1a369895a8d952d04cc0674b0caedd373ebac80254d4d8a50a0895ad8deac32055a5a236019e2dfd4fcec7f96d99a4e63d9e5effdc2b9abf727b4ad07f1534ccfeac116ddd09ac61dff8be04479f219b5dce0367fa252f01728f2b2c0192296b510210e56a812ff5c49deb035b406e0992ae54c728433a7c36f3662de6f5368e366e2246b899bc2f900eb8c6117e4024f34f2934287636a23e095311c08e461ee1677e7b207b0cfc9912ecf34f70bab76da7c4bda39e4232669004c2a136fd26da09241184c7667cad249a3ae6602672edc9fb2e01c9b0d2692dbdbae1f91a55701b25e8ccf0379508ad5802a264d02be1d81b6563ada9a05891e83c931b74ef39982106511d25b5e4a09e6c01cd30f02867fffe69c7874c6c24f2b335dd280e9bfdd2a2c528fca866d8cd0d9a77a36d0bef47e4196c8ceff6e68cee5bdd9ac2396944161cf4ca285f9271c45cd0801e15ea8e5f9a5aa853e10bc18798f5420860dd78e8c0743ed23937dba6609af9aaf6b2e8c29a3c5f0535994e8839b94bfa2eeb956c40394cc50e2beb3a8652c3b1d2118d7bec6dd1d9d29dc61e3f7aaca7fb510bff16e12014896cc871f73986e52fc736191aed8ae52c3b8e70f3a384e3af1c897e7f1c192952557ad78fad1352283c8653e1cdd29b4ff4d1312f007a97acbbeee40a26f757505aa996f81c78b1eabc4247c284be502c29fbf1227c71a6ddb35339e22e28ff633b2497332e92040e16221f13f64120af2f73243dc3c8466e793f8992929122a43a6f668b16093049515553aed2af21516afbba23e9333e0694873ae575feeb95961b15acf660a158f646590deb13bfd72327fe87299d9a54ae802738ed2af5245856de36763c70b328efddbd991938eadee5698123bb5f50a8f8fa60c8399ed3fd82bc11de9f79887deb6187eae8bf3048fc5a5fcb2fc26de38f5a7598cfa1d280a20dbd6105fa660eb8f61445e292a3a498ca1420a4e7e730aca0aa3ad34f73074ee9b151ad179203e3a1b1a988501cb87351ccb62b49c04a734eed9ded6299f8baf2bf6f357de6d5f0175a40c93560a37c6a254e2b5a19693762b2cf815763091d4b71ed23df5ea270f6d834608f8f77fdc5ce20457c0d3227b6d68a72ee45396a69845436a6005997415343750a4051e5515f957b794b6f4a2243d677a7cea96dae1ef405150611f9eae89f16577d8dd2be2111d59a4989fca77623a0f860759f39bbc819c602449ce9eeb029871a270b8cbb6e74c87e95dbb4f6aa417011a7dc7fd13480c74ea88d1256fda8207550ebee89bc90d8e2cc93f86652034851a20d593adfe935bcec4319a52decb0ad94184af2ff7976d7eee7897a53181bda7f4af9650ddbabaac00336dcf94008d8b19f78a30d5bcc03276033eb0e6e967bf7f309b49077fd6b4db66faccf77c0006016fa6d6c1720cee2c517088a554292291573a3ac235b115f8242504da4ad82e17a00eef6d0d2fbd810e012e15db2d50a4f8d62e1826da893860984b915b2e3167a897a0efbe361ff3da1684aaac564f9f8ed6dc85766359da71ca7022693d1baaa901a809ae2d8694bdc19019e8982921505dcc325e3a07b99109db7444027d399ccfd841e884f31de3f8c8fd2ddc1898d96a2b49ebe29c3e97a87a5cb92eb1d51f0590462976bc5d3728845d94db6806688d4fa6df9cf392591e38a01da7b79a09472c87a4dd43d037bf6d0131108f4725eb12f603cafb7bca512938bf2eabfd8dc1d206d2c4493ca640f069814066b7db1473276d439ee98f81ed3cdaedf685242d4d452d5e0faa83bf86c614a40a26dc46b73567d2f87e8e95ceb2d08f73af3399458841a8bbdf02c8d239d17b4d1db3b628c1e62e6df89d2999b316a7b585c0c439e3b42b82bd273a7621bbfbbae0929f8af7ce2c9620687e231800b78d1641e2b991260cf6d0069ec285989aceef62c8d4e6ff9855950140079912df18067d326224f2315c8327f9b10959979823d2e21342a2ed1d6832173018a232fb6daed2c1e2dea1d63e63d50ff5d9a4c48bb60f45e2978a955b0bc11ae9d56cb993e97aaba84a773bbd09bb5848262500ab36c6c9c9e7c827126f22c20672d3804712dbd35f7eec27b6cb34f5403ad9ac72faa10adf7abc4d1b61161b7270d64e249f76de80b69a0cdd70f4a122c2f50ecc18d12be0a0fc234045d59d52d1eee68ed60b3d4fa3e5c1701633c6506f778422468376ff43eea9ef4cd25cd3c75d9544a7e48be58119a570364d10821ad626e52109cbb9b96906e8b7f432f221c1fc29876960fc04bd58e791f12478422349e5517ca4ef433b041bf50a780d5ec7b26efd5135e216bc1a63973056e3c7628410ddab4b89fe7c6abd1b728b0c378afe40bb73d2d279c1e0f87eb6c2dc6ef2d28a448b573846377b4709d3f77932f44e6e8d5f19f4101d1f531d9bda0382ff8d09f1b6ac0ae35231cdc7e5309c4aa35e9d7a0e08b14ddbe60a0b3720efe38b6b59b23d24be411a2bbf5f4591af6eebafcb788be7335537198443e573af4ed8ddb7b2ee1dbe1ca92d66fca894bfc97a8f18134558d0bee4cc5affb3d6cf6261350b7ae90241199e618bc23f6c867f1a7511e4188a48c38c2a50303d3bdd142d1470591b0bf9281f0234f1238aea70115b4a136e4ff918c9148aa73ebb46972ce8d8703f9fc8cffd3be77c5fa3f407b5d37c7e9605e698ff1b07e4c7434a89abe4662526121c1083b4b78d31e46c8a7ce844d4168effb4803992f892934708f16b91e0456aa0fe0fe24976a1cfc293fc08146cfe1c7e443d0782e5cc59901400e4b0376146188c65b321942435ece29b050b3b53da9f3b40d558c778ee8934e2c7317f8d51aa9e6b4318223b50c322def9192c7320036017598f946f3a6a53337666dbbd6e7fbeeb9324465ecf491b076cc7315e9768355866372154f0bbe4c925c6dfdd0e4098e5f63c7dc2b949a2e3a947e64d48802e41552cc0b880a6602b8c70b2eef8050510a78b664bb0abb46352b4e7379bc29435fd722993454cdf7fde741c3f4ef0f53b095a5f45908c357ebabee4ae8f42ad916e521a43b0126c8bd71900f49548d441ee68da409b0ba81fef8b4ab10b9435b5f4ea059db4fe42b162094ea0d04869bb69f1862a84b509041bf3fa40426f5d56f6733a9b1911e6dde15a9011ed60ba4958fd8c5edab857c66fe2026d445cd4a6c94a51ea07fc4f66ace3b787dcd81ba6fef0cb53852614322382a4515fcb6c7f63c2cfe417bfb6f4683da484ac663adc5f63638ce96d5fa2a204c6f4b27f9cdd3fb0c020615248a9105d7dde687aba67252e1c52f03b5e43669fd4e78e5c1b7ef001739eb19cd40e3d5b840b0c899b5900b5725267ef920ff704fc9a224d3c1e95eb70dafaff84911a04a304f5b86bc29bf91bbe520fc66f98f1b9750f0cd4c295090d3614f3ee6a8a35a7e6efa8e10f1d309187ca2f2ce19dbb328a3d58d5fe911ea6151eb038bd8f49a756aa34b337eae60e8f8724109efb75779e99e8117d45103041711ce1f1dd8c339909126d993d73d3dd8f21c6b95960f095f7855fd5513dc5d0a963633c2d9409158954e0895e567b3557927937ab81f9468691595538f6825e9a8e1880c403c80ff15ddb4ec8da639ea7abbbbf09d53cc103f0a3425e452fbc1cfd6d4453d9650c0c3ee54fffd3d5f60d44ed4eb54cb1577ea3ba90bfeda60bdc241d2675e62c26922dc75f97133dc362d47bcd2869cfa50e6ca1f5e89524adbdba230875d2c26a19e038cf43db4940231607f1584c623a6d2b11d2f9bd4d0b7937d8eb3578cfc8b5a7f2cbfe13cbc67fb6d697b17083f0671dea373cdde8112d2a5bb95150aa2a8f43697c89f86a56c1d2138b71778143183f6ab1aa137953db67a4ee1218b5c68e8e231bd8a555c1cea79b9cb0c23ac2dcb7a3dbeb58b783f0fbad9dd35a5585deccc74e3e2b7d60db528253c8774514a9bf0ad47363f42f349a43dc40f921ca8ff04dc347be210714a4923373bf440a7e4fb4d2c202757daf8e899d1999e62922bd3afc3feaa4097474c4dd275547aa69e1e5af9a50bedcf25caa99316cd38e309949b8818a402c31f15740a919378cf1feac2b2166f8558218c9e30bccee7b0b494801584c210d1a1d5cc50c3cc674eab6be778004fcab926cbf184fd52bcbb52792370015103e47d87b3f8a0a5f5d24d126e2cd72b2", 0x1000}, {&(0x7f00000068c0)="c513f3367b8f7a0277ebecae305000bf3a52be4deacd78cc62ff68fad2645546642e90157c76fb830105aac6d1cc62a2d63de0cb8dc4d26350b924e5db6ab5bc1125d258a82a6abdc6c00c7cdf81b21ddd4a5906", 0x54}, {&(0x7f0000006940)="b3ca586de4419c10140f129c6fcef0495df6f4ba81f27c5445383f52b898bab2a06e92d378009324f59670c0a6594d1cebe8f24595e0059f7547e3fe092fdeaf81938c3904bb4d6dd8b1f171306ad0fe4a6a43051686166f6c94561a2bb25a1f8aef82af04e9dbba49711c44712929e9cc4d740585059f013ed7d2b9478ef16caf8ba64787083e0b8861b09f7248d3b3cf05bd4b55fa09ef80bb7564002da96e70b75224a720ed6364fef125b730f18beb30bef60a351451dd22ed7dc2db8afb1bf76080115ced0a23c8b02581b804aeb09ca5dab23bf67a5c38bc7bf34f", 0xde}], 0x5, &(0x7f0000006ac0)=[@rights={{0x10}}], 0x10, 0x4}}, {{&(0x7f0000006b00)=@abs={0x0, 0x0, 0x4e23}, 0x6e, &(0x7f0000007e00)=[{&(0x7f0000006b80)="965548104755f4029ef58e0dac0dd62e9a043075d648305451d73a036e24a0818a1db2e62fb30ea10834d5cee3f9344b57077e2649d6bc46382939ce3ac0924632fa683504bfc7d34bfc19d0004c131b3d538f0daa76d0d2b8b76597a0f02b7968b745148f091fdda7fa45bbb0b249d44f3cf5414b6c4d3641d10bfd7f4d7de5db887ca8789e9e0f777b03b42d8adcf16ca071b0afa7accb91eb917db2cf658905e61cb2b4997ea00e0a32ed669a19ece1388e5d43b82939dd64d352debba5a70a5d", 0xc2}, {&(0x7f0000006c80)="37175124ee5fb82c1ac12eba6c3ee7d3441d50aca21f760a67e9d5ecebf6a2ba6c281ca9d75dcd633a758792a6be051135445610d33719237f7b1916150db81392a7fcc17f6d0eb0e69f47f912e4c212be40e1b0ce458fcde660957c85d2a8413847984f233b904ad25330960694ae6a31b9b967da003871b9371517e4cb50899a8fe02b1493ee8b4be5649edb24d3f2cb80d3502a7eccedde976d2122e66ec175cf3db6fa5376acaf03c8e936a78f7186fdd95f9a769a00e9f29b8de8f7a8e2b9ff20ae416430f315be3e8c34d698946190e4a55780a22a756ddbfa9b7e20324333e5d04d39ff5563c0d7ba70b71bfc", 0xf0}, {&(0x7f0000006d80)="b69420a159faf034189493ade7323ad83dc73f8e9c50d7ea1934862def3dac1cd7298d61f704ec7a75485daa965fea64d8b4e02aff6c17884835bfaa72c02a7471a4e14d4d1eafd987c0502d75c98869fe30c09a7d2170a0d0470798a444519e06f793beebd2d31d0ad49ce0de0a1ec1f30cbc4fb395968517358a8059dddd224b8a1ab6ed828fce23e6f037bbc02b7cb239250474ffc6ecc5ab263715b8e95cd0d66fb57ae00e7bb9e6eb6c29e7cd1ec8df382ad9c40d32725197bdb6f1c92014f1fdc29b4c308de69b1dd7428028060e35695993353549a20b0e407e9e00374ae8109ff0855bf138f569db585b34dbe1be2781ef3e90ab2e8aa929960215ec9d35e4d9f081dd298e008546241cfd847efc50d154a2cf3f710dd6df1a35697baa49c5bfdb1935b27295dcffcb8e4c2397dadebbebd4f5581be7f90d52422a7c50693193fa64f61d6f65e77443f160387ea39147e7670b9112151d93e69e0265ced1839f5523c97bfe1469f7d0bb95758c66b0c0d8ad018d8949bc6f24fb5bca631383c7b79d6b83710ed744ec82fa87f9a9587df162092275e2eb016c00173c4fe9524884ef18dc02d3633acabdb4aadb3de51f9b1a28088d9ea8fe2a69c97eed0e6f05113a5a89b744fad9f5fe8a184450824795ee7cc7947b4a8708fb3990e00e6dda13ca4a4c2e6e5c98db0d7f2b4cded79cebd6e9facc5dc95ca6d92f85065f8a155f86acb2cd15f86a9e6640cbe6f1ac9089f349c660368ff6cd22f55c1ef59650bb463e5c2768b6b9981a5618bd0e79b8d49957c2e6492e70605c72f27a177c093159aff0f499072c508628fac9896986f7ba3bb08d2681e493207517cd9f45eb42d79c2325a6c305c520773f2d0e5f1fef2c0de2af854159357cd86d9f19c8a7c5c52876d6e7f352ce0f60337f433e8234cf25b3b830b84a0517f66b2bdaf6a69ba03c90e2c50f61caf56a156d1083d069f590e9230a6c077e1b5bbba16efbd0f26f3733dc2a7e8709492bbd8ba95e229ba7bf71be41c554b6d320d5a35e0770c925bca0ad84b306483477ebc90513e1dad9b710f855a67251c4695002bb93527cf952a3e0b8c9f24473f7f8653f8a64d96cb81f0465d2af5c688755bde4b447ddb98130627d0dd176f5280492874e2a86f190e16836a934589952a000cdad6a3e308a90b7502487cb953853b0a2996b47c204de7a9dc39e2f6672c630ae635908d4e4c9b1126ca50090a4f4131d71998d12170d31fc85de3a2f9c2188f2cd3b3924df80788b840de781a070477ed5b37cba1d435dd0f44595f98bb09db1d738e72aec9519a0a5fc3b664f750fbd22dd0da6324fa60058f500cc5a38e7f0afd9f333718efd8d023cc31491e822512fd47ef3fcc17c9899a28264148ed00f7015bf062b30809686b1d6b05b1a7ffbdcf8d2d3473182998620066070c7d41c4d98d1f2b6bae35b8f7030066f458de7db2e6ca501df4189b7f45c0ef2e406d4449b9f33f57064fe3ec7ade681cba53f2b4935598115acad0516cdb0860813ff6162f5d13dbe4fc4cf8442d7c49e7ed86a9940282a57ee80adca60fa1b3eb33bb7f8f2f474d1f3730862046f3158b8dc3eec4d529b44ff40093ac8f6edc3e9aa6e2de0cc6f3f75e7b51ef890475a3002aaa2d590e7c5c9a7fce56959731a88c487060c2bab12dd65fc45cbbdf4543a0638cef222db7abb9a84381db17d1d1554a865b5a9ca57f66a061e16780c8746470f9e06001620133fe637b99b813101eaf9f499a66ad3a86feb68b1e39794d20e91a4e4e5bae36139e18bc9cc1c86f575a178990f632af792fb50157175ddd858edb68d61dcdac44f630e8c8cc210b389b8ce9da4d8de0abf11c03e87cb7d8e64f7e0435ea5b4b22952f49bcdbc242aadb0f56ef2fa2f9a5fb34e4b459c4f0e26e13c3cade13513f5372049ac28c75a112ae79292cb0b10b872f13e57f3817e3ea09164aa137836c62264ca8bc4ef3c6c231ca6ce82fec20d128030f1fb7f4a38d4f41e70add7751e1c0eb74a314728de6ce9d360219e16ee2ec87ba1ad24bd2130cec9100a26b0a0032cab3e8424029d6040a3d7ec8c8c284085c9c2a5913a86ee3678ccb7fb547b48404db399947a138248c3e5818eac984dceb9590bc9afd79e75afb7e9c8a39db580623c943545a94678cfd07ad38306361cf119412b431753ae1fcb5f0a2b4f87fb1d70a95cce8831e0eafeefda67c68c27221d74e562e35e36c29a665477a462942552159b8cfe7ab7e66e8f264b38db29a5f6e83b1f06a583c8b6247cc9207fbaaebb0d706c5bbfb32ff267c591a3666727b8f1b5ae3540978c94720c95c7acceff303cf955c2d93fc45a1a7777748ed04b4c576a2f68b5597627e32fabc723e030b6fa0e7fac9bcac65c2ffd9c45a35ee57ce8af209277ad11fdb80f1ec3b98ec396e06160751922c33de2fe99b803089cdd973c01418aad923f1beccd9662087cb7858f3ba8b4cf02e70b2ce2099a41d1940dca9725b658d9aaaa4a35a97db6272c74948d36deaba15ac9c72da0e3813e5691ce3b7e78399bab4a356d083f53cd6f365680ec7d0958e9235574316e9e732af8efa4dd2637e1001f031a4efe45c3a44e76d05bd7f8dd2c6d0fc4630c1477bde7e8a91ea35e9ca151f15f2cc4a4f5c27dddb1c07f3e922ce01b0fc471942b2399e6c959124c140bcc2c15f5b1f2f3e9c9a53f6704de3c64737ccca688ea56cbaeca5d4da21c9e6a0fcca0ca69299e07708217df5e244b70577d8e5f156e25cc06083561e8d777a79243f3d365ec0a0e4343c14a790a40dd09c8c791b5f8e6b6175e9f00067eccd6b619c4bc1dce9460ccd4a8255d307c153dddf9df1692fa7737d0c8109dd23fafbf814bfe18518c02dc35a8030e89b8fa49c1396371f858d55131b5bd05a7bbab80796831cc8ec0c37b936c9cbe1b108737d75dad8e0916730070dd6f22f5d50fe477d0bd72bb510dbf72032625ea1d638ecc9b671cb79ecf6dfe4d6ed186afc7f511ad7ae46f510bf22c92ef97732f1774d5f89549b6bd37eec1f5046f9e09f1370c6b96bf8e05d480686efb3afc4b04e634f3a11cd41067ce2b889e9ec678a3b8eb202f3aae3d6c663ac9d704898b2d905a057bf30f1f86428f88bd6227da6fed209e0e19c42a333d7263f41417c9de6d24e6593b8ca5fbc4f27802e2059fb3ea9f8f7460b52823af6d217940cbd9e74e2e8e538c681aeccbfc3c5e7ed3585f44e9a51dbf4748f485b77c63da632f6fad5bc6ef5dfdcd75b2fcc820b12087700a6884f2f2348f3936b2a7cfdfc8e59eaa1a326b3a686c57ff7cd5f64a515a2b74046ec0b60ac09b033c19913e24dde13cc7816bfe43505ed75697ba7ffe95c7aec475191e44e41363fa342c1973e1798e22b190fb23c5e2da2eab0a29321e7a4de2de38e91625f38e6ead720dc8872d47ab3dd93433b9fdc4244595331cfc6b070082f1e7df785fde9e8760967bf2e6b27d4f2d43b26b255410596d36a12add1fef87c920e8d19d63f35007fbf554bb8bfce40fa345e7c41907cc76402922960a8219974c2e766ea003597ce01f691330b746185fb680d7ed0dbac495de35e8b3eef8f87c2a708a6ee3db9b83ca44afddd348bb8ce5144efdba18643b520bc1b2a02f2cdb670c11b87447412af1fe9138d2f93aebd8eaa797d6b2ad4dac9f9cb76ed65039cfc9392ba8a13227c66698096de48552b082f86bcb4db7647ed49c6387c699bc8a462c948059a64337f9de89e728e6127619354418b83ad4d51f62f3f6547e1fbce6407647cb692639ba5652aa906c41ff8a25646cdecf68e3d86ba41868af948640786fc84b6b7f507f8b37a2679fa886aee65d2b0c58e64542e7f75ff7339c94be5e52e7fff0e066b89fb21785d07fdd20c05e58f8c0a605e78bc7f693a49ea819682c870207728c2d08cafd24366b5c96449929a1f74359c222ff51e681c69d22e711b6e5bf4e210cc643b96aeb079cb0b21af7a13ec35d3931d07be1627fb61852c161ae2d7b043078e598fa6ab98162d00330762a47caa4e830c8465aef13fb9a14b94ff3792793edca46c163b953bb9535680720067f25ab81ae8aa3cc9ba507afe82f271a93ed81f7297723978afff8295ca10bc6caa804460dda1a1ba1af50e8ebe79703b974c97c26618fe78b13e6bd4b3e15ba80652e5ac06b4677029b8e627efe1cf345c41433268367f3e5e418ce0782ca6914182e4fadbb86e48c9912e1925826cbc8bfd10aa9f0f3566277988dfdf12822fee10582904f6ee6a4d1ae2e0bcb06882d322e3705cfc7d857a5de8c721c36246ef734094dee0c16e4bc74e320893873353f749f4e7f293432c277f76a241b6078fe2bb6ca828ad4d93914a66bfe9d331719c42eb58b84a7667b31598ed7ef569e9a3bfbff30469830ef03d35dad5749b1ddd5a077e0a6116febe9a7cf7de5d1af0f47dcc23e3e8d29dd297911038aea4a2f037eca691789e1eaa2c2c74c6a792655a0222656528ba39bb76c709eccfcc55be1d64048c5f3054847279b993543a2562c3753769a5eb1c53e8f5f7e65db21196e9b3f45bef7f9fa0659af35629bc8e1279cfabf40f23911799928adc50263f3a1157794cdfcfec2de2a99cc2cfceb2ab2c75f280faf3da3ede058661762dc5433b012fd916d71ca351a0ddb33ea6099d746a0b5644d19cf15a6d8dd9c86a4f92fd771ec9792c5cb19a0f079707db61d37d497811bbef290145ef75b04811818b0423b9a0268f599986c9f2ba5a21aa3e1bbc159292cc1cfa8b95c964a7c84a7653920359ca32f74892c7554f0b00bd3dc2686f9a7433356fda08c8ecb06dd75530047ae45db3edcf8201a4fc7e8247cff7f602de9cda133ac0471f5c86faa9f3a381c9010d4ed739fe8e29be596a1ae8caa1d9b37d9ef9f5d00778e125cd7902072f9b72fee1947faa6977f84e501cc9a1d26e312599c01aaef89d750f6884ad951131d57caf1323ad63a337490aa673aaff091af6fac4eaaca1230cfe670b4a7d829ee109bc22460391e6d2198c70baff41b58759308ff2db41829cf78af75ccfdc907a94932319e1c29a247b19fd249dd92f7d1232bd7acabf81645383d8728e0c58ad87e368b1d124f1f7d03621e9a4191a981a739a3c8a337a6b4d2d548dae68a893181d29f5b1ad592bab974b8b1b2282c4c8060a358aab6adc826f080732ba5b02dd906ab4e67c4a7ef4404082c1d83d2831f155282631fb782df01e9aab40ae5d44f49d38d3adb614c670ae69a3969632940c3f147fc963ae49674764e048bc3aca7c4d8472c7e5c5fc16c1e6de2b0384558cc43c00caa226ace14044b65873e6cf4443fc4723cf451774b296cffa866a8fa66a99eab74062f39ec8c3d7bf7b585b5b8f47645634774723c9cecfb4fc44dc16000b3c0b7834843b136a6ee3ad9506f8e0f59dcea7bc4975ad611c7bd09f9e25baccbc6c9efa5aaa440e8c07160f4cfe60a02457824133063ef4e8809d53a1621236b757967825a5896cead3f2317f3726fbed57f6b91156bad8f7b9cefd74a51e74634f771195fb2a48f229fbfbe114b79444823262b7e4427550badc16f1f691b0945c2c829075db54b67b8853a78b8adbf228fc45a9d82e2a46b0e3412951ef18cead026ddd00f6f1fba6f050759b54c1ebd5186eaf069509dd5fd3207e0fec179d3bb4a633d37bf1eb591f6138334c464db5b710a8e4b1484ac92700c0d28f04335670563fc1686628890b83a966e05928147599dd1e778365b162e8f745515f4edd3", 0x1000}, {&(0x7f0000007d80)="025c0ee5b27eb267095ba74b3f3b694a104c54cdd14c33ec7038fee3b43b724de1332824e514d4eb15968206fa3d7f5011dff0d85220d0a01e90d41476470b83e50bc878771970fc3a825c", 0x4b}], 0x4, &(0x7f0000007e80)=[@cred={{0x1c, 0x1, 0x2, {r4, r5}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee00}}}, @rights={{0x24, 0x1, 0x1, [r1, r17, r2, r1, r18]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5}}}, @rights={{0x1c, 0x1, 0x1, [r19, r1, r0]}}], 0xa8, 0x8004}}, {{0x0, 0x0, &(0x7f000000a000)=[{&(0x7f0000007f40)="15efc9b2550993b1a0fb0036c17affc7f0acb0e40feb6ac0accea9ea3bb729f22013e8800b532b7eb102dccc59370cc4487b247d8b22def4fc9f34ca8b77eeb82c03f9cc2e3aa157c922c12f7bd55c0c3a86aff67f499cdbfa4287653f23fe3076de6d480715b756c803366783e30225e621de0e167f1c0e184f1c91ea833d343973dca5ed0e1c949df26f3f6178", 0x8e}, {&(0x7f0000008000)="02599d24d4f1543923a8f7e2494f77e08c86e276c40624596bcfad95537947f27efb0a18d35d36ae2d26d1cacf8bc286113c754cb05d7046ee581be96cf7b34f599b4e385ef995a5a9fcda67178455dc4f238901797222d4befa6da6ed5409a2b2779b4f66b9c622e30970a14a027ceb88b6eb4c7261a1ac314b07af4bc82ab981748570f782fa0b0921a16a996e04f6a6ffd3339e163a79a2c22745b1614717d718599f25fb2cb553817655c89fc57dd0d897e68b9d440ae3843f9c3b4cce7e3b67e46faaf5d6d592d6f858173c1c7bce902e5c3a9e23cc3ee832797c81aeb049a47e0c3b780d972f369d1b967557385dcc6bdb3bae02dc25c8cf70123d15d5c0ebcf02c2118ff6c59f1150a0c6e1102cc5ecbb91b4d0271bfdb7eb2fb1dc4af06f41a46d84469ac867913de1bcd3e4099d838daa71e70bdfe8e20d55c87c69384dc38b2ecaffc82f531d92efee548e021b0e8c61b5c63fa23513aca31dd963fbad443f31ead140dcc63a68ec387ac55bb9fd8948315249a9a058b01e85a6c1d9bd04c4aacf13ee4069a1f3d9c583acbb51dadade58c90720c1cc03b4d0385df4409c0a74fb54a02ceba88928c80e9bf541be02f44112efdebc8285214caafb43d77391751637dfcf9f2a9947c962be47edbdee035c95d1d5c9987257df2dc82cb39f19c6fa4bf2752824eaa26fa1c299c9b0d237ba290f74fc8fd5275abed35cda588eba30aba211f053465eca6918cf8d1909a41adbb2b7f41ae6adac0068c0234535586dae12af803ddaea6028bab83876565bc34c49d12fea754006a2f4c317b402aab989372771d2664a8a01ef46dd3bd2a8c367f1376fb495f54d4156e8a040e37eb5efeae899175cfc1bfd66eadd24084655433c108a5c8de2342bf9c166d2d1e67fe07f707b6a273ce10bb7fd4f09b5211a1602465c520eeb8c14dc47098a39d473fd7304fa28e48809ac87817e41167d97cf139a999195de316ac112909025bbbb63f5f84f5dcc879d3832b12f24e6f317c2692ad74e76617183e432cd5eb470fd679ff3773432e15bd6d4c8ffe73d2f0074eeb49d6dd64be98778daf269d5cd636f7c91ea515874edd88eabf4c1e32cfe3b8f27508c8f932d48a580b3fdc4ff7aef6d88b04c3223c928cff8ee7f7fd47b39ecfbe1faba5c9910715f3744b2ce685b7c806f4a81eb973f00c12ef3f0b1ac96aefff7e1e5a35f5fc2da16e0fde338b92d9048bf555cd65e6e1cd45bd1974edc0ae89f84470ed865edd52b81918c8cf00ba3a2c83d49b519349744f47c42cd8bbf685288c0efb84b4305c1da387f21e2455563ef41fc3f1b17f85696ebef9bb39fa917a7cc01616da835c6be73004fd8abd9d4027ece59f7c67a1019d9570e040c9551007c512e5e12f0b6f3ceb5070503d2cccb283ce51e3f350ca83c0c3a058860c07be80219b9cb14640a5c3984a1a0d95c27969e5a857cba99f55e17e0534e39136f72a51ee81b661f8637f963dd0a2e52ead459f6eb90933ea10020954cc5372122a412509254829707467a1f6daf141fcc9c29fd5baa3b0f0a7c6cf878cf78104472dca96a9cdb19bdae205d4fa79f533d40af0c782f50f4208417beea7121ab2f6d025c3e1e926209b43f46618b8b257a6fa736931e264b82dd55d9d6c23b8e170844ecaadc4c4b479e36cb05b2303ab8d24af1a89121419483fbe6121ab50af9e48f3e9603b7383914a0de869b38c074a4e7ffc7578559ce04c8438be8363561e5b3f953b82c4ce10059378d748800783c5e5de55c61fa6b15e9cb094b5f24dbb6b550c16549bf457cb00aae098389a404f2afa03b49567b88f309973b3a6bf206c7081b980c4abeb0197ad54b9e97dff610de519a4b8b057a4ebccc08e27dc84c45ddda1e5df1f212fe8834403c6b5304ffe9245a5469d53647ef0d214d82f51786b176369ca5e5d8e0dc7409b9c969b85fb43816919cfce9ac5bec76f82be3d16df8b1f9e24db81fff072c72216ede52362a2c370c11e1dd2f79325c6d85dfd608b8a8f2556bfd5145b0555631860f02a083ec6b0b3f3bf84bcc7a04a9121f5c32aab482a7023e78bbb414a362755f55bbedadb3d7130ad1637de86e3abd08827d4ded9d8307ea5b90b94f2e8afe73ab9acc2eeafc21c923cee4f6f5c710c068f4a9a0eb1e3c32ea36cb115eaf4f66d8f6d067babce0aa341763aa57202a55a6656057d60107ca7bd7cbb5710f3ceac160becf95a8262416be1d876b9d9e1430246ff041503f5aba8153bf9fd5ada50431631c0f2a062eff4d03119500b854561db8b439bf5fc50ebbe88fe77e751893720b2d30cf9103fcb8332bae1c283c0ed748d09f3617d3f54d8a661ac8e3ff3cc759884537844ed9b44ccc64944412f0a2c87faece13585cfd9c28477b392109d46462e3025fe86a49bd7fcd6675f8ebac0b1fc884c8ed5b559b741958e752cf68abd90476c7865637a526364e91e261b1fa0b3f15e1c74917c5097493b52cafb9106b13e805e43f35667b8bbca206b79745a97a3544bfffefd396273080e30b5a9c3ac22cb2bf2ed149e6c669a834f6168dd31ff2f6f54b139dffe31895ce3ca52db8dc2fd00c92bf62b4eba771229dbbb377fce289267d4d4c920f6c15db728d0acb523bbc5443868a9f1fc4b0645a55f3c966cf7082de0afc0c941ea5bc4705bef3fd903ae8b9185323ca52717be09c8da33a7b8e421d768e7fc945fb1e4ced28e21c2a3b9221b11a9348989130451aa177e6362a964c61e98c09cfe74812d9c722d418a34a9f4b0bbe76c97051a8ba4b9bfe88e737c946bec63897d02d57a0e422b06480b29ab742116c453cb9c0c81aa4bc522fe7865d641847a42de4188cde311751f7bff6c24a2177f9738c35676103c17ee65629f08abdac11cca5a63a84b0b4ba713d0acc6ad8403a50dc16b50e1fe2c46b2a7ae3e6ec483bc8fdefe4c440ab32fcc858c9fc0c39a120174c8992cb3be31a2c9c4c0f913ee91a1dbc6ae99d741dabadf461444259202c7ed01b2ef5ea8e6ea78b0f97e94e6e2b885a2bbdae86467aa59a42b4532dbe744b22a75c94f5055ab6c084cdc433073c32db9f6b224c619fa4013bfef91d68592c42bcdd69335b84f6137571b968f63d9562a936b1a1ec6429e44f737e8cc1243d65ca8e00e134deb20587daf23cebe203c3b3e87b6a908ce3f0fd716ce360d3df059f718458c83876ff241e5dda5045798019505fdc40a0089cd229c8ac8bb71d86e00b8c002aa953bc93ea919db01b7d6099c94505206750a0e76f58242109b751b15a4d357da7cd93a5408331278d2045c18f1396aa4bc26220091c3bda1d2aafeaef03558cf945df1dd270a0efcdd0d15df5a5b341c7596103c79a4ee1efd1b5021364d3b8dcab9dc8f76c73d90438d86c56f1229c4e9c79d4f55e52f7a2219fe7fedca07c3bc5b83841986cc28db1f0505f39c3c511ed752339b86ba529daed8e166705d5881cf24ecd0d653243cbba4c882fa5efafc77c9e14917d195a98ea55729d80126f4234e05c079d327e6b4323be8c901c1c98ca58aaa544ae66c39f515722f7f9d3a95a7e7d3bb1df7e20cc2d828dae935b5a2819b766f56b808a932291379606d7ec6a855f329893d4c9f74d03e0a5a374489ec752615b0d02c338ab35c8cdec5830975eaea7f0f44f615836d30bb2b6304184d5b7acc96de04de860664b65c591d11fc63851c068bf36b68d1c07544c4bf007293f4969137b03373360b6249a98f14aecc074ab1a86960b3ab3c81e44f5e77707a038a6c3b51aede78c38b58ec7c97f5fc13e9a1bf492056328a90e9c45948fb4447b398ddada3b831b9b1dddf1ae3984b478a185c2e644d72c8a84b0837d7d3690262e92b113e6e84272be13963bd714886300a8a2b5c3a804f8e3fab70a85e93389c6127da112a0b8b391ead2dd1c971cdeac81e042620b74a383969ed4e458779aa30a0bc1dca997b5e833896fcf9ad3205d3301e233cad43324fa30bb3cdd04ecf4bcd13994854ee51484567ad079b98f4c3e11c6772027a05959d0482d09ca9108c4d35401c9ba2bd3ea2c2f66a23292fc0ee6616739c0438b05cd2da4498ee4dd78864517c21a1d32fe9f0883f2bbdcfa8591dc9d017866d78cea38d8b944e596b5d2d7476687e846c511888db78bd16034e4fadc7dfb23e301502ae9887bfc19b745a47df5677d3608e224b26124e46a075dab373029cfdce53752b1001e0259915777e8583bff43615c204575a2030033ca8a803cda650f0caf0f10cc73db5aa3c9429dc284a4751ec07042c681a04c54c36ef172d97b78a3d8470055508883d55bced1797279790cb39022b166ddcce7d6c50e64d93ca498fb060de42b503ab7cbec38a6675445933d649b98e4a21e7327145d5211df01311c21d27b1efe3db301b117ce431e992c022ac6aa447c7086e49f82835768bf2f4455491e0928fe7d987a84f870564962f238b3037c2678b1427d931aa2c0c0a7473ed0d55316eafd48c988d2ca455330bd8196876134d9e5162a0e1b71399f7607955d4564543e4081cc5b43c860e874d270f021d93c64d8eefc319afcbcca82f8ea4ef1ab8ef8e60a94de97e74e6974d7813e7275ca1ddf2763b19b2159433849d427e5ad49c7b54db8710c7590f3d6d52f6fd10a0df64f36cfa6c56c9341bab34f248741ba2d262efef10d0df7596a5111ccb4cdf20e45bba0a5fd04f72afc8930c6b3d8946f7efabdae50d797503bd4d21b95472bf8dc6d8f943c9f4b4ffd9ec24dcf22aab49b143ab41506a3cad30e39b9ef59ca7f1df664bc8b72c7b53701e7833e45e8867c48c14919393f9c489e14ef1c3e665bff5fcc45582b94507a1e2e15b8ef71e249ead603e2ffb744172117b78965a055fbc4c0e9af72239f2379ea8b611869f8ad3a17f4c8474523cca66e3890d7616225a2e9051b865466b6a277b97f05b7b4ea1e05be4cb72eba6a8a7bdc0acea61389c2a89657febfda54acaf3b83570b7fefb76fc93984671f54c0d68bb2b60d7dabb02a356d2f58914401bfb8eace23c53cf3b3c40d43a29b23df65b1496cacc28ad2652e1adce9dd90bd0312ebb7278465bda4f9e1fc4759861dcea669b7d8c73a087afa89b6e6a4004be49e8f71a1cec21f0a5118e0146a0f6de189cf6f6eff770906ac8e917e46a21b0e8f0978dda0dbecc7dd2f05ecd368c82498e91c11cf5d4256262256ded8974fabf81b6c791e27eb9772adad71288ddeb9c7fa5a0de40a5e482915c101e96038f492d588a25456addb8f2dce0b50a24e4095cfdd89f0dd4dbd9b4b3106746821ef8c41b9d5defba026e27671d0b758ee567a35a57e68f668b11f610c78d601a85b513021b8a45c524a7a609d51ef97d643369cb831f1951ac58b3b7223de27ea1e1bf007f44eff780c6efecd18cd64b4f1e0922633fbb339254ec1c90e25930d2e575dd86492cc1fb0fd1669465eb412fc7587b5d506acc96e3f1a3a407fa25b44b354b37e8e15f27a876f19aabd7a2fbab6a89c67f00591c5d0beab22cf1d5f9168cb8a9d942d76fb13ba85eae998b4e7c35dfc86225b7b91c5e2272fe85c998fd7707e8b90ca1af1a7fdbaa619ac696b41404665558bb67fd312f8fc6ec6502c89e9a50fa8f945125e2fb98cd4f89af1e4eca007c04aba6a6701f484435189d17edab86406389e97d6d48bcd4f127d100ccba814d43dc8e57e0d0ac24ad587030e69a688c753cd6c941f19508ff7195fe4b6d076b4681c3cfaec843c74688261a44f05b1df2bad0e8ed38096ef", 0x1000}, {&(0x7f0000009000)="aff351691af1e87d27f55334ce6fdb6d70e649da1e19b8f77105ef8378b00eea382503ebed5474b507fbd5439a0a634d7fcd9235c5be1b92a7c4825a954079e6f76962302a3f4028e7b7c8aaeebe39e8a485cc62a4e4b855d214614645ae8210dbd4774c276693a516f0a4627ba69bc6d27b8480ece0ff7572c22981b6052a40f88e4946c69845147a5de35fa7659107fab2e2152775960798b426620cdba8d9fbf957a487b2fff2a0efae002ef00729fb5cc940e3ac1198996dedc8ed0c7522bb9bec88c1488ec5b92db363649fbe752b7944c63bc0d53b922c7e3350298f4efd69768a93259bf682562103b1c8ecac356d5924cc6144ac6ac8f8d5dfc3a9183182d9f58e2ce8495abfd9f4ff1174216468c82f65ba18caf398c60a8dbdc655a31664b1d187cbd8597d7ec585200b601a59dda748409eef16a44dc4db704e9ae937fe84cadf83a7cd1ae760817f5f95d7fe5eb12fa24baf26b5d0950ea3af1a7acbb8c815dd0907169c2286d953dcda6db3390a19471790bb7f3602168da2114e88508b2cb54f3f8fad02a924114add94ab0dd75cb813cd839308a0d9d71f47b0961ad31ab93a66c544d6f5806b6d10a58596ecd1fdae12ac26d99798e2d7cb86f7e3ec9c5e3dd67bc19cd624879b4d933a277c52db3936c274d3cc0fce0a00b672de9a51f140a94d8a28ecf9a0561937a2781f84d838f9253be2f6fc4bb9d098d675196e87cbabae5e4d3c27689528f3eae6df2e2dc4bf1de58d29077c2180ed0c7d9d6d2e9f97d8fe5a11e015c0d4bf499e2c049ef35ff873ce0dbd9f80f017ebd09210657d98dc9be1c22d5ed2f3eaefe9deaf78ee127a4eb0d7e517a20f89286878a2182072a479ea9f05ea112000c3a6d6e1ced57a0c061f9a36e83d15ff719f265f8efd1f8c90c3d10d7adb21dee1bd15494848f703120897ea9df84147f7dc38179de1fa844c3ef75e13bfc4cc7cf9bfb7d30f0d7a268b1ca4037b48e51cd1be5265a2c5035089543d34a9aa460b77cfcc9cd3b098a44253dc4fc793dcdc80ed1598efb393a5d37b32a0d504b39888060fe7f9d85d74ed34c8a27a51128f4700d283cad9e8d33c11e60475b42d621758c5802065199b263a489d8b50c3e072a727bdf27decd148f2c8c6f24dfa31d2b85d426931c35731aa3a7063be3e36d43152a09476b61a7680d5a94bb55d700f9df35411e761d471c085d6760375c4e5770b98c149d25443b0e8593906775157cd33bae3140b6b84775ccb40b4366575775dada72ff88acca5ba4149586fdd19e767fd7d14c4d9559f425b15c559dd6588f5cdf01511eb738a164933cf3fa6a80cdb624c08b055e24d9a0e2d399084450e9dc0da26a99cad657081dde9fb1e27eb056af65e453a539fe2df46f7329696a4841bdb635e9b4e6664262d5c3feab4dda6286573479c46ad5252e8869bce6ba3c04a4ebf1d346a6d9f35b21fd7a87b19fd819677ffc26ad5f4bdcffbab104eb3b703579b27b5724bc1e4ea1e8aea5dc58261f3b5f2c09d58c0d3c6a8c0d9cb24aa9eebb6bceb442b4b90899b61b756580a8b9abab2981073cc8f7ae48cd28a7e25bf00bf46af1198291aff69ed5ba313a5cf58df2e5f7d4a2c0ada2613f4efc4fcd251bb12657c658c1ca0fffed4927a16510da0a2c6988a443b3260433f6f294de59659b0a59b83fd144f00b442bb1500e4df6a63c2c358a1b858bcf7a71e4a24fbd5f1819af4e737ab2e2dfd365b27dd723be72dc936e34eae96dc5a9b458679038982ee080b611abd0124d9efacdfeccc13df76673238045d6874531b22cb7b8723eabf8593a5c01117b403948e2277cf57c8e6a8510e711b7d983ca967ce7ea04699c88817a70a9d68160623df1b4f8bb4f04069084f9a8978c6fb0bd7725d7214875bc63bd0f996d87a4049e0db0ca377482baa1f0ccdc5e29b2aa9b728dcd464828deea9f6774143a746a32b4ffe4525195b6de95a35b56ac2496003cbd0984d44b0775246111e58d457e6ad4d07650414134410b7d5f0dbde35d306e41aabc7c3ba19476e3490f08fd153dc637d5ecdf82e2166215a7417c4c499bac7adc22db2bfa672acb1dabfd00e4f7551050710e05d061250b37760649372aeb9d485a13372fa6ff125970606b97383ad84f7174e5c44a84dbd30825b023e77a1addce224cbe9f14981382ca81afc27fe2ec78fd43fa1d22db3ed4ba73948a32be351f50165930dc1904e0479b763946d3874fe8bd31b28f98750959aad9dc764372bee138b8ed06d76c7e5b7b1b0e97d9bc2e55b23a1738d22e11c7448abeca0255a63947d9517bf4180e5ba44a6c2b4aa724c6445f8c39db4473defb89556e25356e14cceceb9ab3e20d4f5aca625852cc44d3f054ecf131ad7c619ad230a57ebf21cc1ff576f5b485e8c3d958b2c106fb4657a5ccb8c13fdaf747d625bd101237d1e14a740ba903b59de052067ad609606a15b113cf19fb68b00d2c4fe14712456c32246d96b9e6d54765e3925edc40ae22544fa0a84b42c57f84a45299dd0921b50c7f3f8c8a6610c30379ebe5378858d4746e78e419202daf5245039474cfa03b9116e338909d9e21b49dd956f819e1e3922ca9b775b6f272b5f14fce90e70c2f7ed94df7c43b789746e8a597db4314a0a657abec7d5cade3c39d3d47b6405bfaca5e0b647339f39961f09d6ce646f7d22493dcea389dbbf182163ab744055f00b14ed5ca0cd3b1c5d441397b0a66d084622b6d71d5434b6e7cf01831982f7404ba905401ec9a9f25207628dfe16fb623b476b1f643c212afab6e64db408e5d437f36806c104554156bb208eabd04748967a81919af045a61b286523327ef8e34a52d2648302be0f6058b8724b1bd7296bd7665690418f4933ac31400a076fd2a6e05d4179b9441de7c99d218c1a2f5f137fe95a19221a79d1f4d15985013ab23f06ed5a3af36dc2aa0374c68b4c540ae3161880843c813947aebc8b4e43064051ab1c398d9df3762de5d397aa6d05881053313c2c397339e477ac01fb2a91136fd54e6b08490df7865735fbd2a92c9493eb46b3e3fb20ebe1d067ac01b1b6657e87dd4a8f34d7df39d085f818adb4537cdf2f1656281ed1b70ae1dafbe7d15c1bd2b18645931b1a6e987fea01b545dd9c6e4b78ce41abc6669ab34f2d24705ef7b8dc83dfa6bd06e43e5da8025e7a31387deb2ca4764b92310e507770ff26289565483b3dd910199bc8cf8aa95f4246234fa1a2779dda84280e7cc3c9370df4ea6b2ed6b1d89f1b86c9375e507abcd54c02abea547768b99a4a256e23af984d73bbbfa3eb547bf97ef15c7eb91b07dc0e5143def9f17f6e946f028838afa3182d29f0831ff191f79780d4a95059262b844bb6a1edb9f5ebc0720e70683c88554f044bf9e38e45a3708235007a81b8d85956648ffe0a23021bd7907c47ccade5a6011fd0bcbddd3b34be66d72ea0d0ef88492355a6ff5b023f4dcd6e06886fab4010beac590db1933f41645f8f2c32dcc29590785c0f725854416cfb4e0674f1827d2592ac6e7b42e5707181b39da193a83486800cba66d62848e9d1e9574f864525d98cfadf93a8d744a6011dde076d4b4b3d668f7290f7e99c4ce90da0a08eef2990c06c21902c8c8b9c7bdac59a483b568a9313d3039ad26d16eb66364ef3c8b5086deea9b8bf74bbd8d13f6af82ec8f54d5a9330fa9721e4148efdb2ff522fe1b5c7e1b701dfc64de14da40fd90eef947f80c4ae3ac4a4c9c0deb0bf2ccc0ce18c2f5a32d7ddb55d949b8b979c9cd5395f651cfd07179841dbbb01b534f7df46b09f836bcdce068f9d1cc99cd790fcbe82a30d3ef624eaf04ab41a34e0219ba582b0f29a7588e712e577dc0b1080ff724debe6cd47950aec435f892040a1b2100199c16a06f9b08a3d3fa57a9b07ddc930098dcc595c124d1c73336a02201b867f16b4e39a9fa6ad40017f9eaacb66753b30b418b04ab2b91fa2e8ef3b223ffcd282f9ab8cfe9068874d9e527d3cfd3aa15f0117e7f7c73fdbca9f4b23910118d96a39d0d9e8d82f15184ae1688771b4e2ade4393ceacec43249d066938e9311318a34a44a965ff16ebdee3ff822a1d0832d64bc3c29ed4311ab85b203c4791e6a315964c9c18ddb2711596d9579ceb7cce1a6c2eb7127dbccb74a23e5105895dc4aba98f78608f4c095d7089142482225aa9002da623e0918d5cc6cb404baec41a1959e368ea5be999a0e4060646d5146706ae88635fbf6834db4656160854b6c30792c71b81840ecbde042f99af816a0ffa4e56c22d04107840b52604109fd99edc66b088d1b78244b1303598a693edb261510e6d9ed731b468aee49ebdc2422953d65abf08b1618aecf1bccfdb8d0752f916c03e140fd64d554134400ba22cc4ecd7d59cd6739c18f6741deccc68899e6249fa60131c201ff6f9967aead4a41dc8c77e1132e74c55028d8d3dedad077c0fcc49c3333abda7c5e8ecde84b3290f04ef5e9c8b58abeeb701ffb3f1a04b17d045dfe9b2ea3c78e48a8615e2ed725ba8b5f81602d00ac0b214352ad06a5ef707ab52f196a1d83bd47d75ac123fee9076191e1a1b01f57bce4a57aa3983a6d064fd062da57659d4a5a3a0d6c46fa81a1a341c53b938f910fe55101369ae4b0882e804309faf4fbee721439d8e1a9b5236b71b473b16d5dd020e3e35dff8ba1a62cbc90f37e652d3f7ff325f42fecefcec08a99ceecd2eaf8529b6f495d2910086285fbb6ddce76a2cf0fbc7e0eb96936999865b36b7d3c0950ef1b21c9eb2b5a1a27d96c1a84acf4f96a61fa52469e4ecabe3e4dae71c4809d1a65ce24e97bf5f43dc194cc4d17ae4063fa13daac53716b89c32c9cbaf979cfdccbca22ed91920735ecb10890545d3694b1fc57f46eeb4ede46d7ca44bc126fb02fd212108409f3c8494f2560497d72b5d1e6be37302415c54b8f0ad9db71fd3ff9fc91a72b4ec49ebbcfe4c1cbdc6ee869f4c1026cc76b1991f0b57f77f7b7efb4db2412dc28648fb2ed165f47dc32ca615acb762e6a2025074bfdba89d3f7045782fb0653462545d3f27dc942293dd5e79d49412d3645624a35b218ff3c80df589246765d59b646c094c56f209f5db3fcf494c95a8934d33d13f6ca8c57e4c8aab594aeb80d1c06ba35f57302485cb130241f63d9f2c623b4f4e53fcfbff852f5cb392e6cb45eb0982cc0cab84e6fe70541f3467eb85aa02e4dabe5fce6ca0c61b3e44fdce4dd6f6a042a01904d8c13dd85edef3b5f0f610253eac920bb9f15607784f1da928bf5df7799cc432babd75fdd1d53ed1180d5dc78c4ebb171554693abb736f72de8b1f87f3be5553affca27452f22977a43ad677159a2ad3bfb6a2784f87380712469b4c0947d0ebf9f184de3b249cc6787e6efe0edc36b691f857868ffbbe7a91f503f71688078d0aa52c57b81e50eb5d707e68312d48627cfb8787fd8b7477c06a83ca29b4f625a778d4b9e8dc717009a7c825324567e05f05fe996d84452880a77dd3707f2b8fe6c1da18b5e4ad1237fd07a6ca35243232b3f0e30b0c00905ae5616335a137f4eb31d4725b728fc2d3ef77f84ffabc20aeade30626b034ee13e55560b1b2187cbf6878a6f6c4d86757d4223d682670ae03da735a671184a5686e9cd1dd796bd19250010b3bd233fa451ca428ee95d4ce04511a583c69849e24e426aa9c5db99f1d550596e580aa763b6acde9539345e2516ca954a33c1b59fb0305f6f01aeffe4cfad667f78675ecf4f6b417f04508f7d70cfea60f33753c030d136144037", 0x1000}], 0x3}}], 0x8, 0x11) getuid() 00:28:45 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x17000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:46 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x200000000000000) 00:28:46 executing program 3: fsopen(&(0x7f0000000000)='ceph\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:46 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x18000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:46 executing program 4: fsopen(&(0x7f0000000000)='ceph\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:46 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x300000000000000) 00:28:46 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x19000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:46 executing program 1: r0 = fsopen(&(0x7f0000000040)='virtiofs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:46 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) (async) r1 = socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = getpid() getpid() (async) r4 = getpid() sendto$unix(r1, &(0x7f0000000000)="ea560b454b9b9006ae3c5a5319201cca294cbe0d6d0d6d4d504442d5f170985bbe41bd691ae5b37bdef99ae2553bc8ad9c93b51d26241ee676a4ce69dc4ae66f88375c2b7fc19ee745837058ad9244f20ae8012ac623e7120e8e4542b6753a22", 0x60, 0x20000090, &(0x7f0000000080)=@abs={0x1, 0x0, 0x4e22}, 0x6e) r5 = getuid() bind(r1, &(0x7f0000000100)=@vsock={0x28, 0x0, 0x2711, @my=0x0}, 0x80) bind(r1, &(0x7f0000000180)=@pppol2tpv3={0x18, 0x1, {0x0, r1, {0x2, 0x4e22, @multicast1}, 0x3, 0x1, 0x4, 0x2}}, 0x80) getresuid(&(0x7f0000000440)=0x0, &(0x7f0000000480), &(0x7f00000004c0)) getresuid(&(0x7f0000000500), &(0x7f0000000540), &(0x7f0000000580)) (async) getresuid(&(0x7f0000000500)=0x0, &(0x7f0000000540), &(0x7f0000000580)) ioctl$TIOCGSID(r1, 0x5429, &(0x7f0000002c40)=0x0) getresuid(&(0x7f0000002c80), &(0x7f0000002cc0), &(0x7f0000002d00)=0x0) r10 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000002d40)='./binderfs2/binder-control\x00', 0x802, 0x0) r11 = socket$nl_generic(0x10, 0x3, 0x10) ioctl$ifreq_SIOCGIFINDEX_vcan(r11, 0x8921, &(0x7f00000001c0)={'vxcan0\x00'}) ioctl$UDMABUF_CREATE(0xffffffffffffffff, 0x40187542, &(0x7f0000002d80)={0xffffffffffffffff, 0x1, 0x4000}) (async) r12 = ioctl$UDMABUF_CREATE(0xffffffffffffffff, 0x40187542, &(0x7f0000002d80)={0xffffffffffffffff, 0x1, 0x4000}) socketpair$unix(0x1, 0x1, 0x0, &(0x7f0000002dc0)={0xffffffffffffffff, 0xffffffffffffffff}) r14 = memfd_create(&(0x7f00000044c0)='/dev/virtual_nci\x00', 0x0) getpid() (async) r15 = getpid() socketpair$unix(0x1, 0x2, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) accept$unix(r16, 0x0, 0x0) (async) accept$unix(r16, 0x0, 0x0) r17 = openat$loop_ctrl(0xffffffffffffff9c, &(0x7f00000001c0), 0x0, 0x0) ioctl$LOOP_CTL_REMOVE(r17, 0x4c81, 0x0) socketpair$unix(0x1, 0x5, 0x0, &(0x7f0000000000)={0xffffffffffffffff}) sendto$unix(r18, &(0x7f00000000c0)=';', 0x1, 0x0, &(0x7f00000001c0)=@file={0x0, './file0\x00'}, 0x6e) (async) sendto$unix(r18, &(0x7f00000000c0)=';', 0x1, 0x0, &(0x7f00000001c0)=@file={0x0, './file0\x00'}, 0x6e) fsopen(&(0x7f0000007e40)='cgroup\x00', 0x1) (async) r19 = fsopen(&(0x7f0000007e40)='cgroup\x00', 0x1) sendmmsg$unix(r1, &(0x7f000000a040)=[{{&(0x7f0000000200)=@file={0xc563a0953a94baf3, './file0\x00'}, 0x6e, &(0x7f0000000400)=[{&(0x7f00000008c0)="26df6347310b2ce65912c6171ec0949f16097b0ea5a1bb75508e545f2350baff8a9e88c623f80fb50d9d23a97eeeca24ba8c256f3414f3de03b745de410f5e856172998b1318bfc12b0eed3806fe4abf662dd7575092e018629c901d2209927ebf953de877a7191044421708704820a0d28a40d00a94d1a9861992e84748d09140ddc41c1ed5305e5302421ba6461020a85d847d219f17e4fb1fb6e51986b475f53278227dce83eb443b1e48e80507ddd9919296d6c26843b417488df70ec501432155c2d1fdf91c0fdbcc69d8235b1446c4fa27995e6e68e96f3fe27efb4aaad3c7359dcbd04a78c210b10c9099c551ee715dfb5afc842eef0bbb509ce39b26120da98996de3cd3d17d19c1221a82684b2dba475eb1a89d1ff7122e6fb52f13e09fb5b7c03f10ee16d56ebad4571041fbb22c6a9afe0a1e8e35438b9bf3ab2c1b28de96b34492de0acd00ec696c969d9207efcb3896c5a8563d1ba2246c217d9f0c492634ecd78cc5bde70e82c3e26f6d208094f3bdf3b5a14af05362c46141bc9cb38d8f1701de16d36ad62a1d39412a9c9c06da88f01512cfe13cb5899c4ce955f918323ec53c03e2965ab48263ff5b65bbaf7ea84d44892c80a1b940b39922cb9c7a9800888c2ac625d6d23ac43b3c655e46bfcdb2fd033cd2ed803d9f4a0d264b5cc6fa02ad652232b5744566f5d5fb7fc7b9f0570f1457a781bbe77e3b70c0e2649e43df265cfbd5a6c38927f627481a343cbf4ff168adae746636658932ce2f0388d996ac2a9bbe0bf6bc7581fa624237f8ff556083e7200c26ab8cf24fa6c1e363618614f716d9a2e1af85982cc414e4ce48d5a7bac6dcce4441febc7cf54f20d6657d1fba1d543a63b90c38871b1beec8ef62281a191c0b5e3cc36c0df6e7a38d03730053c4f46f422de91ccde20ce8a282f605965bb1ace689a29ab4c327af1f99491b24ff08e73bff49027ac1ad4aab135727d893ffb86346002295994d335798b99187a034f1ead9e18524639ca60f69a90ea6bdae6d1fa3cc051900817d1cef2c06ff45577933915d47a892a36cfbdedffdcd1b084b3c29ba5b284d1efc953cdb9c4a329f9087c2c557cd0256b8260786933ea588cdb5898a6a187d0bec3135db5fcc35178697c14bab4d2cf6c861f6427f1c697d3d074167494830ff0e074277177135f4b1a5ebfd7a63ada11cdfe7907620b26676867dd597a6d54eabc30a1ba724d488ffe8c7a949ea373dadd27e039c17cd4cc3ef92ccecd73dd3db2dd9de79fe89d4f1626e7dbe1f3a6081a197220235974f8f63ee590445ac78bf2c346001d397481b197c2634a2a38e5f46e441a0583fa12d545c7de80fc3e8e137839f1ce5c0a2817b46b66e43152e41f04a4e5adc78f497b54c1b9e21d8f02bbd1148c0e21194d88a59e4f6a2b5065c050c053ed6e0810c03cfdd469ca1abd6abc11ae2df6475bcce9f2916c0e4fd9da5e28165c8cbd459d03a5be723697deeb9af93b15a637c5b0779c4c7cf76c3cab9e333df9b9f2297c4370c59d29889bcc363a4a7a24e1ff1836d373c45c7f7b9fe5f528b6e18b7af9924276d3b4612405fbf0eed72f5592affad74fce0c1098a6b12fb8455d348ee364e50dc18193ac587682b2d37d5d8427f6513c74e1974cbda833962611f95fb9cef61cd65e11ec3e328ac19180f3a313ee88685c95990470702db1bc7b0668d5c1839a22af1c365c5c303fcbdad7904e6609505042ab2b0d199abb93601cb00be3eb3c4652bb8bb44634ba11ed0e736c49f68e9802ec093aeb417df1c93bd7a622cf499b9c96525f54f27293edfc8bc0c4cd258084aec56ea02e8f276de1dd903cfc2beaf0e1754365e740805fdbcbe33b2bf804bbf0f8c5862be396e08d09bfed47fddc751f3dfbae6ea3224f3185629217eb5ec7505859aa26072c3aa9a5b1b3dc95f8964c48cf6dafc9cbe0ea83e715e7bf56f92ff3847e6524c92fac30eb48a178c79ebad3bc04772c8b96c5ae0b082f99d090ee29cbcc9e84897b413df84e6af5cf74e277c0e519baa9755d03be6996ac57a07fefeab749fcccc7e85c0ae137f2c5b1705535223a900f04fb04f54066123d9679f96e4c85f1bc909706997b34c6c42d705e089af12a9430bf32e6f3cfdd3fec6ad5a53666580386e604a241ea4ad95f9ededc70fb5ea7664aed7f78861963c4829ab0a36f08068df0c1c198a35e21bf02b23d9891ecdb3185f9820b93d05d3de4f089f85b95080fa94720a4e50e1593bebcd9ea123a04b02e15bf4e3ca2193ceadf9ecbf25c013a6ff1c2607681b56974cb046b58540d6396669fa4454894486016a6f783fc46793687979cd469e5eed81b8e8da97acb00f55cb887851e7023db6671ff3260d67dc041e17bddccb3692cba50061718b97b7a6964dbb23f523ef0826dc60dd09c34f05695f0e07a50c8c475ebdb96111c226a9856d231060ed65a270664da7fed3393ee69da5e27a3c142c26ff10f696f0a24cf0cfaa7b9b86b7a3e063c31273c8d8068f1ed10963b41c899082ff807e9072a09e33aba36d072540f4efa32b7c9bddf20aef302b8b8800aeee6f0e33ad2de2e4f37cae8f390ccda0fcbd854354cf7e191dd7d37da3318c99936cea27d820683058117337d10f19ce9a81b5803f33087d4387e87c84ebf8f4a24814e963a0db630c68f95f716543ee36b22fa8fc1ef3126938bc0203975bdb38c9e38045c124519b2c394e915e6e7ada1ba71406b580aa549f08561938f61c704d09264a0080322d409672c8413b17664861d334353112c1d4161bac59e87bff00d8fd9c186c8e76c8db0377380bc58f441006f40c85345e0f85af8a4b9ba062b7d816616b2d19ef699643eb891eb1e9f0e7588e4b0481d0d77e0773e9ae33b132fe6eaac99914ed298897846ab21b350422cb4cfd16334fb2b3d8ff850d79921899a31bcc4b2c4710515405f2279692b6ce29fdb79e4c7ba7c91ad21112ae2aaccf2f865fc2ad8959d5a77c69dee6ecfad26a604680b35577b621cdcbfd0796a937a0f7a4e2e075cbe87fb8cc14a4ffd938945243acf77c969edb7a619166c8c0d0fd5b147d8c322376c2d6e2e7637abd8233013f56965ce97d1ba7cd4a3c187b3f475f0896db9f376a2ae9e7e6ef093072ca39f1009d5934a9b1c9608eeb5e1bcea1aa6d9c9a0d05289f5f335b52ea773f9d652f81e607440e7a6264b9c4ae69dbbdb6036251bdd7606db2c70202bf55dde67e3e6f4c30ac4562797e55c4b14a6173ef45f19ace8cfa742b62f2d40f120f0831f9ce0274fb312ea311b210d8ef66338e72a69ba172c313c665277ee0e0bf9877540b32c9ecf3cb91191f4e95a2ac651a9f048f1b262264a189b4f211edbe43f75f6552cb5a63395b99f36e69eb7b50fbd3e71fe6011b2328f20bf2f63ad5b43bd496eccfa6f67407b97a02130d0de6c234279f742b26d7e848018e234d0425ccd92e9b6406a119f3627b7bd064e45a546a6fafa8cf214e2bc09ce4becefc7e7b6886da6387b4ee15e2dc89381d19d66a15540c0552931a170c4fc59e8dc4cabf9e8006d3c9082c37ba0e931ccef7855f89801c124192a0d8d1de93e0557ee17237aeb6f0c52d7aa6a8e4724fc97b28e7ff719c8cc61e0ea4b596bb0d121d96a82157b26ee4a7a7cd13ddf1d5e810ab219e8745ca50ffa866afd74155e9f55b4014a704ab27214d5f61068935b8f72103a56efa9b602fdf55afca083e7dfb3ab01e21d4669a2d954e4fbc66250f5b680d9074527c53f1e7a30eb428738cb7c84d110636d299da1aa8366c9e0a90def0808600ddfd20fafb662f4f0db9e2360a9e6080a3473d46f8e9fbf62c3bf37c8b848f87b7f53aae1d82d1486776a2cf44a1d41dc5265ef1ee1928c7338ef56ff3235d48fba64c076dba2986db6371b16b19160cd1a3a32e0e7ba97317856e8a1fde3d2ab32a498cde0e1f1e88519e01f9c4b51fdd068f5c396e220df70f1f9ddeea09134dee5aef17618c98eda3c293b519e89ccad03578e76f32a1f4d7dbb3c80345bc64b8978406da2866c868158575daa2858f89d9ca620d207385ab31a6ac5fa26564db0c94d4146164c6df139f756072f19fa7e1c3a95daa747c52a0f0eb80a2b9f060610e58f84023cc7834188d62db93e84ead8195f9561f8d17dabf463bf4ada7a895ada07fd0b44c7fc92856458cb9cacaf1eb8b15857de5b0654851aaa2f09cc969a2e7819ded0819f93774eaeae2521258730ab77540b253f1f2ca1788f404c641821ae1251abe6e63a9aac8ba4a18a29dc00e26352c9ae8705874a041b681fcc97b16567d3796648a9457f90dd98fb9cb627f40899aeeed36f955633464c8ad28de88fa0a07ccc50b9e194771e476d5a34905fa3e16d00abd99feca6ef5a21d644f90f4bef73f819ca57366834c8f2ec0c3bcc0e9bc8297609ae5fe5bbe97d11c7aa233ecb454db2a1319e30e58562a09863a7df8d4b48f5b3c8674d10e7d65c5f0675ef366818e7036990cd85527d50708dc6cd632f9b488e0392cea205e384e56e844d09774760d36005ae2f959cecae20c028659ea0576eb2bfbb486552bf9d0413be08ebc63aba0f96fcb765636f7fff0363e031217756244055e3d5ea5044016186b9eb43e0bb93fdc1ebec69da15b2b504ec3244e4c542a31c89748389c99273e7643e57c1f7bf6d421e4fdbe42c164114da9728ae11518e780c4cec968e9e55189f11090bc7a9b2a81f98b269e16f8b1c13e95c234ffc276906d8ac3dc10032f81197942bc29b215d73f98454b72cb3967d57dae0bbb84b1903676d8ba9ed1dcdf720d21985a5e9a1e38c5272d55e8c541e4c3747808fd23f75c47b11a21e8c715e677762310566f119bdfb3dead609900e73d5c423839dfc93c0dbe5a0de304d29f9b0e3afebbdbb179aa3702b06409b85643a51cb995a2bf47d566629ef76d674c39fda4eb4b9dec532d36eea32eebb05ce05d5497219b29bbd7b8454ff7315a0a7c473002c51f99478c0c27993a6d38f8b68efe51f5338ac6675c4b42dbfb049a42ee76a29ab7ea93613a2428aeadbc5cf34bc5702598c5bdad67602a4302ff55c36cc2560003c9d2d6273c72a15b913c4501838792231f85c0cca8b8f81ecb80acf96bdfc77cc1a80f2783eb37395b1b02e99a30c9f3cb583c61c162c94871fe3a084263b5918a1ff0423b39df5b76ce5f431aeebd92ffe6e5ed06b64ff3cd0ea9cf12914f6f076e31eafa1f1b26dde6f0340f21dac464b3b943036fa11f913f69da02b69fbd2b586bce8217ceccc05536fca81ffc179a494ac6b1a5e450bd0d316db8359da4fa1672deec9d1bf0411bdb943a2b601cac6ab8f3275501225131eb5193eea3f9f05090dfc021bccd6453be0428692eaf40f780da6210de0a05ac6ec3c5feabd72ce394313e57456155eea0b9daa19d4e3e3fbe5b0b65648f3b78c09fecbee0f00f1bf4a56751618106126d39cf065b5f412878da68538a0cf9457a50bbc363b1e2e5236128d30846e2f0cef94d961190130ae0acc1ab29ae7c419821659005ebd8c9167f23267316f2e4881c1d5b4a355659e9582c208b684c865e1cbeb5193e8c8a0d8404fe67113163d0decfd408e0dc91310efafe4f9c3f5628ac5ec236b104fcf904b5cea909f40fed759586672ff4a131256d53f3ae3ece76ce5606ba1b505f70781c9a6b26348181a976b658940041c78cffd56431d6b992fea4909a71b88b1d97a188ef010a8f8e2e8acdcd6d36fe0feaeda4268e6cff61f802e345e029ea2e838bc71996", 0x1000}, {&(0x7f0000000280)="8b7c5aa06b1f0149f7d035c4acae8eec44a4253f8aa16c76adfe395114ec1cd615360d5909ace6d43af55e83bc8b825a9d3b9426f4a6cd34a074a2c0c205afd479a1b0f6ee7749416ea304ce0de125df459b", 0x52}, {&(0x7f0000000300)="23754504d47058e50a708c5d928a28caa8863177a96628736cb5514a721e5e14082c8125cacd7b1ca49288e2a351d938c44d57b53cfc1c3392de5c3cbe9650ca88ecb9b5fa3237aab74140c2dc28b60209415eba667e171b072f172fd84a18f24fe5e36c5f4caf697c9e1190bd1499b902364c762e6c8d3835de51933bc6266002322bc1ac1bfb4117c6eea2207e55ff2572ce210743a78141736990e5298bf8c5a1cc9201bc70e6e3a3fb0a18fffbdf2974b1e0297017d7a391b50bc11aff9b8191e03694eb2506c7892227700542ebf45a5f9e02005ca245d86d6903da06495a3dc6955bb12ce70f4a8782076a4467eb27a897b052adc56f747399", 0xfc}], 0x3, &(0x7f00000005c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r5}}}, @cred={{0x1c, 0x1, 0x2, {r4, r6, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r7, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}], 0x100, 0x4000}}, {{&(0x7f00000006c0)=@abs={0x1, 0x0, 0x4e23}, 0x6e, &(0x7f0000000800)=[{&(0x7f00000018c0)="84e7419de7807afe986786d254a3807962232a70ee8f75d457990481b7564587feecf8d7d55e2fa705364d05d7c6adf0b1449042a26bd9c0ae420277e5597cfe15c63ccdb11efb11d6ae76cf854029798d1237d06221f55d501aab6e566157835e22f8a6cc877614db7eefa8e2095cfbfc6f3030145738958da325bef7aab490c1d5f022a233d991b2dfaf699b42a9ea2ebcef84abbb2cbfb69ae8f2cd58b7daa131cfdd1cbe3bedc4efe548ecb05db6d67619d30b98045f07aa23e9fb137ff01e5176fa50b8dfcbc1f5ae3c525aba864a8a876214db4a05b3a44be0c4a3841c24483b6dffaa0a36e54e1035b23040cc7dcd0dccf4d20e87d6fe706480ca5da9df715cfe8e37cedd19b09ba40406052a402aa0ca9542ffa99df2b61fc1e7b309e3d80a94f64efd69f35ad37a0fc665745f1d9d59e404b28a7ee300060de4f6678dfc1c0aea60e716838703107e3c24a9642fa6ace21ad8c4980793377040aaf3c81267c1bc86eefba67b226a08afb5592e89ccdcc785dc2e4a526d9ee7d610fea7ff5119908b87b41f9005779665218fdedd34cc7079e746a61cb7add96b9098f31255ff8711208ed51b733a5b1d5a29f422368ce6da55c38ec755a09772f2978db8c04228fba8d88b9a4118aca83707c0d792fa0155c52b015b0e23342d1b9c2456b1dcadf5a423187008aa43ab7d83bb9d41f3ee747f5ca9efa1dddbc4302ba8e9d82df09848cd78d9d2c2b7a2851177c2d458dd038f5dad0b6c63023afac0428f9638bfb1257febfae9a1a03ab122480c671748897131bdd0d3ef834ecb040078c967c67a14ee5f61a4f64f2922d120f773ae3fd8dd32acde92db5eda6ea891a6cd8d2c5c19702c1fdae8daf1ba0c16f3835651325a4594aafcdac04e36b4452761a13af62de2659757456594c7c2433741f3846a4c35f62207ebd1126436551baae5ea6c8bbdf51cec3ced03e51eaa2be3b7207c6e71a8808aca9cb3a1b84cedc903135dee31427e6f38a9f36dee06bde9d49a6091357bb15839c36039f8ddb0fa342edae7770ab0c5aa66de174a5813db5d97d9b55bbe0063d603fb851fb48b7b5db12961444b8f378d8e387a06a1c1bde3c6f5f850779cc2ced53793a462edbc2511bdce50535b47bfbebdc0444bcaf1aa130ab7d92437dbad442456d56a837966c253bf5780cd068620c31b49d10e45bd39d402eaedcfc524c0add302437df601e60db5f7e3488296e26305aed3317063d02c05b763e33e9bcf8dfb9ee520887aa44e9c4ea068240ca025c296d1e38146a6ed81abc6659d64295cfc55be5f422b546b81d0bdfa328627f6bcc866fbc88c00c33024c589535c7d97f6ae3111ec9f3e0f7814eccaeb927a520cd04119c99189f08c3983af1c560fbaf215ee57d67a602115b1c58128ee9ee4f017d6117d6bae15218d94c25d9e15f6760b5cae13626eb8688a76313bbef45d247a41eb17623707edf00ba12fce0080d007bf6204e5df6ecd83eab8b56e760d3e537bf5e8bd7bd6e8d0ab617f7cecda59e49bfca69e9fb82dcf492de21e59c2a81272ab7667db42af684621abe24a50eec332a93c365682a17ac3bf917dc746be4ac418d5d404756b62024fab85d0dc24483e69692f14f5abf1b24c85ded680af43aa4c38d0031680591f7458a75a6b2ac30f9a4a47c0badbe00c6a5ced8d4c2d1c4c320e8104319e2dae4fd59161959f4d6e3489f5e91a335d44a0712e4e37a40da1b4673ed199550dbee9a4ace0b00652f8a3bb0488b6d2f2d0dbf581ebd77c354d304b5be4ad560cd3cd6cfa01021b453381f668d73ad317a980fe4ba2fb1520363f6534dfdd29100a01fcd871961571670aef2ac026392f6e7bdad9c26e6fd7c5a8379f4671a3493443f179cfaf061c04752a1798624ef381700ce55a0424b2f8765a3e785513ecd0ad04e51adfe15c9dee0928913d87fbf87daf391bc5579085f08d022b4ed21917858d67b45aea2038803ee691d3c5bfbe4a0e6a4290216f1230137c3acbcfd0298dfaca032c551ac624b3fb24c26d002725fa796b75b4fcd15458841bc2a5c8b0d28f1c9a614e71994234f48fba29817531e5b7f56644727ede975921efd19741823b40bc49d68a2e2a2e0c8c6de7f676e3f545d2a347d73579e961f1096198576e3f1edde72c3b948f5c9fc7657d1406065394f0bdca8241fb0ef72b085e98a0fddd7ce299a9463223dfa34357f116dbf693762a7e66578a319b781b4c5fbcdf8ad1e994b4835d2f9c2298c05c2269cb5d31905dc0c7dc2d82e0d92db9f0e3b5de0f21dcd11080600bd8c8f6e0df8a1e0b6bf8423e592348b55041307e2d939d284bafbaebcd16f38ad56d71f8d2bdd2ccf40475f3ad6615213eb9bd7aa471213dc7fb449c75bc6cb9cfb334689a18fd0bac548cab3ec8c63d8c8e3dd42bb421d20845ca5a39e7a60d8bae06bd3d6888424f566358ab128a6e17396afa1c2bd121b6b6941aaf83dbe80eac4fcde184e4c66112c16c8d80d666fb32230cc7066ace8a4fb0fc67567280c0a91cccea6e03b45ccd6b0b04796ea31160d165932b63dc97b4298f0a39b6caf0f7115f65a238112c2bb387a64bfb9a61b03c9965b0ff80a2742730d05ffa4535232eed3e3098d40cbdc110ead18d7bf707451768f636375733927905fa5e24eab71300d09a63b42ab21cf2f146c0c540b7a520f0ea8b1ae4dd8f57b0c12d37df106769ae9c0e77010d1d2ac392bf2b291cde15420600c445c4b47b0fb2dc4038d71c0b6e25e22d232100694f6d33c6f8470848546bc601f0c6336e08dcb6cf661eccc5cf2432d00c230a21eabed73310b6fdd9d654ee16a904b04e79e493c855a80c2fbadb75e08cf946b58ac1b3ef6548771e9d0a11df91653e168753a58a611ff5910cee2c6447f5c565f838abca04dc6cc1959a4192d0f62eba7d9dbca297a01afae6095c49948d1b6e0289184e2311e6e3f6a7257452e9758911d47529a6ee1e238fdb93f4c574176e0805d8e3a7272d9086969fd7f3eaa5ed4c7f61e91abe6dc22b6ffc89cb48f15dfb47f56c1b15c07199709741300d4bc638246a9f69840f9736cce01f910e5bf6ae3af3c9affa5dc7268d26e06196c2a187f22fe2c3ea78017e0d643254a3847b583d34bc1965900cee2403f14e00f310eab0790b8fef28e8d2727631fa32058c9150910711969df9586d4c7d7fc60fd738de49ada181afcf96a921e31b1ef85eb48f05d2821323b0a1db183ddbd196bb72f77b82a29ab012379b15be331d107ddf4d9132199ff320452ec134a9114ffe4c9e4636790b99aa291ddc91e8344fdacd070cd3db029b3725b82b13ff28a7caf1dd7227fa72dfc867db6d86861355020ae83b705cc9fbe62af9aa1e4476707ec57a86998dc3f418ff894bfba45ff6a967ba22724aeb2401ed8f4c943ef5d8fc96b211e6632c752514d282c46ce494f11134becc50a51d7548354c8293b83f2b55a88029eddee89cf8be62339d23a7ac72156e26ea0902e1f7b445bc65e61c8a76b6611f742c6661a652d82da36bf2e457aa76b9c3340f19d25e8d5a0e8989b54ebf87711879004b9566e54e8b9df0d269f70a72ca8096f1d6966d66a7cf9a081090cf3ea99e1f91d683083c5edb3ae529a5bd73caabe2057dfd9d0584668b195a961df750e564c34477ffcd58b4333abc24d257ed76e470ff183386959982c9be61d0617ce83255129b98b70c46de328ca5ccd76f72917cd279bafeecc529dc55b0729d90515a08d5575ef21409db69b771448a75fa8489b2401c1bf2d3197346235dba1264e60b0a4ad8e0225963da86af0035f74a106ccd3947d4ade26ea5c6865e65ba6fc9ad51dc2a11f7aa2747312d8c27501b6e6f339796561e4210f32bd3ff88e557cb2d51f43d1104564318d16c227c35349effd23e93093bef312b09c14adf3ed3cc45e5b27d1e696a8038ef4188376c754027d9bc4746e44d3e78c598d9cd5c428215ff3ac8c10a2c4bf245454de8a43aab0eb5d975573c51ec2496a2b14d62f293b578f42f983789f31d6aeb9e2e8eeddfbd3091f57f94678b6b1906d173d99c467a1920fb4ea58ef73758c33352759f99d3302ad87388a301512908286f712084c818e737ed1cfd42092ae7e28b3c8371a036968b080464072808fea7bd950fa4aea78fd1055042ade67b52f739bcad9b0cc3355dd709c92455b7f8d6faa8d61278473bc0cde703e4788e01f7ba2eb555f3a7fbfa6e876829ecc1b7c619ea45336ef97e933ab514ec77ae4d4649740959b9bd9d897b06dd53dacbe35286a9aa5ab2d2550200edcfe1435fc78d9a855c48147a313511d7a895d5e16f6b7615551239e53bf0e5698bfab18e33bfc3d65e247ac944f62ea827b7c1a8a664c2150a9c51608e0719c22e3643cf0c5c7582710c6613054f3e22c13e42dcf15db6a78e39752415df811c843af56223789bf4708e1c108787a21e3f9608c72702ddf194306f573966374fc3882af7e77cf21a542502c5b6edbc54a8169da047de3371a146b6183563e4bc379ec44cda559534797fa4c6db16679ed20b6b474bb47322ee03ada85bed6a1de7fbb0fa84d0456d3953e4f10239ac9311e03a29009e880b89bd2a4a05ed0f7862d1e05e06423c136ecfe6af159a850ece88329b0c0758ae5b156303957dabd56ab7d989cb184ad0b91627ec1929b0e25b75a7c9bd2c480a4e0999f1a5422cbbfef0c6036d96bb9284c445d07b920e277a7f25858dbca51ba28c766e0755fb85ec40882ecb2399f7b30a6e90f4974208f45a400419eb9a484eb7f1189124b8a8a2f92e846528ca2be8b512326b5cc3ce5bfdff8e3cf1d5de76858d57d94992cf6a32dc4e1510a6189ca2e649cc90a3893be81ecda834dee5ae099df7f44e244548c580893e756c20505e77470db4f84b48a3b8ca1e4e2e317af499c5c12abbf69609632f7b9e2912bcc133d881ce760d1f3925b922f34266f1016ccc29ef56aad61b0a210b8cdb0b7d2af4c0d60ef273808bd9de06c4d84bb81c7675811beac7bbb503859245c4054967d899606d8674f9bf6fd15fc6c120e3315e62ace711b3c6d403a9c1785bfd5f42b095eb13ca94e5ebb27eca868101c5f2bdc464b6472314f22d73cbd731af745b4a13f7835cd8d8c60b6c8051cd445166ddc392f48412ea37b90792b470f7916331234ea2fb613d78db9defdb3e1f3e219b4e2250c64fc90c8b334e00c83403f04ed24d08f1b9862cdb82b2d725b1b4ee027451e553d1c07eb72b291db650f2141da3463270e6848aa172e6f41dfb59c5838c278ed17c853d06db43ba439ec09705204c6349fa8c9fd18a31598fbd432b186d3fb91e39954a300c35150e968bf4c722b9e544ffc073bb52ff3efd7c5234422a24bbcb2a1245d8b4349bf851cb5ce2528be8b9bdd227fd4e390567b209f70f3386ca8a6e70aabaf3a1a533fa74c823609c597a239301fd03484983e5bb898db683fb8c2d66c549b8c0069aefd0aba1d35ee9c415ab468e3a7b23858cc4d85fdf91a258c68fb150d785ca22875a7952f6ed43d7e08791017734dabebf002b54d6667a5ef951047b4834aa6ec2e384f217e0317840ceef9219eff31b6ac1795e5528488ddef90f5aa6a64cced8bb237ef70b94ee472c0083d117c4660507e32dafc7ffda13e0b20e079bb02005b16cf8eee8ea5d2c7a85cc42cd1899b9244eeede6adf33b32d69ede79bd8712c602975e5fde2425922b417d84014fbbca029ebbb70de2b8cc441fcf9c76c5f4edbc97c4fde1c0613cdca0ff603", 0x1000}, {&(0x7f0000000740)="466421a47d47df50c6365679ed631823743dd09002fdd7a9a79563dde7e30b25fce0944d50f303513e9d22098f2355010fb04a64c4756e0cb840a44f7c9851520c472122ee49b180cbafcb207f3f29cff39c71c04c18175286fe82af9d46eb06b9b9dcffddfef69eb48a3bc3ef8fd81e01b1c1", 0x73}, {&(0x7f00000007c0)="0f43e5038345cfb5ab33b5a14856cd27558c33fc625df021a38fc60b8d56d9852065a5047bdaf87390203f98e711ec90", 0x30}], 0x3, &(0x7f00000028c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee00}}}], 0x20, 0x804}}, {{&(0x7f0000002900)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000002c00)=[{&(0x7f0000002980)="df08eba0bfc089261b639ec7cb50b403d8d9e041f6f764edd1bdcd32eee1c4af1666b3e3eee568b9145e55a3deb0998c51d3a653b146028c0b4b856e08c6426cef98c188d27d215b4f8caa104920366df4d7a298f2f57a971ac3034e3394f2c4c4f9ef22488da0a83551da6f1f90a0805190f5d9ea75ea1dafef99ad0ba80683a7c8fe0526e7f133c703d0f88ad78688fd41f22455d1405fcedc19795858228a9fb18e830c737594a78e68443b775d75092ebc434b69a57cdf", 0xb9}, {&(0x7f0000002a40)="3f923d00d3c136dd8176ad686725ea1ff9314cb016f5dc00a41bfaeb9a918fd454c3702e6a815d3df04447bd24bc106df6b0bcb21d060484a8509bb4e219639b4db7ef0815cb1fa80e5eff0e1608c84984743a92c839af92ed8d9945499a5aa95bb998e1045babf7065dfc6cd1fe6d9431c9f5a561256d4f940eea19250c4d3e8e3bdc13111d16736c4a0026dbf92aef4d3b796fd5dd7f5c6044aa8185c572a7f38c7a58220cb0", 0xa7}, {&(0x7f0000002b00)="b21ac1057f742d5e21b705c1f3b78fa7db7c71b1a1b98c25a46d4b6187f83cb7e93182d09316f4ebf8cdbafe41eb693e0b65709893d2f562812e1e611ed90860f73c40632a6a6c5caa6028c2bb1ea762748444f2f22dbaa98416b5ae7d39dc9dbcb2c3007a63f0f11208b09ee2b61cab45a104d9e537371084d2af99a579469021b3a476906dd34b6124b9a5444ccdc962486ee5fdfb9d05f4a2b6a04bd43a538451562b6f31c3a244230578290fc9b1edc619a45fa1cea5d9df974313a7067daed8de30ac463a019218c60d71317fc25664a6d87abf5792e18061b1225028646d1166e0278cedc8621519ca3759", 0xee}], 0x3, &(0x7f0000002e00)=[@cred={{0x1c, 0x1, 0x2, {r8, r9, 0xee01}}}, @rights={{0x14, 0x1, 0x1, [r1]}}, @rights={{0x1c, 0x1, 0x1, [r2, r2, r10]}}, @rights={{0x34, 0x1, 0x1, [r1, r0, r11, r12, r2, r1, r0, r13, r0]}}], 0x90, 0x4008081}}, {{&(0x7f0000002ec0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000004440)=[{&(0x7f0000002f40)="c151b349f05d56f9f4c372dd0fed519c7ecc625e78e802d29227a974cfc9a58c7df6d2299d8a347b0eaa458ce9482b65943feee3e3d055726444119273bf6aa56a6c91ef7b466e6718b1a1e8e752c199b72cc2928651bbc1427ed15baa56cf9d8a82c0600887cb0c2026e6303be70d8381fffc2633b5452a1278006b7b21e3ff97775439985d16f87e0c57498ede83120a9785c4ff9c9c982ddf0f34305632b6d0cf7a60b24f6a65dee0655c6ae3d952b68439d5491fe2a6cc9c480c365fe31a6a9d5be34de0c1a5eb525af7b0a4a7b3fabb876cb0ce771831e6e9899c117f8027c95e2f3337aa5ea3a1b2fa5aed6736b7d7aa3811e0f310489013", 0xfb}, {&(0x7f0000003040)="c092ad957bd69724d3d352a6d42ee4303c6ec7f01aa6cab1503840b654382e922052c23717d1b59a2704db44c442d19a5cd507fb5df80dafb93d4c89a48b31b5c75ae790190b8caa5ce50713bb86f2db4778e42939c378091f5bd0d8606b0cdaefacb3bf7776bc625567029119a4b221768d2b7fd68bc3d00a9f19", 0x7b}, {&(0x7f00000030c0)="cc9163f5439c8454d4a7b6211cef87f1d67d99b5888988258fc5d2d2aafa73dcc6849fc4a36898952d7c1182c051f749f4a03ae4e1489f2b3b8af996f7975d99e9d15577228b8012ed02259e3c16ea2da7ba4aceacfc6cb97b321365009efe14d29d8d4a8399f5caa243cb5b1f334ec443e7f7da553064eb2443816eeaa8423ee05d5ce6900ae0019df448e1ae6f54c18b9360f73096d4aa7057b21857679d0728087854fad17bd16757a601d265111799f581ab7bb68b416d26d82e98d4cc96b4fb0ea3a6c338b35e7e4b958c06ad9f1261f8131842a56528be17b675c4e741abced708ca97732920ce03141e42b56f595e776e79151def5fbe34e78fac06e86e58aef484fea97316f74b19495e6c493bf7f7f73532806725291a1b652e82fab1412204211978046cfe2cea844be25b2f871e56f165fb1ee7e93e540091ada2a0c43dc466a2a8a8b3079719bcddc79981f654d6761fb1b3c8332d37173d81664f8bbae24501360517912a917257e78df356af97e2095df5673edbf83a2f687f1b759dc7aad090163b8fa85f4f5004ad22b9b0126bc9e029217f9b157df67b8b216eaeca948b79228b32fd9d6ab25e4dacd0bb7210813d2ca0de70d7e8bceeae291875a875d9e5bf7918847bcf4acf26841b31c67a50fb4994528e9325a0b0621ad3f4a5a894183d8afc92695c49f3cc71e83d9ef9e43d3eda710643ca737be30279b139a590e6d38a86c30f2c53eed782ca2eb5f0e1f7b623cf3598a3edea9dfe0352de884f3d0d12820f222f21b474d6f8b9062e5adcf00d8b5afcc2e4b6ef351e34d11668dcb8803c5040e5193660c3175169443c2d6b78bea6b0f91cd1942be4eb9873fa74865b7e544a279e2e58e3e940426bce26c6cae964d48be534ae11921b52007a9e8b7178c19b4a32785c1f55597f104afccca13426a0f7adeaed7f7830d29eebdcce2c461ed82b682423466b3c5ac60145f5549e98fd3b69c631eb391b3f0b064bdf6f6b398f0b3680319520f3cbd8dac50a02fa876f241289af83e002c1d15b52983edc343c037a613bf67acb243c7c3f89943d14e9d0fa511cd4ea3c376f75b6a4958173d9a8dc8b14d883252cef31d154f962b55ce4f8e9a2d2916dd478299456346e65c19f518cfdfef6166b202eb566cb293884e616625f43ad01839ea284fdbe109f793f78540f4dac8cade633583b76b524dfedeb02a4b8710bd568edb04463c4d7f00f65650c8b23dc21caf3629ab0769aba78a999df9905ed625be3d293c5f6d4e8206e4163352fe90cb8fb998674b0cdc12df8d419205c1abf21b8f2afd3384ee37dbf476be4d34ea7f1509fa401597e3515adce73af004d825bfd065c877f2f0a5523f7b9879ad1fe4869fcb31e4b6474be73b69ba633e89d9b69f1009aab59d6c028c0ac6933f36d8bb1857318daf1b1327f0582aaa44560b0b69ecabc92da7e9b2f48fdec905c4db9ada1568305595b95978ab42d4d2c9b06d5f04b6c7ede84a7b31e6dbe0458e824747329a8c3028a614fb3a7d1f29bbfff396b3f625d9295ced9e3d447804cfa68dbe05f6ec5536aecba5b5689fee6323ecd06e2c23ac55032dbc3856fe626fcb65fa4d414985c7bdefa026410de0265700801741cbf98c9db2c1e70d8520c1cbd38d3df1f1ff397ccd65d76826a00ff7594b319c3c5f0b3db6358983072c65efc4f236b6d7ae847004f537350d06cc2a265d7fb2355a3a6d44305b15cef42b85082c3d145eb53bbf08ad2084ca3edd179e7f46e2ce0e7afa737d0d5f100b2f8e897359234c04b5084aa163f249a0692596780b1ba36b21d1a57a24bede6fe0c61f5abf369549016e86260333708cf6853058d11c2bc95649216c81fd520180475cbe0f345f4ee02973163a24ec0c52943760db105dd000504c28ba9229c17d83c3ec3007c7647638f1ff5b8be84b7bc6f1f8076e1e19ffc4678cf33daaf2ad563102100457796a9491c7d9617dcbb9f254bb5d942795afbcc6bf7005a26492393256ecd9d55447ee03d8c611dd9d6aabe423592efbe98320f20e6aa230f18e632cecf2acfe7a4de9fde0710093c472e4cad938730456af8ac665bde962df136b604008771bd101b506d239eb74a529e9b231688c71ad4732931ee083835afa6cd126d691ef9a114a77b50f1310bc0d4b1db4609ca83a3c90fcb15fa137341f8e2f80b810a396cc7cae8ca524c3a695a0968513372995ee53ae01ce8ce080618d14fb8ab5c273153d748b26f52bebeb46e67cb7a9e9fcaae2a57ddcdd6422d51dbc8ea268cf78e33850a45354d67fcc78edb087f9257f5e293d71ce2491bd1706a1175037c3a3d3493d02cfc81064b30ab545b8e7e23fb3d5d67893b9a6fc5df1499755e9d4503bb4b7f12c9f3f6e6485cff65d12a30ea1258c311d3e16ad43f869fa2a457432985ed6a7770a9094156b339b091649438d7b9626376f73038fe8df181a81f8dc049f2aeca6cbd9892623fb7781c470986eb476edefcf1c3660fd66e0235a7bbb1a14a87a870723935243a812c302e0fac7d78ff148063f76abc24d546b484b3caa318623f66cb7f36e35ebaf42ed7addf410bcf9926418ff3b9ddcd3a971ec6554e700de8688bbcda7fffc6bee48b44d6d88ba71da36f05b7a64ba4f9003beaf04f5841960267dcd487b85b7aa55d3e15a98a2856321b4c745ebd3ace72d3226ede44f40a54eb18832575f72237c0119ecf4a5bf0021d9a0bc1fb8c47a0826c25926fc10ad8fc60990956224a7a3b9d4b087ed62935660bce228620ffe248729479314e7cd0589927445e07168385f63f70fdbc9e5120d3dfa566d539f608fb3e81f336f97a53b3d05e4a43cac42179888db4eaaf249206bb2688ae8416718b886629d3d0fcd5c76e25510b01f2f440eec0c13132cb7f8a1f67f20aabfb18d3ed4b5b7f101dbcaa26b66ab85c77faefac77097441d593bdb6ece85e4265ebe7c930af2f6e5170942d8eddf45502b1f33a7e805e15219e3b62a7926c2fdcdd2d587e54db984fae2fd9688292adc483deb3001b6e3caa76a8d4b77fc4dd84dec906df91557aa36d85a77dc70f6d410b3208a9be60dcd95cf23055752ea37f5ef2657617fdcb43c0dc4e7928c706e6404d1b50e94c8ee11a2ab174d24048d772cad0d013c30b769d62d25dfd8f9f913a3d5527504e31c5eb290a9155d30a92bdc32bd3402ab91f576b048bf9c163b81904399be97d7293e91992e4f7bfa167092535d40cc7b3a31a784f1825ed7faf465b9ee1293aa7b266aa07f4c4f3af59361dc5db2b39a2dbd5c1e1d3a1a39c108f50aba6aebef6084ef5a9b3a7eb607c04ba389e25d4e0fe6e261ff6a3607a896f72d3119e8d2c5f14d7a304fddf29fdbd6394850628ce875ad3609c28be9d5eeb8aabd07d685cfc4718ecfd3dc84732784e82cc625ad4072aac8c337174f3b532be0b12af2b7bbc32613f844a01a68c07608178b6a7c968953672301b33122fc335cbf01e35a03a4878839a1d21c742d0309c7392804ff5beeeb16ad04ce5a51587735bf86af627c311c19097cc0273b4a91c991b5ba21cdd6f4f8089770bcba35c92762f312cd6b9622d00174062e5253f97a85cf3556ff48222e806423add05542db3aa36fc19aee99bd53824d0c821dad5cf30a50aff4b8eee65fbda77354f768dc398c98ac2afb46647a9db99a50cced19edda7394abb1e54f4f772b081f58d47ec88d5f87305294ca8a42876ec15f8d8fce56e89022682fcf510fed21d3eab2456b638f44e127a3b53a7d4bab531d8d607508e8618a66dcf57b59940412124c58c40a42f68c4e0987491b3d9de9cfbeeba411d7cec5060ef1c78a004587f42df01d7cbc867a4f7de049c4b0ff05d1cb31fcbb7844f6a22e0052dcbe64b208430b7e1cf2bc59e93a6312eadd0c46f1b6704b0720d64ac6a8e899f050d0da6bd52dafaf834ff0f0ec4c4ae2cf66abd08d3aa73fa74edb208dc55c71e27a232408f81f5ebd57099a6f7f5294f19ada8d6244177b5ad997c81469865d8b60a057602251e616bf629561f4339066c35f97b49ceed51e9e22eb33f5f090c94b3a94569ba1c503b23634e034bb399dab9a62b6154b33c7813eb5442729524d7bd47b3d902378c20f0ac2d77c1e9c057083b0487508f5814c90ff44edcf643e54a05dd7e84c64796d40e50207743151544fc59c2eb6021a39d8ee4d73bd50b3d4951700d8d476d7c90c3f464edfba29ec1bc14c7e62aa553233a049406c33aaafa7b08ed318827b3cc6857201ecd74fd5e99a3509311b6d655712bc618a5216203837a2d17302c2e3fc5ebefd98b4691156e27360981c461c935813ee0f7c15cb99f88f8f1e01435bc6535a3f9778f60f7d3cfb4220144e0935f4a10c7410ff42f351b311dab6ddc70aa3e11de65674b6bb321101e81c07e486578d6a9ba44d210918bf178ff1927a3455bb357a06cf1b9f241eec6ae72bf2fa586370a8eee76df0ed4de0b8ab05626fd872a7b9666ecf9981c8309a8428c6c7e9a93739b60f5a77fc8c71f191a1bc37e763cf7b890e8f118d77b5c2cff9152f3624418e9f3970bd143454bc05994a8c7db8c7c644cf75241296cf2fcd06fa313c0e974ebcebfdc6de06a7d1e1ef4669d75573bdb7df0dbd351103dbf0e405a3937afd5b70db51e825f309138411f9b36099f57cc7e00ffff71346f34ac197fa0b407677a8f9e11574e290dd44a2b05c11485db2c07b1b9d653a8c4a86563a0df453ad0c45ddb807b038696f12c0a7f26057b3f7512d3634a9d20e28fb9d288da39c7b0e37375f12806fc0b16b5bf8d57293d85d5b1af9019643665737406d9ac21b07b75c80f24fa82c2e4b283e259d59d7b6059a89c1eb6331556c13749d3a843208f534d71f2709a42a16b1338b6c29ddecc790895ee23c897b47003fc4976bed645034c439237829198ac6722f5a0f6aef8b1c4ac3d363398f988386d80fd0a4d22a03b093f061b2ba37f1dc01cd18b81d33443d7fbe1f431fad632e8045aec524e7fc0c8f5b952d6d68556e34faa5aaff2ebcd52b178454a90b2dbdfad800d45756309c7526f0cff7cddb13ddc7304a2030a18505b251e3453b3286d6609561b4916d6022cc2c38431ec5c31cbf7ff630845c2bc7424cc97984dc8827baaec1d2161d71557dd78b4500ebef89774c91352f67589e4195f0f7ecf777ff50e960a3546326b4e1911561779a8316c0f4f1b280ad5bfe2f3afa4c429921641a3e665bc704dd7d13ae750c1c49ea8f411b92285f79fb8241480b8c362e95c9d2f994799114c18277a154197cd97a642a4291576b2f50636d426483f0a6c96c87cd7003ce727a4bb11153f23a9de5f6a072785e4888d34a15c8925d3e60c8a3fb765fe1c332aeb56c3a8134b14f2f6667a96e30bde68b2e0cc34ae02d617a9d6b1ef7299e582d924f5be910bea4784a7890d89a9c15b5b097bdcfffb4d4d55519d5c798ed1caf5be3c909c5372a78f0bdc6191f6c606b2d57dd30044402cb66a03a6d51b16fbe13fd4545dcfda6c64b8854a43e99f62d715c912e8e9fd273e994c7b318de19586e032436ce4c9fb7c41099472363df7eb41a1c1d6ff7cdc4810b14920df86e5563be67a4595f133123c471f26dacf43b1226e5ac9e07a9627822ed86a05f96d91e12b108ccf792177b3c76c1e1c4ac0dc318cb17ab40bd471526355a58a3ec11a01b8960afac2edc56c66f7385ae483465aee276b005af3b207ae27b6d831f97fa43b9f2a90e61ec24cdd95317825c9190e1f5b5c834e113b6c", 0x1000}, {&(0x7f00000040c0)="e75822fbe9625ecd5b69f80d49cd60e8235d92c010578eb7194457c781b1d62f2ae5ca9a6224a7807e190e6927dc1e2faed1a8c5d6b26d8b73865aaceaf82faf11b78ffc6b2754a9869b237c61c4b299c0e968ad59841fe510aab91ae4c6cbbfa6f367b44621c88ec59b3b04b19eee7e3ba632f0acff67971f0c3761869724dd890c3f24ae0426f3c25bada8400f155265a2425a7a077d13925e73aebc28654087b2e49b8e9a22abd89f732d9b4c412b617a8251aa403a8f5c46f917854ad2a2644fb50ff543e2c8f45d3c4080a27e7eaf60fea40eb05d5a338dd92f", 0xdc}, {&(0x7f00000041c0)="b4a8a8bc5480413ee82f9e03e7c3c1d0adae77ea1e9425f4638c7a43bcf1e6242e43c43ef0942360f22a1bf554e0963ac78edd5e59cf0fab9dd7c700c18921ce9dcf4b014dd22596b161f8a4c1cfa836bbb8b7196cf7eb497b668aa34d208cdb3b63d9bcfe897fc5b89f667cc12681db3a8e79696fc89fd6567f9d44fa2d51374b7e43c28ca217152acf572f60a3f14f7581014f71302320c4db13f08ca10e177c50069cf02c3d9cbb09c79c423ee7291ba48fbe14a281056b33ef1117abe7e431686083812fcf654cedfe", 0xcb}, {&(0x7f00000042c0)="066c4fe0ea89fa64a1830ba9ffdd0c79177ecf0cf6016284edad1e5857c33a2019bd08df5c3af7b1b719e145d0a54df2cac25e04", 0x34}, {&(0x7f0000004300)="8b89f5f34b669f16768879a584ec59dec3db0b1e63c6c685169c2bd5a61b08cc49a5b1c3dcaef1f6845e04407903a0707c469d15fdff8c48b7", 0x39}, {&(0x7f0000004340)="0650f75f11651aa660c58c6d278159288253915b3bf991960c54aaab6ff65f0fecb237bea89616d962844df02ab6be53612df2ca1de4db4906b1a1e9e6765dc292b17452d2276d31c078aa8ec78e395aa81004d2e331cf2f6f261fb0127a217a5a4b9cb195d7a39fdee97c1869c03ec64ef8cb73a39ac810426473863a6d973ad2338b293cf62b8b06566b73dc5d80a51bd10ddd064eb476f7a718a7f2ffc61342489fd01b2d2f48a6e4048eaa6896ba763fe7ae62ec84cd2b8fb94ad245f746141e80e9fb", 0xc5}], 0x8, &(0x7f0000004500)=[@rights={{0x28, 0x1, 0x1, [r0, r0, r2, r0, r2, r14]}}, @cred={{0x1c, 0x1, 0x2, {r15, r5, 0xee00}}}, @rights={{0x14, 0x1, 0x1, [r16]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee01}}}], 0x80, 0x20004080}}, {{&(0x7f0000004580)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000056c0)=[{&(0x7f0000004600)="1b4117e9ccd919767ab836f65396ae33c1e0385605bcdad46d5b9815d5d9caf2d266d3ea1c74533b89cc9ac60309e585c6e7a0507edc190e49f6a5a5466b306404714a92f6b8d7334d7715ea0148e52d0301fa01dd2d24d4", 0x58}, {&(0x7f0000004680)="e6491f54ab4f517320840bf4ebd4a671dd885bbf0358c6528ae3a6dc6a8f25961218e0a8f2a4d30ede01b01698de4ce1e5c00a4d74050b2468d4f4dd2c340ade1775239cd304174ab769780be54949b7891d9b955b0cceba38b78523d95e5c73a2f676aac2207481aea5bc5b80e02077baee1064ce925b5c2da9bd909a393f044c60b61efe55cd8baf75906be93fbfbbb9b0c9f4a37f9da79a878e583eb5677e850d50e147de0e213419df215611f65ca95eb5ff58c6e5a6d675f5f94ddfb79a2eb912be6c2d5a6cf01a00eae0e51e304b99785f7b135e3569dcf41b5f679c81f06a52cac7e30cb66ea0dc5ea68f819d407713f3e0f85e58ba8996be080224395584e32ea9416edb78d142bb7c3c766cec59bb4e5c1c711a374cc5f1a51868f4f4ba535790fbde0b7d44141a0c94a0cda32c2f1c82cd58e9db4d4c01359d73ca269481e6ed056c7b7a1166c6384de2cf9cdb08797b0d922be283b907e5c63418ae4baafb73fc04a11a44eeec0580bc6910866ffffb66072afc4bcb9c0008009913c0c23b66654d760a369c6667eeeb0ed8b839f2a142cfb5203160afc068090b5906496e8db3f5ba90e525533c62025ab6026c443390155f7b06eb8eefc006af7701a1dc49b65e15c22e47f4be9a5b98bed4cd8be19bfcab9aa1c6999577aa23fbedc0d011c00ac8e95e08ba5c707450205edaa8ddb0ef2b582994dd9eea08a88d576a6b0f210f472494a20c38c0391a8cec3316f08a7ac4e134ac1628d7d48b41c624ae20c2ef5c37aca0017a71b32ff8fd9fd81e779b2e895a5e2263023a46140e5c9421f318ebe094247f4eadb35ecb6f7aafeaefff6d59a7353ac173d2c818bc2fb4c9cd81b3b0ed77a4db8c3711a0090aad543a4f8e85c121587c2b17ce81f12e61a3cf74e52bad62637b4ce9e55a1e0d3640728ab54bfe14c99a71cc0284fb540df6b88c871e983c45f9091f9329f59ca47a5b6d2d721509feb198d5a4512dad2f65a812c30438626b6e0f38c09e8766706fa5128cbe80bee12efe9fa8f874f2cce1ae27ac202b0dafd9553364d3c5e26450af67a54f78b64670c4eef95e703f2187ecd22344384aeaef9447b1064357c231f4c8e6bc1a8df0ebef5a68d49faa9f444af76adee21f8cfb46082025ba25a0c27ef859382bc8006817f3a3994bc867bc2b2e0acf66b904b2e49f7bbe27023c8c7f4bd27476522c8272e609e3ae9d23f058ca783f9b63b27737e0bf22d30a8bd03ae8b80075de1dbd44812ccbf3619ba93345ccd399db2ded1cf9bafbb00eb02efcb8503377af9701ee16107adc04fa3cadee04c6b2c54c167c4169dd8c75d01b89e46e0ad79f220b2d21ceded3f10fa17efd887260020e7e6d01e62ac02ca9956cd20f0fcfd462746ea94c8061c3cdd6bc1c1cc25ae625d0190d77603267930eb06447de81dacd18e2618d6a1c3370174569f88a74ccf3a9721ea5fccab573b29d4b4d2a2fc639abf544beb3a81fa1798f098a14030f4c8a4787193ca12b8b3b40f5910ba57bb404a669040ee58b3660cc47e1b740e62470dd67130273f4a517ab568733644457efba8d2dd32892115e9bc6706617828bedbc8d73f516c58eae6a1dacadebb18b8a4bc3228546822b4ebf971f64a077598e71737a296bcc7f124561cf2731bc3f2f28630dc957452fa346e898e558f7fe45b54de17ff48befea3c1b47acfced86c0e94559250d403e9d9c3c945c827c7baf433fc1d7e9814649f00c824993b97e3a963c83d85921d0780a52fdbd72334faf5d53036a8d9704406071681bd72381979d1c9172b43f5e4b9b3782df1bc6d3d638768a8307c1adcc8573dc0e93727d8c7f176fd2a9f91a1bc56150858d9b3449e3b5e34519198e468c1bc80bd3e53d4a29ef808e2e6e0c3a62ed16bc19545409380a3dc9f04adbe6c871965a6e8756e85ed6c58441957b92e9cf9abf7d4e33ec366c93702bebc3062782838af4076e7a48246112a0b509d8a5d4fc8ecc78add2926ce8dcaae744a94618d04a2ca1df7082150cc5f1d288aa414f3136e1356e23cfa32234cb94d1aeee195c0e37857a712bbe30901d187c47b4b45ae30fae64bea0a8672d3099117985543c0a3b9f3abf306f0ab539842a798fbddf1b25c1a69b33513834e6094fcd315a86756c09343ec56ecedf6f87d4b12e64381ae1233afed806fa1ee7855ae7300a425a7f3dd200f806545d7416622bc90de53c10a2f1eb8e66422fff528e87eed5701f5af00fa2f230354409e6ef4b62a9de22f2f8b65657a050b2f5c17dbc9ea9cb9796e66d376ac57559fc61622050c0fc0b266130c90703d29039b54819dd36faca85b7fa2930f41f3c90cc65530fb19686994091b5026161033d2cf14ace36e0a10a817c533c967ace67e14810f128d010305111bcd895177fb997e93301c882957a382429b1fc774bc7ddf36d745e807391c37236fe9fb44656abc6e664616146949a39b081c32daeaf5be7d010a5b9dd81942eac33aa997681b99770db5fe411d864f1cfff401a907915ad96b5ca3fde6c6b4fe98ca7edc26b5e7110cc7611350767e42343322b075f2622f624c566c4d7fbec163fbade3ddfbb6052384c5679f58fdc8728113e9b50cfca98d9ec92e252aa9f6df6f69ac2dc0d264fbccac4b54673905372e91305f1530d18f42189e65e69bda6adf5487c1aaa2de7932cce8f7ea4edafe076d67c8a54d00d7a2dbc8765b900dddf133a39561f12f16fd7bd696cd68abba582ef8f7249665337e5f3f9f65f8bb45d79606b1942c7108288833f98873970939d1a58153529d3039b8700e4ae245b95456da2368be31aeff7e9747aff3dbb5b60b155fc5f846bbb61ef2ebe416dcb1c6a9fd896f4052ee9e0f0e67e6105b0c13f708a480964d8bf0c95c5464688c0732175e5244bc39baf11d97763e7dc7ba8bc380adbe7a626c4275920e46c2684ccab605ca0709fa84b61cfa16312ba1d864a6f18c20a34f00f18687c3b9c2b0fd9604624a5a09cfcf35f61feec87e964dbc4d60a4ab0be55b4f5e4a5d9817269596c08fb8a4c3b3e23367c0923316ab5ac0252bd937cd5f7f47968df3270907b1622a6d6ce19dbbfe2ea4d1677755cf30e1458431f456a749a4dfaaf62222de25cf534e551ef0605fd027d8ae06497f57ad3d35ccc7b6318d17e93aaa0c9504174662e530e5d4e070afd660db7ded8554fb581fe6e9cdef2cc867e82b50cb43f1cbc64ee81edaa52a29c02faaf3b95aa62963907b9de04f657baf12037901edd46d8bf2922323d65a142a210fe6b7039bdea6c3dc318b838fa30856a3ad61ddb1a5d038eb6a1ce0a88266deebb1d1ca847fa5910b78841b04e043f513692beedc81fc0f1d1597a104ff42940e02c6b2b50f3f0bf3a523652c6333ece693c70750cc8864168bcc08bcbc52aec0dbda5b59c85746821ad425f0389fa53041048aa708b403278f3cf0e434bf35e6f0ef8b88d54bdf7a599b2b5ab74cd9fc116e252eaa8d734504695273f5f034123f375a0cc70fa25385f08fa2687e9dcf39e0a44eac24fed06fa62f744a00e124e8e7931b84ae3b8a8de6bc834c685cbc7b8bcb76989eed71d808d4eca8aefeb593fc85ac20c4bd737e841e2379696f9593aeff223c301524a81ef877edaf98aa87100b27b6031d45f4c054100294286bf23c6057e2a8da3c8a633a830e4a4ff697891783bcbb2ffee3a0d896b77542b3028e93812e7d2f004bcb62b903eebbeb23b427f2d38aa09fb35323a1de4ba708769ecde4cc5d9150afeeee6639faaa2b546b7183144d362290ee385b224e60f5834427160fce114d8c0558ec1d32b5e66bf2ecef4e0a3023dade4df26abc72efb9a1ed119d37962974295a5150c73ba27aa7bbd17c0b322e50f9b4a1c2405dcc2955cce0dd46b8572749d2b30a3900d0d5e53b7473d80e26adcf15671ad7f4082df0a8f535e21b2d2839e4db381ddfd4124accaccc1cd0e1d88f5279bc53a520e8622244ff6f19dc63ded3ee5b520765df03069c5f61fcecdad99c00512849a53f8b27518a9b43f32ad53103f6636fa65977816bb0029fb3fa799002eaf66590c1eb7ae34a0ba33ec988cb0266e720367c5d24cce4bee8cce08b20de5825f224ed1198858ed62cd3c399ad47cbff5b8dbf7a3aac9b9029c2ec336e916e80ae479d321d17535d28904f7196bab1f13a12c74dc896bba0c989d1a71a493002461f024adf6eeb5d2755cb7bd33d34feca85b66c66ec40993b3b5ce955cd853dbddf6971f88869e021bdea59a671047101299674bf854813c8d32b5d3a0b52bf7951d5508c4428770f8d63aeb35a2650cef4e7a9898e31156c74b843949c281219f1423769673a4f61e55b3af35869f065d458326ef954a1c2afccf3814c5ce91141467351b2c15e426c82fa5de611324d328cc296ac0fc625c1e9d7688d6c432feb2214fe2a0aad8467ccdf5638a1d9ac099d57979bb8077f11b2d7bc446bc67cb3d57b8f60fd09d380ebd8f31e7df0c07934f8394e98ac2c119417a9bfdbe46b4dca2fc21d4565ccb42616b67cdbbe8745c0b4c5d9c2834c1c3128185e3ff4b43ca15acef713eaa1055534602e27ebed79d33a4db2f99bfcdaa8ff1da188ebf7312474512349d1f9db1a3b5f8fb83f69024adb051cc5a002b2ac12a4d8cf6566947dbd9acf53596154e88439473e08a5d554c24fda4e68aada6ba1911e16994078a6d4bb16b5c73ed09e12a4e55363b48498fe18d5726c6db2df1fdcc598c7bc9f3a25233c300cf8e9e9afb68b756db9e72681e7c59acf74f03ba86178be153a6b2325b7dc9e1ae34e0bceb290ab4c274225eeb1ba36db9f5ce72f49129c470a846bcfac870ac1e3e2f096c60796633c4bb6ec07eed2f9e6be6a87e95436c79600046c49c3e9dcaf46f6d52b90234b8ab4c55b9b822fe529c72db6299163315f59214b980fd565b215950c2a7f45502b3089d627b3b5fcb791908612c8441a00da8b7bcfdc89166fb0645636d9a0d479b6ab5f10fcc373a4d61e88ccbe105686f63457f26869fe30423432cd56c9f9883de2365456601cd3330bf2cd79598427a82034312b2765be76563eeed60ce12ef6e50db0fbbfe0f29bc9f3d03a1301432e06917682b5a42a580c8580d146a9e15313e9aa6b5008d0143723bb094afebf9f47b2d6004c9a5ce89eef0287c98d0a2204aca3f8e1fa8f9b81906adab2b53d69799f55fc09a8b0ea87ea77b303022b52c6dab8f138cd71a0d681d312e0a7b482cf189ed93d156669a8c86dfa8673fa12c2ab18498a0fae991e6f5287e35326acd9b375962fa44ce69683fecb7b1f4afca0b8e75017548774fd51409635b20400ac194714bd998da37a939480b9148345c2b4ba5888a53803d81e0602ce0c2233e67d7c66629027dafcc4f5987ca79c57aafa21b4ebb60a716a4c90fd5b8e262998b12d6841112238676c10e55536fe0ae27de72ac71440576563e067b20412437b46210adda9cc1cc7197398698c138e090bbe4c22c866c66ab15e4046bd3e6f10aea90db4b16db898607af911e0c41451978d1468e774b88a200d292109f5a106cfc369b020db2dbe3675be3e6749d6579274c858675b9ffa64b0007acf916587fa16f90e21b75b409d1d58809a9edb04a6de29a8b21592121a312c1fc0df1d488a77a6546d5daa67c10ffdd40a955c9b8b7dbc71bf013405efa69b127593a6bc4d4ff211fb37880bb358e709720dcaeca3f81a0e9447f82b63e7cb584f1497a21c78a43527fc3ed8f3be395e48c0f78f71ff9c4b8c5", 0x1000}, {&(0x7f0000005680)="65658d9cbf0cd93430a072a9dcbaf7d94b", 0x11}], 0x3, &(0x7f0000005700), 0x0, 0x40000}}, {{&(0x7f0000005740)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000006a40)=[{&(0x7f00000057c0)="a704126d4b4ad9c4d814c3bbabd928c32f47e828fda9583d793fee07257d5fb74283a630", 0x24}, {&(0x7f0000005800)="8182839ceeb209564f5191a97f469ab55e9c47d3998f44882c2b48a175dcaf769061a0a4000daaeb7bb835ab2bca5119e6098cf81955067de253a6c72cf68157b45f6aa6c12a54f32556701de45ee8efb9658473029e68deab72ac4059aa48114a1faa919f2e3c9fc7369a0e408ef3ac2abb2b311641efbe187875532a9992e8da2a99d3dd85a81754e88bf83ac1e19eed6399", 0x93}, {&(0x7f00000058c0)="cd38b8d253f650df0b0996e7e4f4261e732ba65f2a3d3c10a9e2b7ba18c74f59abaf5182f309862fde98fd1d16814c8a1868a93215bb212b8f0ce3af19851875ac946c5127cf29c1d56531add7992bbd42e310ac315793aacc2b3af0b9d7e6fa13b0ef29584a65d7caf9b1c543b9b382d5207b05fde34d474e79a3563a66ff873d1c2dd362a4cd06bd5cc4b3efb6b347b623fbd89d8ca7ec33ed7d0b81e1c937c10dfb2992d26df73ce9536ca14d30f9e6c162a47f2d3492a1963cb8ca3e8cb2343f8301a286f96ee5ef9f4a8391b548b8e1aff201238be24c48e3dc304aecaf9a39d66338a8c01c715b080ea2a0ce7cba160b9dfeea0c293fdde9e2d792ea35044986e4cba76eefcab273756d43d990973b1b535e74507100af1c57e4f2e6fba71ba60ca60017e28db9e295fbc11d1912638a189ee41dc78ed801af4570bcb84fed67cb52c834a1e6c49b53463962c740eb150c3b6b4fac7813e2e632c9e2a5de0a6d53ffb01a8039080ecfd546e65ff133a9b4e11dbfbee8305e3f6202c68db606c0edafd99e45209936082ea652ff82436a6423edcff8ee5fa114aa50ce4bb2dbcaffeb0ad7f2c6f7c692cf3d8da260853c4312bbd016507962925ba973ad589d03e64bb0d6e4262318c055b9f09609849114ab21cd9424438be61e1d20fad1e40cf89476ff3b59dcaa83813157969b16f10b0f6472113557659506bbf5d4fa9d6ab2652ae18e81f33d548899bd3e282a9ff602af0556df8317f8e970e2426e42adc793b71e98824acfc7e88320239b6e51401ae68e0b8fb7740dc612020cd72adb006bc87dafb068039a8322f664cfa3d5f3c48c8a153fd428343f25863000f9be69038b4b5bb60efbc6da59e755d1856998903bc1c363f1c22697259f920f42b59dbd684d9e56e64d75c0ae8cdff66038d5a9d0052d520349508e461cfe93ef9592334dafa55d1f0d79add50e102ec388fc577eb4a41d00e88069b537a0534fae958d77929a10023ea54c0f2b48cdd3a301ca99f7245dca79532654402eb23758f3b978278af0e8f77b9dd264d61dccbcaa843b633329d6e82b644a6819a4974084de7c5caefa88c407e336b4db965d41eadf99bd8324beac0792154cf27eab2c417dfb1deb467ece9b899eb89a231d35aeeb1f996be166270011d433ba5c71d28ae9da02006797c616ae948ebd23e2c8d9eb78f2bc8d994ada2d6d58fbad5688a09a337b54c7b349957818b51d6b1b4c1cdd53fc4b0ad265dfb8c1548e7078911098a81e4f7d4686bde8b53eb2f59a5a5482db30db76579592e57c66dbcc5e2b8a9edd72ecf1ed579089f00e5deaa9fcd0fe83fa27b9400540428f7aed1a9b1687ad72feedac2cd0bec6501528d3fd0490b24d357bca6e41a0a7139ce24e8ec152da68df18be94d2a7fa7954c3721d9b7ce1887115ea92fda1cc6b1d17b1e8cca7b45c2e292aef68b868f2edc89ee1da523fae93ac285c1b1fb056e40d799a438d9ab67de4a63d7e02c083ac55770fc0d7634a90412fd164ecb8aa98897e3015c925eb5af3488be9a831696bced14e7de1ee5bec3857cc2cccd1a369895a8d952d04cc0674b0caedd373ebac80254d4d8a50a0895ad8deac32055a5a236019e2dfd4fcec7f96d99a4e63d9e5effdc2b9abf727b4ad07f1534ccfeac116ddd09ac61dff8be04479f219b5dce0367fa252f01728f2b2c0192296b510210e56a812ff5c49deb035b406e0992ae54c728433a7c36f3662de6f5368e366e2246b899bc2f900eb8c6117e4024f34f2934287636a23e095311c08e461ee1677e7b207b0cfc9912ecf34f70bab76da7c4bda39e4232669004c2a136fd26da09241184c7667cad249a3ae6602672edc9fb2e01c9b0d2692dbdbae1f91a55701b25e8ccf0379508ad5802a264d02be1d81b6563ada9a05891e83c931b74ef39982106511d25b5e4a09e6c01cd30f02867fffe69c7874c6c24f2b335dd280e9bfdd2a2c528fca866d8cd0d9a77a36d0bef47e4196c8ceff6e68cee5bdd9ac2396944161cf4ca285f9271c45cd0801e15ea8e5f9a5aa853e10bc18798f5420860dd78e8c0743ed23937dba6609af9aaf6b2e8c29a3c5f0535994e8839b94bfa2eeb956c40394cc50e2beb3a8652c3b1d2118d7bec6dd1d9d29dc61e3f7aaca7fb510bff16e12014896cc871f73986e52fc736191aed8ae52c3b8e70f3a384e3af1c897e7f1c192952557ad78fad1352283c8653e1cdd29b4ff4d1312f007a97acbbeee40a26f757505aa996f81c78b1eabc4247c284be502c29fbf1227c71a6ddb35339e22e28ff633b2497332e92040e16221f13f64120af2f73243dc3c8466e793f8992929122a43a6f668b16093049515553aed2af21516afbba23e9333e0694873ae575feeb95961b15acf660a158f646590deb13bfd72327fe87299d9a54ae802738ed2af5245856de36763c70b328efddbd991938eadee5698123bb5f50a8f8fa60c8399ed3fd82bc11de9f79887deb6187eae8bf3048fc5a5fcb2fc26de38f5a7598cfa1d280a20dbd6105fa660eb8f61445e292a3a498ca1420a4e7e730aca0aa3ad34f73074ee9b151ad179203e3a1b1a988501cb87351ccb62b49c04a734eed9ded6299f8baf2bf6f357de6d5f0175a40c93560a37c6a254e2b5a19693762b2cf815763091d4b71ed23df5ea270f6d834608f8f77fdc5ce20457c0d3227b6d68a72ee45396a69845436a6005997415343750a4051e5515f957b794b6f4a2243d677a7cea96dae1ef405150611f9eae89f16577d8dd2be2111d59a4989fca77623a0f860759f39bbc819c602449ce9eeb029871a270b8cbb6e74c87e95dbb4f6aa417011a7dc7fd13480c74ea88d1256fda8207550ebee89bc90d8e2cc93f86652034851a20d593adfe935bcec4319a52decb0ad94184af2ff7976d7eee7897a53181bda7f4af9650ddbabaac00336dcf94008d8b19f78a30d5bcc03276033eb0e6e967bf7f309b49077fd6b4db66faccf77c0006016fa6d6c1720cee2c517088a554292291573a3ac235b115f8242504da4ad82e17a00eef6d0d2fbd810e012e15db2d50a4f8d62e1826da893860984b915b2e3167a897a0efbe361ff3da1684aaac564f9f8ed6dc85766359da71ca7022693d1baaa901a809ae2d8694bdc19019e8982921505dcc325e3a07b99109db7444027d399ccfd841e884f31de3f8c8fd2ddc1898d96a2b49ebe29c3e97a87a5cb92eb1d51f0590462976bc5d3728845d94db6806688d4fa6df9cf392591e38a01da7b79a09472c87a4dd43d037bf6d0131108f4725eb12f603cafb7bca512938bf2eabfd8dc1d206d2c4493ca640f069814066b7db1473276d439ee98f81ed3cdaedf685242d4d452d5e0faa83bf86c614a40a26dc46b73567d2f87e8e95ceb2d08f73af3399458841a8bbdf02c8d239d17b4d1db3b628c1e62e6df89d2999b316a7b585c0c439e3b42b82bd273a7621bbfbbae0929f8af7ce2c9620687e231800b78d1641e2b991260cf6d0069ec285989aceef62c8d4e6ff9855950140079912df18067d326224f2315c8327f9b10959979823d2e21342a2ed1d6832173018a232fb6daed2c1e2dea1d63e63d50ff5d9a4c48bb60f45e2978a955b0bc11ae9d56cb993e97aaba84a773bbd09bb5848262500ab36c6c9c9e7c827126f22c20672d3804712dbd35f7eec27b6cb34f5403ad9ac72faa10adf7abc4d1b61161b7270d64e249f76de80b69a0cdd70f4a122c2f50ecc18d12be0a0fc234045d59d52d1eee68ed60b3d4fa3e5c1701633c6506f778422468376ff43eea9ef4cd25cd3c75d9544a7e48be58119a570364d10821ad626e52109cbb9b96906e8b7f432f221c1fc29876960fc04bd58e791f12478422349e5517ca4ef433b041bf50a780d5ec7b26efd5135e216bc1a63973056e3c7628410ddab4b89fe7c6abd1b728b0c378afe40bb73d2d279c1e0f87eb6c2dc6ef2d28a448b573846377b4709d3f77932f44e6e8d5f19f4101d1f531d9bda0382ff8d09f1b6ac0ae35231cdc7e5309c4aa35e9d7a0e08b14ddbe60a0b3720efe38b6b59b23d24be411a2bbf5f4591af6eebafcb788be7335537198443e573af4ed8ddb7b2ee1dbe1ca92d66fca894bfc97a8f18134558d0bee4cc5affb3d6cf6261350b7ae90241199e618bc23f6c867f1a7511e4188a48c38c2a50303d3bdd142d1470591b0bf9281f0234f1238aea70115b4a136e4ff918c9148aa73ebb46972ce8d8703f9fc8cffd3be77c5fa3f407b5d37c7e9605e698ff1b07e4c7434a89abe4662526121c1083b4b78d31e46c8a7ce844d4168effb4803992f892934708f16b91e0456aa0fe0fe24976a1cfc293fc08146cfe1c7e443d0782e5cc59901400e4b0376146188c65b321942435ece29b050b3b53da9f3b40d558c778ee8934e2c7317f8d51aa9e6b4318223b50c322def9192c7320036017598f946f3a6a53337666dbbd6e7fbeeb9324465ecf491b076cc7315e9768355866372154f0bbe4c925c6dfdd0e4098e5f63c7dc2b949a2e3a947e64d48802e41552cc0b880a6602b8c70b2eef8050510a78b664bb0abb46352b4e7379bc29435fd722993454cdf7fde741c3f4ef0f53b095a5f45908c357ebabee4ae8f42ad916e521a43b0126c8bd71900f49548d441ee68da409b0ba81fef8b4ab10b9435b5f4ea059db4fe42b162094ea0d04869bb69f1862a84b509041bf3fa40426f5d56f6733a9b1911e6dde15a9011ed60ba4958fd8c5edab857c66fe2026d445cd4a6c94a51ea07fc4f66ace3b787dcd81ba6fef0cb53852614322382a4515fcb6c7f63c2cfe417bfb6f4683da484ac663adc5f63638ce96d5fa2a204c6f4b27f9cdd3fb0c020615248a9105d7dde687aba67252e1c52f03b5e43669fd4e78e5c1b7ef001739eb19cd40e3d5b840b0c899b5900b5725267ef920ff704fc9a224d3c1e95eb70dafaff84911a04a304f5b86bc29bf91bbe520fc66f98f1b9750f0cd4c295090d3614f3ee6a8a35a7e6efa8e10f1d309187ca2f2ce19dbb328a3d58d5fe911ea6151eb038bd8f49a756aa34b337eae60e8f8724109efb75779e99e8117d45103041711ce1f1dd8c339909126d993d73d3dd8f21c6b95960f095f7855fd5513dc5d0a963633c2d9409158954e0895e567b3557927937ab81f9468691595538f6825e9a8e1880c403c80ff15ddb4ec8da639ea7abbbbf09d53cc103f0a3425e452fbc1cfd6d4453d9650c0c3ee54fffd3d5f60d44ed4eb54cb1577ea3ba90bfeda60bdc241d2675e62c26922dc75f97133dc362d47bcd2869cfa50e6ca1f5e89524adbdba230875d2c26a19e038cf43db4940231607f1584c623a6d2b11d2f9bd4d0b7937d8eb3578cfc8b5a7f2cbfe13cbc67fb6d697b17083f0671dea373cdde8112d2a5bb95150aa2a8f43697c89f86a56c1d2138b71778143183f6ab1aa137953db67a4ee1218b5c68e8e231bd8a555c1cea79b9cb0c23ac2dcb7a3dbeb58b783f0fbad9dd35a5585deccc74e3e2b7d60db528253c8774514a9bf0ad47363f42f349a43dc40f921ca8ff04dc347be210714a4923373bf440a7e4fb4d2c202757daf8e899d1999e62922bd3afc3feaa4097474c4dd275547aa69e1e5af9a50bedcf25caa99316cd38e309949b8818a402c31f15740a919378cf1feac2b2166f8558218c9e30bccee7b0b494801584c210d1a1d5cc50c3cc674eab6be778004fcab926cbf184fd52bcbb52792370015103e47d87b3f8a0a5f5d24d126e2cd72b2", 0x1000}, {&(0x7f00000068c0)="c513f3367b8f7a0277ebecae305000bf3a52be4deacd78cc62ff68fad2645546642e90157c76fb830105aac6d1cc62a2d63de0cb8dc4d26350b924e5db6ab5bc1125d258a82a6abdc6c00c7cdf81b21ddd4a5906", 0x54}, {&(0x7f0000006940)="b3ca586de4419c10140f129c6fcef0495df6f4ba81f27c5445383f52b898bab2a06e92d378009324f59670c0a6594d1cebe8f24595e0059f7547e3fe092fdeaf81938c3904bb4d6dd8b1f171306ad0fe4a6a43051686166f6c94561a2bb25a1f8aef82af04e9dbba49711c44712929e9cc4d740585059f013ed7d2b9478ef16caf8ba64787083e0b8861b09f7248d3b3cf05bd4b55fa09ef80bb7564002da96e70b75224a720ed6364fef125b730f18beb30bef60a351451dd22ed7dc2db8afb1bf76080115ced0a23c8b02581b804aeb09ca5dab23bf67a5c38bc7bf34f", 0xde}], 0x5, &(0x7f0000006ac0)=[@rights={{0x10}}], 0x10, 0x4}}, {{&(0x7f0000006b00)=@abs={0x0, 0x0, 0x4e23}, 0x6e, &(0x7f0000007e00)=[{&(0x7f0000006b80)="965548104755f4029ef58e0dac0dd62e9a043075d648305451d73a036e24a0818a1db2e62fb30ea10834d5cee3f9344b57077e2649d6bc46382939ce3ac0924632fa683504bfc7d34bfc19d0004c131b3d538f0daa76d0d2b8b76597a0f02b7968b745148f091fdda7fa45bbb0b249d44f3cf5414b6c4d3641d10bfd7f4d7de5db887ca8789e9e0f777b03b42d8adcf16ca071b0afa7accb91eb917db2cf658905e61cb2b4997ea00e0a32ed669a19ece1388e5d43b82939dd64d352debba5a70a5d", 0xc2}, {&(0x7f0000006c80)="37175124ee5fb82c1ac12eba6c3ee7d3441d50aca21f760a67e9d5ecebf6a2ba6c281ca9d75dcd633a758792a6be051135445610d33719237f7b1916150db81392a7fcc17f6d0eb0e69f47f912e4c212be40e1b0ce458fcde660957c85d2a8413847984f233b904ad25330960694ae6a31b9b967da003871b9371517e4cb50899a8fe02b1493ee8b4be5649edb24d3f2cb80d3502a7eccedde976d2122e66ec175cf3db6fa5376acaf03c8e936a78f7186fdd95f9a769a00e9f29b8de8f7a8e2b9ff20ae416430f315be3e8c34d698946190e4a55780a22a756ddbfa9b7e20324333e5d04d39ff5563c0d7ba70b71bfc", 0xf0}, {&(0x7f0000006d80)="b69420a159faf034189493ade7323ad83dc73f8e9c50d7ea1934862def3dac1cd7298d61f704ec7a75485daa965fea64d8b4e02aff6c17884835bfaa72c02a7471a4e14d4d1eafd987c0502d75c98869fe30c09a7d2170a0d0470798a444519e06f793beebd2d31d0ad49ce0de0a1ec1f30cbc4fb395968517358a8059dddd224b8a1ab6ed828fce23e6f037bbc02b7cb239250474ffc6ecc5ab263715b8e95cd0d66fb57ae00e7bb9e6eb6c29e7cd1ec8df382ad9c40d32725197bdb6f1c92014f1fdc29b4c308de69b1dd7428028060e35695993353549a20b0e407e9e00374ae8109ff0855bf138f569db585b34dbe1be2781ef3e90ab2e8aa929960215ec9d35e4d9f081dd298e008546241cfd847efc50d154a2cf3f710dd6df1a35697baa49c5bfdb1935b27295dcffcb8e4c2397dadebbebd4f5581be7f90d52422a7c50693193fa64f61d6f65e77443f160387ea39147e7670b9112151d93e69e0265ced1839f5523c97bfe1469f7d0bb95758c66b0c0d8ad018d8949bc6f24fb5bca631383c7b79d6b83710ed744ec82fa87f9a9587df162092275e2eb016c00173c4fe9524884ef18dc02d3633acabdb4aadb3de51f9b1a28088d9ea8fe2a69c97eed0e6f05113a5a89b744fad9f5fe8a184450824795ee7cc7947b4a8708fb3990e00e6dda13ca4a4c2e6e5c98db0d7f2b4cded79cebd6e9facc5dc95ca6d92f85065f8a155f86acb2cd15f86a9e6640cbe6f1ac9089f349c660368ff6cd22f55c1ef59650bb463e5c2768b6b9981a5618bd0e79b8d49957c2e6492e70605c72f27a177c093159aff0f499072c508628fac9896986f7ba3bb08d2681e493207517cd9f45eb42d79c2325a6c305c520773f2d0e5f1fef2c0de2af854159357cd86d9f19c8a7c5c52876d6e7f352ce0f60337f433e8234cf25b3b830b84a0517f66b2bdaf6a69ba03c90e2c50f61caf56a156d1083d069f590e9230a6c077e1b5bbba16efbd0f26f3733dc2a7e8709492bbd8ba95e229ba7bf71be41c554b6d320d5a35e0770c925bca0ad84b306483477ebc90513e1dad9b710f855a67251c4695002bb93527cf952a3e0b8c9f24473f7f8653f8a64d96cb81f0465d2af5c688755bde4b447ddb98130627d0dd176f5280492874e2a86f190e16836a934589952a000cdad6a3e308a90b7502487cb953853b0a2996b47c204de7a9dc39e2f6672c630ae635908d4e4c9b1126ca50090a4f4131d71998d12170d31fc85de3a2f9c2188f2cd3b3924df80788b840de781a070477ed5b37cba1d435dd0f44595f98bb09db1d738e72aec9519a0a5fc3b664f750fbd22dd0da6324fa60058f500cc5a38e7f0afd9f333718efd8d023cc31491e822512fd47ef3fcc17c9899a28264148ed00f7015bf062b30809686b1d6b05b1a7ffbdcf8d2d3473182998620066070c7d41c4d98d1f2b6bae35b8f7030066f458de7db2e6ca501df4189b7f45c0ef2e406d4449b9f33f57064fe3ec7ade681cba53f2b4935598115acad0516cdb0860813ff6162f5d13dbe4fc4cf8442d7c49e7ed86a9940282a57ee80adca60fa1b3eb33bb7f8f2f474d1f3730862046f3158b8dc3eec4d529b44ff40093ac8f6edc3e9aa6e2de0cc6f3f75e7b51ef890475a3002aaa2d590e7c5c9a7fce56959731a88c487060c2bab12dd65fc45cbbdf4543a0638cef222db7abb9a84381db17d1d1554a865b5a9ca57f66a061e16780c8746470f9e06001620133fe637b99b813101eaf9f499a66ad3a86feb68b1e39794d20e91a4e4e5bae36139e18bc9cc1c86f575a178990f632af792fb50157175ddd858edb68d61dcdac44f630e8c8cc210b389b8ce9da4d8de0abf11c03e87cb7d8e64f7e0435ea5b4b22952f49bcdbc242aadb0f56ef2fa2f9a5fb34e4b459c4f0e26e13c3cade13513f5372049ac28c75a112ae79292cb0b10b872f13e57f3817e3ea09164aa137836c62264ca8bc4ef3c6c231ca6ce82fec20d128030f1fb7f4a38d4f41e70add7751e1c0eb74a314728de6ce9d360219e16ee2ec87ba1ad24bd2130cec9100a26b0a0032cab3e8424029d6040a3d7ec8c8c284085c9c2a5913a86ee3678ccb7fb547b48404db399947a138248c3e5818eac984dceb9590bc9afd79e75afb7e9c8a39db580623c943545a94678cfd07ad38306361cf119412b431753ae1fcb5f0a2b4f87fb1d70a95cce8831e0eafeefda67c68c27221d74e562e35e36c29a665477a462942552159b8cfe7ab7e66e8f264b38db29a5f6e83b1f06a583c8b6247cc9207fbaaebb0d706c5bbfb32ff267c591a3666727b8f1b5ae3540978c94720c95c7acceff303cf955c2d93fc45a1a7777748ed04b4c576a2f68b5597627e32fabc723e030b6fa0e7fac9bcac65c2ffd9c45a35ee57ce8af209277ad11fdb80f1ec3b98ec396e06160751922c33de2fe99b803089cdd973c01418aad923f1beccd9662087cb7858f3ba8b4cf02e70b2ce2099a41d1940dca9725b658d9aaaa4a35a97db6272c74948d36deaba15ac9c72da0e3813e5691ce3b7e78399bab4a356d083f53cd6f365680ec7d0958e9235574316e9e732af8efa4dd2637e1001f031a4efe45c3a44e76d05bd7f8dd2c6d0fc4630c1477bde7e8a91ea35e9ca151f15f2cc4a4f5c27dddb1c07f3e922ce01b0fc471942b2399e6c959124c140bcc2c15f5b1f2f3e9c9a53f6704de3c64737ccca688ea56cbaeca5d4da21c9e6a0fcca0ca69299e07708217df5e244b70577d8e5f156e25cc06083561e8d777a79243f3d365ec0a0e4343c14a790a40dd09c8c791b5f8e6b6175e9f00067eccd6b619c4bc1dce9460ccd4a8255d307c153dddf9df1692fa7737d0c8109dd23fafbf814bfe18518c02dc35a8030e89b8fa49c1396371f858d55131b5bd05a7bbab80796831cc8ec0c37b936c9cbe1b108737d75dad8e0916730070dd6f22f5d50fe477d0bd72bb510dbf72032625ea1d638ecc9b671cb79ecf6dfe4d6ed186afc7f511ad7ae46f510bf22c92ef97732f1774d5f89549b6bd37eec1f5046f9e09f1370c6b96bf8e05d480686efb3afc4b04e634f3a11cd41067ce2b889e9ec678a3b8eb202f3aae3d6c663ac9d704898b2d905a057bf30f1f86428f88bd6227da6fed209e0e19c42a333d7263f41417c9de6d24e6593b8ca5fbc4f27802e2059fb3ea9f8f7460b52823af6d217940cbd9e74e2e8e538c681aeccbfc3c5e7ed3585f44e9a51dbf4748f485b77c63da632f6fad5bc6ef5dfdcd75b2fcc820b12087700a6884f2f2348f3936b2a7cfdfc8e59eaa1a326b3a686c57ff7cd5f64a515a2b74046ec0b60ac09b033c19913e24dde13cc7816bfe43505ed75697ba7ffe95c7aec475191e44e41363fa342c1973e1798e22b190fb23c5e2da2eab0a29321e7a4de2de38e91625f38e6ead720dc8872d47ab3dd93433b9fdc4244595331cfc6b070082f1e7df785fde9e8760967bf2e6b27d4f2d43b26b255410596d36a12add1fef87c920e8d19d63f35007fbf554bb8bfce40fa345e7c41907cc76402922960a8219974c2e766ea003597ce01f691330b746185fb680d7ed0dbac495de35e8b3eef8f87c2a708a6ee3db9b83ca44afddd348bb8ce5144efdba18643b520bc1b2a02f2cdb670c11b87447412af1fe9138d2f93aebd8eaa797d6b2ad4dac9f9cb76ed65039cfc9392ba8a13227c66698096de48552b082f86bcb4db7647ed49c6387c699bc8a462c948059a64337f9de89e728e6127619354418b83ad4d51f62f3f6547e1fbce6407647cb692639ba5652aa906c41ff8a25646cdecf68e3d86ba41868af948640786fc84b6b7f507f8b37a2679fa886aee65d2b0c58e64542e7f75ff7339c94be5e52e7fff0e066b89fb21785d07fdd20c05e58f8c0a605e78bc7f693a49ea819682c870207728c2d08cafd24366b5c96449929a1f74359c222ff51e681c69d22e711b6e5bf4e210cc643b96aeb079cb0b21af7a13ec35d3931d07be1627fb61852c161ae2d7b043078e598fa6ab98162d00330762a47caa4e830c8465aef13fb9a14b94ff3792793edca46c163b953bb9535680720067f25ab81ae8aa3cc9ba507afe82f271a93ed81f7297723978afff8295ca10bc6caa804460dda1a1ba1af50e8ebe79703b974c97c26618fe78b13e6bd4b3e15ba80652e5ac06b4677029b8e627efe1cf345c41433268367f3e5e418ce0782ca6914182e4fadbb86e48c9912e1925826cbc8bfd10aa9f0f3566277988dfdf12822fee10582904f6ee6a4d1ae2e0bcb06882d322e3705cfc7d857a5de8c721c36246ef734094dee0c16e4bc74e320893873353f749f4e7f293432c277f76a241b6078fe2bb6ca828ad4d93914a66bfe9d331719c42eb58b84a7667b31598ed7ef569e9a3bfbff30469830ef03d35dad5749b1ddd5a077e0a6116febe9a7cf7de5d1af0f47dcc23e3e8d29dd297911038aea4a2f037eca691789e1eaa2c2c74c6a792655a0222656528ba39bb76c709eccfcc55be1d64048c5f3054847279b993543a2562c3753769a5eb1c53e8f5f7e65db21196e9b3f45bef7f9fa0659af35629bc8e1279cfabf40f23911799928adc50263f3a1157794cdfcfec2de2a99cc2cfceb2ab2c75f280faf3da3ede058661762dc5433b012fd916d71ca351a0ddb33ea6099d746a0b5644d19cf15a6d8dd9c86a4f92fd771ec9792c5cb19a0f079707db61d37d497811bbef290145ef75b04811818b0423b9a0268f599986c9f2ba5a21aa3e1bbc159292cc1cfa8b95c964a7c84a7653920359ca32f74892c7554f0b00bd3dc2686f9a7433356fda08c8ecb06dd75530047ae45db3edcf8201a4fc7e8247cff7f602de9cda133ac0471f5c86faa9f3a381c9010d4ed739fe8e29be596a1ae8caa1d9b37d9ef9f5d00778e125cd7902072f9b72fee1947faa6977f84e501cc9a1d26e312599c01aaef89d750f6884ad951131d57caf1323ad63a337490aa673aaff091af6fac4eaaca1230cfe670b4a7d829ee109bc22460391e6d2198c70baff41b58759308ff2db41829cf78af75ccfdc907a94932319e1c29a247b19fd249dd92f7d1232bd7acabf81645383d8728e0c58ad87e368b1d124f1f7d03621e9a4191a981a739a3c8a337a6b4d2d548dae68a893181d29f5b1ad592bab974b8b1b2282c4c8060a358aab6adc826f080732ba5b02dd906ab4e67c4a7ef4404082c1d83d2831f155282631fb782df01e9aab40ae5d44f49d38d3adb614c670ae69a3969632940c3f147fc963ae49674764e048bc3aca7c4d8472c7e5c5fc16c1e6de2b0384558cc43c00caa226ace14044b65873e6cf4443fc4723cf451774b296cffa866a8fa66a99eab74062f39ec8c3d7bf7b585b5b8f47645634774723c9cecfb4fc44dc16000b3c0b7834843b136a6ee3ad9506f8e0f59dcea7bc4975ad611c7bd09f9e25baccbc6c9efa5aaa440e8c07160f4cfe60a02457824133063ef4e8809d53a1621236b757967825a5896cead3f2317f3726fbed57f6b91156bad8f7b9cefd74a51e74634f771195fb2a48f229fbfbe114b79444823262b7e4427550badc16f1f691b0945c2c829075db54b67b8853a78b8adbf228fc45a9d82e2a46b0e3412951ef18cead026ddd00f6f1fba6f050759b54c1ebd5186eaf069509dd5fd3207e0fec179d3bb4a633d37bf1eb591f6138334c464db5b710a8e4b1484ac92700c0d28f04335670563fc1686628890b83a966e05928147599dd1e778365b162e8f745515f4edd3", 0x1000}, {&(0x7f0000007d80)="025c0ee5b27eb267095ba74b3f3b694a104c54cdd14c33ec7038fee3b43b724de1332824e514d4eb15968206fa3d7f5011dff0d85220d0a01e90d41476470b83e50bc878771970fc3a825c", 0x4b}], 0x4, &(0x7f0000007e80)=[@cred={{0x1c, 0x1, 0x2, {r4, r5}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee00}}}, @rights={{0x24, 0x1, 0x1, [r1, r17, r2, r1, r18]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5}}}, @rights={{0x1c, 0x1, 0x1, [r19, r1, r0]}}], 0xa8, 0x8004}}, {{0x0, 0x0, &(0x7f000000a000)=[{&(0x7f0000007f40)="15efc9b2550993b1a0fb0036c17affc7f0acb0e40feb6ac0accea9ea3bb729f22013e8800b532b7eb102dccc59370cc4487b247d8b22def4fc9f34ca8b77eeb82c03f9cc2e3aa157c922c12f7bd55c0c3a86aff67f499cdbfa4287653f23fe3076de6d480715b756c803366783e30225e621de0e167f1c0e184f1c91ea833d343973dca5ed0e1c949df26f3f6178", 0x8e}, {&(0x7f0000008000)="02599d24d4f1543923a8f7e2494f77e08c86e276c40624596bcfad95537947f27efb0a18d35d36ae2d26d1cacf8bc286113c754cb05d7046ee581be96cf7b34f599b4e385ef995a5a9fcda67178455dc4f238901797222d4befa6da6ed5409a2b2779b4f66b9c622e30970a14a027ceb88b6eb4c7261a1ac314b07af4bc82ab981748570f782fa0b0921a16a996e04f6a6ffd3339e163a79a2c22745b1614717d718599f25fb2cb553817655c89fc57dd0d897e68b9d440ae3843f9c3b4cce7e3b67e46faaf5d6d592d6f858173c1c7bce902e5c3a9e23cc3ee832797c81aeb049a47e0c3b780d972f369d1b967557385dcc6bdb3bae02dc25c8cf70123d15d5c0ebcf02c2118ff6c59f1150a0c6e1102cc5ecbb91b4d0271bfdb7eb2fb1dc4af06f41a46d84469ac867913de1bcd3e4099d838daa71e70bdfe8e20d55c87c69384dc38b2ecaffc82f531d92efee548e021b0e8c61b5c63fa23513aca31dd963fbad443f31ead140dcc63a68ec387ac55bb9fd8948315249a9a058b01e85a6c1d9bd04c4aacf13ee4069a1f3d9c583acbb51dadade58c90720c1cc03b4d0385df4409c0a74fb54a02ceba88928c80e9bf541be02f44112efdebc8285214caafb43d77391751637dfcf9f2a9947c962be47edbdee035c95d1d5c9987257df2dc82cb39f19c6fa4bf2752824eaa26fa1c299c9b0d237ba290f74fc8fd5275abed35cda588eba30aba211f053465eca6918cf8d1909a41adbb2b7f41ae6adac0068c0234535586dae12af803ddaea6028bab83876565bc34c49d12fea754006a2f4c317b402aab989372771d2664a8a01ef46dd3bd2a8c367f1376fb495f54d4156e8a040e37eb5efeae899175cfc1bfd66eadd24084655433c108a5c8de2342bf9c166d2d1e67fe07f707b6a273ce10bb7fd4f09b5211a1602465c520eeb8c14dc47098a39d473fd7304fa28e48809ac87817e41167d97cf139a999195de316ac112909025bbbb63f5f84f5dcc879d3832b12f24e6f317c2692ad74e76617183e432cd5eb470fd679ff3773432e15bd6d4c8ffe73d2f0074eeb49d6dd64be98778daf269d5cd636f7c91ea515874edd88eabf4c1e32cfe3b8f27508c8f932d48a580b3fdc4ff7aef6d88b04c3223c928cff8ee7f7fd47b39ecfbe1faba5c9910715f3744b2ce685b7c806f4a81eb973f00c12ef3f0b1ac96aefff7e1e5a35f5fc2da16e0fde338b92d9048bf555cd65e6e1cd45bd1974edc0ae89f84470ed865edd52b81918c8cf00ba3a2c83d49b519349744f47c42cd8bbf685288c0efb84b4305c1da387f21e2455563ef41fc3f1b17f85696ebef9bb39fa917a7cc01616da835c6be73004fd8abd9d4027ece59f7c67a1019d9570e040c9551007c512e5e12f0b6f3ceb5070503d2cccb283ce51e3f350ca83c0c3a058860c07be80219b9cb14640a5c3984a1a0d95c27969e5a857cba99f55e17e0534e39136f72a51ee81b661f8637f963dd0a2e52ead459f6eb90933ea10020954cc5372122a412509254829707467a1f6daf141fcc9c29fd5baa3b0f0a7c6cf878cf78104472dca96a9cdb19bdae205d4fa79f533d40af0c782f50f4208417beea7121ab2f6d025c3e1e926209b43f46618b8b257a6fa736931e264b82dd55d9d6c23b8e170844ecaadc4c4b479e36cb05b2303ab8d24af1a89121419483fbe6121ab50af9e48f3e9603b7383914a0de869b38c074a4e7ffc7578559ce04c8438be8363561e5b3f953b82c4ce10059378d748800783c5e5de55c61fa6b15e9cb094b5f24dbb6b550c16549bf457cb00aae098389a404f2afa03b49567b88f309973b3a6bf206c7081b980c4abeb0197ad54b9e97dff610de519a4b8b057a4ebccc08e27dc84c45ddda1e5df1f212fe8834403c6b5304ffe9245a5469d53647ef0d214d82f51786b176369ca5e5d8e0dc7409b9c969b85fb43816919cfce9ac5bec76f82be3d16df8b1f9e24db81fff072c72216ede52362a2c370c11e1dd2f79325c6d85dfd608b8a8f2556bfd5145b0555631860f02a083ec6b0b3f3bf84bcc7a04a9121f5c32aab482a7023e78bbb414a362755f55bbedadb3d7130ad1637de86e3abd08827d4ded9d8307ea5b90b94f2e8afe73ab9acc2eeafc21c923cee4f6f5c710c068f4a9a0eb1e3c32ea36cb115eaf4f66d8f6d067babce0aa341763aa57202a55a6656057d60107ca7bd7cbb5710f3ceac160becf95a8262416be1d876b9d9e1430246ff041503f5aba8153bf9fd5ada50431631c0f2a062eff4d03119500b854561db8b439bf5fc50ebbe88fe77e751893720b2d30cf9103fcb8332bae1c283c0ed748d09f3617d3f54d8a661ac8e3ff3cc759884537844ed9b44ccc64944412f0a2c87faece13585cfd9c28477b392109d46462e3025fe86a49bd7fcd6675f8ebac0b1fc884c8ed5b559b741958e752cf68abd90476c7865637a526364e91e261b1fa0b3f15e1c74917c5097493b52cafb9106b13e805e43f35667b8bbca206b79745a97a3544bfffefd396273080e30b5a9c3ac22cb2bf2ed149e6c669a834f6168dd31ff2f6f54b139dffe31895ce3ca52db8dc2fd00c92bf62b4eba771229dbbb377fce289267d4d4c920f6c15db728d0acb523bbc5443868a9f1fc4b0645a55f3c966cf7082de0afc0c941ea5bc4705bef3fd903ae8b9185323ca52717be09c8da33a7b8e421d768e7fc945fb1e4ced28e21c2a3b9221b11a9348989130451aa177e6362a964c61e98c09cfe74812d9c722d418a34a9f4b0bbe76c97051a8ba4b9bfe88e737c946bec63897d02d57a0e422b06480b29ab742116c453cb9c0c81aa4bc522fe7865d641847a42de4188cde311751f7bff6c24a2177f9738c35676103c17ee65629f08abdac11cca5a63a84b0b4ba713d0acc6ad8403a50dc16b50e1fe2c46b2a7ae3e6ec483bc8fdefe4c440ab32fcc858c9fc0c39a120174c8992cb3be31a2c9c4c0f913ee91a1dbc6ae99d741dabadf461444259202c7ed01b2ef5ea8e6ea78b0f97e94e6e2b885a2bbdae86467aa59a42b4532dbe744b22a75c94f5055ab6c084cdc433073c32db9f6b224c619fa4013bfef91d68592c42bcdd69335b84f6137571b968f63d9562a936b1a1ec6429e44f737e8cc1243d65ca8e00e134deb20587daf23cebe203c3b3e87b6a908ce3f0fd716ce360d3df059f718458c83876ff241e5dda5045798019505fdc40a0089cd229c8ac8bb71d86e00b8c002aa953bc93ea919db01b7d6099c94505206750a0e76f58242109b751b15a4d357da7cd93a5408331278d2045c18f1396aa4bc26220091c3bda1d2aafeaef03558cf945df1dd270a0efcdd0d15df5a5b341c7596103c79a4ee1efd1b5021364d3b8dcab9dc8f76c73d90438d86c56f1229c4e9c79d4f55e52f7a2219fe7fedca07c3bc5b83841986cc28db1f0505f39c3c511ed752339b86ba529daed8e166705d5881cf24ecd0d653243cbba4c882fa5efafc77c9e14917d195a98ea55729d80126f4234e05c079d327e6b4323be8c901c1c98ca58aaa544ae66c39f515722f7f9d3a95a7e7d3bb1df7e20cc2d828dae935b5a2819b766f56b808a932291379606d7ec6a855f329893d4c9f74d03e0a5a374489ec752615b0d02c338ab35c8cdec5830975eaea7f0f44f615836d30bb2b6304184d5b7acc96de04de860664b65c591d11fc63851c068bf36b68d1c07544c4bf007293f4969137b03373360b6249a98f14aecc074ab1a86960b3ab3c81e44f5e77707a038a6c3b51aede78c38b58ec7c97f5fc13e9a1bf492056328a90e9c45948fb4447b398ddada3b831b9b1dddf1ae3984b478a185c2e644d72c8a84b0837d7d3690262e92b113e6e84272be13963bd714886300a8a2b5c3a804f8e3fab70a85e93389c6127da112a0b8b391ead2dd1c971cdeac81e042620b74a383969ed4e458779aa30a0bc1dca997b5e833896fcf9ad3205d3301e233cad43324fa30bb3cdd04ecf4bcd13994854ee51484567ad079b98f4c3e11c6772027a05959d0482d09ca9108c4d35401c9ba2bd3ea2c2f66a23292fc0ee6616739c0438b05cd2da4498ee4dd78864517c21a1d32fe9f0883f2bbdcfa8591dc9d017866d78cea38d8b944e596b5d2d7476687e846c511888db78bd16034e4fadc7dfb23e301502ae9887bfc19b745a47df5677d3608e224b26124e46a075dab373029cfdce53752b1001e0259915777e8583bff43615c204575a2030033ca8a803cda650f0caf0f10cc73db5aa3c9429dc284a4751ec07042c681a04c54c36ef172d97b78a3d8470055508883d55bced1797279790cb39022b166ddcce7d6c50e64d93ca498fb060de42b503ab7cbec38a6675445933d649b98e4a21e7327145d5211df01311c21d27b1efe3db301b117ce431e992c022ac6aa447c7086e49f82835768bf2f4455491e0928fe7d987a84f870564962f238b3037c2678b1427d931aa2c0c0a7473ed0d55316eafd48c988d2ca455330bd8196876134d9e5162a0e1b71399f7607955d4564543e4081cc5b43c860e874d270f021d93c64d8eefc319afcbcca82f8ea4ef1ab8ef8e60a94de97e74e6974d7813e7275ca1ddf2763b19b2159433849d427e5ad49c7b54db8710c7590f3d6d52f6fd10a0df64f36cfa6c56c9341bab34f248741ba2d262efef10d0df7596a5111ccb4cdf20e45bba0a5fd04f72afc8930c6b3d8946f7efabdae50d797503bd4d21b95472bf8dc6d8f943c9f4b4ffd9ec24dcf22aab49b143ab41506a3cad30e39b9ef59ca7f1df664bc8b72c7b53701e7833e45e8867c48c14919393f9c489e14ef1c3e665bff5fcc45582b94507a1e2e15b8ef71e249ead603e2ffb744172117b78965a055fbc4c0e9af72239f2379ea8b611869f8ad3a17f4c8474523cca66e3890d7616225a2e9051b865466b6a277b97f05b7b4ea1e05be4cb72eba6a8a7bdc0acea61389c2a89657febfda54acaf3b83570b7fefb76fc93984671f54c0d68bb2b60d7dabb02a356d2f58914401bfb8eace23c53cf3b3c40d43a29b23df65b1496cacc28ad2652e1adce9dd90bd0312ebb7278465bda4f9e1fc4759861dcea669b7d8c73a087afa89b6e6a4004be49e8f71a1cec21f0a5118e0146a0f6de189cf6f6eff770906ac8e917e46a21b0e8f0978dda0dbecc7dd2f05ecd368c82498e91c11cf5d4256262256ded8974fabf81b6c791e27eb9772adad71288ddeb9c7fa5a0de40a5e482915c101e96038f492d588a25456addb8f2dce0b50a24e4095cfdd89f0dd4dbd9b4b3106746821ef8c41b9d5defba026e27671d0b758ee567a35a57e68f668b11f610c78d601a85b513021b8a45c524a7a609d51ef97d643369cb831f1951ac58b3b7223de27ea1e1bf007f44eff780c6efecd18cd64b4f1e0922633fbb339254ec1c90e25930d2e575dd86492cc1fb0fd1669465eb412fc7587b5d506acc96e3f1a3a407fa25b44b354b37e8e15f27a876f19aabd7a2fbab6a89c67f00591c5d0beab22cf1d5f9168cb8a9d942d76fb13ba85eae998b4e7c35dfc86225b7b91c5e2272fe85c998fd7707e8b90ca1af1a7fdbaa619ac696b41404665558bb67fd312f8fc6ec6502c89e9a50fa8f945125e2fb98cd4f89af1e4eca007c04aba6a6701f484435189d17edab86406389e97d6d48bcd4f127d100ccba814d43dc8e57e0d0ac24ad587030e69a688c753cd6c941f19508ff7195fe4b6d076b4681c3cfaec843c74688261a44f05b1df2bad0e8ed38096ef", 0x1000}, {&(0x7f0000009000)="aff351691af1e87d27f55334ce6fdb6d70e649da1e19b8f77105ef8378b00eea382503ebed5474b507fbd5439a0a634d7fcd9235c5be1b92a7c4825a954079e6f76962302a3f4028e7b7c8aaeebe39e8a485cc62a4e4b855d214614645ae8210dbd4774c276693a516f0a4627ba69bc6d27b8480ece0ff7572c22981b6052a40f88e4946c69845147a5de35fa7659107fab2e2152775960798b426620cdba8d9fbf957a487b2fff2a0efae002ef00729fb5cc940e3ac1198996dedc8ed0c7522bb9bec88c1488ec5b92db363649fbe752b7944c63bc0d53b922c7e3350298f4efd69768a93259bf682562103b1c8ecac356d5924cc6144ac6ac8f8d5dfc3a9183182d9f58e2ce8495abfd9f4ff1174216468c82f65ba18caf398c60a8dbdc655a31664b1d187cbd8597d7ec585200b601a59dda748409eef16a44dc4db704e9ae937fe84cadf83a7cd1ae760817f5f95d7fe5eb12fa24baf26b5d0950ea3af1a7acbb8c815dd0907169c2286d953dcda6db3390a19471790bb7f3602168da2114e88508b2cb54f3f8fad02a924114add94ab0dd75cb813cd839308a0d9d71f47b0961ad31ab93a66c544d6f5806b6d10a58596ecd1fdae12ac26d99798e2d7cb86f7e3ec9c5e3dd67bc19cd624879b4d933a277c52db3936c274d3cc0fce0a00b672de9a51f140a94d8a28ecf9a0561937a2781f84d838f9253be2f6fc4bb9d098d675196e87cbabae5e4d3c27689528f3eae6df2e2dc4bf1de58d29077c2180ed0c7d9d6d2e9f97d8fe5a11e015c0d4bf499e2c049ef35ff873ce0dbd9f80f017ebd09210657d98dc9be1c22d5ed2f3eaefe9deaf78ee127a4eb0d7e517a20f89286878a2182072a479ea9f05ea112000c3a6d6e1ced57a0c061f9a36e83d15ff719f265f8efd1f8c90c3d10d7adb21dee1bd15494848f703120897ea9df84147f7dc38179de1fa844c3ef75e13bfc4cc7cf9bfb7d30f0d7a268b1ca4037b48e51cd1be5265a2c5035089543d34a9aa460b77cfcc9cd3b098a44253dc4fc793dcdc80ed1598efb393a5d37b32a0d504b39888060fe7f9d85d74ed34c8a27a51128f4700d283cad9e8d33c11e60475b42d621758c5802065199b263a489d8b50c3e072a727bdf27decd148f2c8c6f24dfa31d2b85d426931c35731aa3a7063be3e36d43152a09476b61a7680d5a94bb55d700f9df35411e761d471c085d6760375c4e5770b98c149d25443b0e8593906775157cd33bae3140b6b84775ccb40b4366575775dada72ff88acca5ba4149586fdd19e767fd7d14c4d9559f425b15c559dd6588f5cdf01511eb738a164933cf3fa6a80cdb624c08b055e24d9a0e2d399084450e9dc0da26a99cad657081dde9fb1e27eb056af65e453a539fe2df46f7329696a4841bdb635e9b4e6664262d5c3feab4dda6286573479c46ad5252e8869bce6ba3c04a4ebf1d346a6d9f35b21fd7a87b19fd819677ffc26ad5f4bdcffbab104eb3b703579b27b5724bc1e4ea1e8aea5dc58261f3b5f2c09d58c0d3c6a8c0d9cb24aa9eebb6bceb442b4b90899b61b756580a8b9abab2981073cc8f7ae48cd28a7e25bf00bf46af1198291aff69ed5ba313a5cf58df2e5f7d4a2c0ada2613f4efc4fcd251bb12657c658c1ca0fffed4927a16510da0a2c6988a443b3260433f6f294de59659b0a59b83fd144f00b442bb1500e4df6a63c2c358a1b858bcf7a71e4a24fbd5f1819af4e737ab2e2dfd365b27dd723be72dc936e34eae96dc5a9b458679038982ee080b611abd0124d9efacdfeccc13df76673238045d6874531b22cb7b8723eabf8593a5c01117b403948e2277cf57c8e6a8510e711b7d983ca967ce7ea04699c88817a70a9d68160623df1b4f8bb4f04069084f9a8978c6fb0bd7725d7214875bc63bd0f996d87a4049e0db0ca377482baa1f0ccdc5e29b2aa9b728dcd464828deea9f6774143a746a32b4ffe4525195b6de95a35b56ac2496003cbd0984d44b0775246111e58d457e6ad4d07650414134410b7d5f0dbde35d306e41aabc7c3ba19476e3490f08fd153dc637d5ecdf82e2166215a7417c4c499bac7adc22db2bfa672acb1dabfd00e4f7551050710e05d061250b37760649372aeb9d485a13372fa6ff125970606b97383ad84f7174e5c44a84dbd30825b023e77a1addce224cbe9f14981382ca81afc27fe2ec78fd43fa1d22db3ed4ba73948a32be351f50165930dc1904e0479b763946d3874fe8bd31b28f98750959aad9dc764372bee138b8ed06d76c7e5b7b1b0e97d9bc2e55b23a1738d22e11c7448abeca0255a63947d9517bf4180e5ba44a6c2b4aa724c6445f8c39db4473defb89556e25356e14cceceb9ab3e20d4f5aca625852cc44d3f054ecf131ad7c619ad230a57ebf21cc1ff576f5b485e8c3d958b2c106fb4657a5ccb8c13fdaf747d625bd101237d1e14a740ba903b59de052067ad609606a15b113cf19fb68b00d2c4fe14712456c32246d96b9e6d54765e3925edc40ae22544fa0a84b42c57f84a45299dd0921b50c7f3f8c8a6610c30379ebe5378858d4746e78e419202daf5245039474cfa03b9116e338909d9e21b49dd956f819e1e3922ca9b775b6f272b5f14fce90e70c2f7ed94df7c43b789746e8a597db4314a0a657abec7d5cade3c39d3d47b6405bfaca5e0b647339f39961f09d6ce646f7d22493dcea389dbbf182163ab744055f00b14ed5ca0cd3b1c5d441397b0a66d084622b6d71d5434b6e7cf01831982f7404ba905401ec9a9f25207628dfe16fb623b476b1f643c212afab6e64db408e5d437f36806c104554156bb208eabd04748967a81919af045a61b286523327ef8e34a52d2648302be0f6058b8724b1bd7296bd7665690418f4933ac31400a076fd2a6e05d4179b9441de7c99d218c1a2f5f137fe95a19221a79d1f4d15985013ab23f06ed5a3af36dc2aa0374c68b4c540ae3161880843c813947aebc8b4e43064051ab1c398d9df3762de5d397aa6d05881053313c2c397339e477ac01fb2a91136fd54e6b08490df7865735fbd2a92c9493eb46b3e3fb20ebe1d067ac01b1b6657e87dd4a8f34d7df39d085f818adb4537cdf2f1656281ed1b70ae1dafbe7d15c1bd2b18645931b1a6e987fea01b545dd9c6e4b78ce41abc6669ab34f2d24705ef7b8dc83dfa6bd06e43e5da8025e7a31387deb2ca4764b92310e507770ff26289565483b3dd910199bc8cf8aa95f4246234fa1a2779dda84280e7cc3c9370df4ea6b2ed6b1d89f1b86c9375e507abcd54c02abea547768b99a4a256e23af984d73bbbfa3eb547bf97ef15c7eb91b07dc0e5143def9f17f6e946f028838afa3182d29f0831ff191f79780d4a95059262b844bb6a1edb9f5ebc0720e70683c88554f044bf9e38e45a3708235007a81b8d85956648ffe0a23021bd7907c47ccade5a6011fd0bcbddd3b34be66d72ea0d0ef88492355a6ff5b023f4dcd6e06886fab4010beac590db1933f41645f8f2c32dcc29590785c0f725854416cfb4e0674f1827d2592ac6e7b42e5707181b39da193a83486800cba66d62848e9d1e9574f864525d98cfadf93a8d744a6011dde076d4b4b3d668f7290f7e99c4ce90da0a08eef2990c06c21902c8c8b9c7bdac59a483b568a9313d3039ad26d16eb66364ef3c8b5086deea9b8bf74bbd8d13f6af82ec8f54d5a9330fa9721e4148efdb2ff522fe1b5c7e1b701dfc64de14da40fd90eef947f80c4ae3ac4a4c9c0deb0bf2ccc0ce18c2f5a32d7ddb55d949b8b979c9cd5395f651cfd07179841dbbb01b534f7df46b09f836bcdce068f9d1cc99cd790fcbe82a30d3ef624eaf04ab41a34e0219ba582b0f29a7588e712e577dc0b1080ff724debe6cd47950aec435f892040a1b2100199c16a06f9b08a3d3fa57a9b07ddc930098dcc595c124d1c73336a02201b867f16b4e39a9fa6ad40017f9eaacb66753b30b418b04ab2b91fa2e8ef3b223ffcd282f9ab8cfe9068874d9e527d3cfd3aa15f0117e7f7c73fdbca9f4b23910118d96a39d0d9e8d82f15184ae1688771b4e2ade4393ceacec43249d066938e9311318a34a44a965ff16ebdee3ff822a1d0832d64bc3c29ed4311ab85b203c4791e6a315964c9c18ddb2711596d9579ceb7cce1a6c2eb7127dbccb74a23e5105895dc4aba98f78608f4c095d7089142482225aa9002da623e0918d5cc6cb404baec41a1959e368ea5be999a0e4060646d5146706ae88635fbf6834db4656160854b6c30792c71b81840ecbde042f99af816a0ffa4e56c22d04107840b52604109fd99edc66b088d1b78244b1303598a693edb261510e6d9ed731b468aee49ebdc2422953d65abf08b1618aecf1bccfdb8d0752f916c03e140fd64d554134400ba22cc4ecd7d59cd6739c18f6741deccc68899e6249fa60131c201ff6f9967aead4a41dc8c77e1132e74c55028d8d3dedad077c0fcc49c3333abda7c5e8ecde84b3290f04ef5e9c8b58abeeb701ffb3f1a04b17d045dfe9b2ea3c78e48a8615e2ed725ba8b5f81602d00ac0b214352ad06a5ef707ab52f196a1d83bd47d75ac123fee9076191e1a1b01f57bce4a57aa3983a6d064fd062da57659d4a5a3a0d6c46fa81a1a341c53b938f910fe55101369ae4b0882e804309faf4fbee721439d8e1a9b5236b71b473b16d5dd020e3e35dff8ba1a62cbc90f37e652d3f7ff325f42fecefcec08a99ceecd2eaf8529b6f495d2910086285fbb6ddce76a2cf0fbc7e0eb96936999865b36b7d3c0950ef1b21c9eb2b5a1a27d96c1a84acf4f96a61fa52469e4ecabe3e4dae71c4809d1a65ce24e97bf5f43dc194cc4d17ae4063fa13daac53716b89c32c9cbaf979cfdccbca22ed91920735ecb10890545d3694b1fc57f46eeb4ede46d7ca44bc126fb02fd212108409f3c8494f2560497d72b5d1e6be37302415c54b8f0ad9db71fd3ff9fc91a72b4ec49ebbcfe4c1cbdc6ee869f4c1026cc76b1991f0b57f77f7b7efb4db2412dc28648fb2ed165f47dc32ca615acb762e6a2025074bfdba89d3f7045782fb0653462545d3f27dc942293dd5e79d49412d3645624a35b218ff3c80df589246765d59b646c094c56f209f5db3fcf494c95a8934d33d13f6ca8c57e4c8aab594aeb80d1c06ba35f57302485cb130241f63d9f2c623b4f4e53fcfbff852f5cb392e6cb45eb0982cc0cab84e6fe70541f3467eb85aa02e4dabe5fce6ca0c61b3e44fdce4dd6f6a042a01904d8c13dd85edef3b5f0f610253eac920bb9f15607784f1da928bf5df7799cc432babd75fdd1d53ed1180d5dc78c4ebb171554693abb736f72de8b1f87f3be5553affca27452f22977a43ad677159a2ad3bfb6a2784f87380712469b4c0947d0ebf9f184de3b249cc6787e6efe0edc36b691f857868ffbbe7a91f503f71688078d0aa52c57b81e50eb5d707e68312d48627cfb8787fd8b7477c06a83ca29b4f625a778d4b9e8dc717009a7c825324567e05f05fe996d84452880a77dd3707f2b8fe6c1da18b5e4ad1237fd07a6ca35243232b3f0e30b0c00905ae5616335a137f4eb31d4725b728fc2d3ef77f84ffabc20aeade30626b034ee13e55560b1b2187cbf6878a6f6c4d86757d4223d682670ae03da735a671184a5686e9cd1dd796bd19250010b3bd233fa451ca428ee95d4ce04511a583c69849e24e426aa9c5db99f1d550596e580aa763b6acde9539345e2516ca954a33c1b59fb0305f6f01aeffe4cfad667f78675ecf4f6b417f04508f7d70cfea60f33753c030d136144037", 0x1000}], 0x3}}], 0x8, 0x11) (async) sendmmsg$unix(r1, &(0x7f000000a040)=[{{&(0x7f0000000200)=@file={0xc563a0953a94baf3, './file0\x00'}, 0x6e, &(0x7f0000000400)=[{&(0x7f00000008c0)="26df6347310b2ce65912c6171ec0949f16097b0ea5a1bb75508e545f2350baff8a9e88c623f80fb50d9d23a97eeeca24ba8c256f3414f3de03b745de410f5e856172998b1318bfc12b0eed3806fe4abf662dd7575092e018629c901d2209927ebf953de877a7191044421708704820a0d28a40d00a94d1a9861992e84748d09140ddc41c1ed5305e5302421ba6461020a85d847d219f17e4fb1fb6e51986b475f53278227dce83eb443b1e48e80507ddd9919296d6c26843b417488df70ec501432155c2d1fdf91c0fdbcc69d8235b1446c4fa27995e6e68e96f3fe27efb4aaad3c7359dcbd04a78c210b10c9099c551ee715dfb5afc842eef0bbb509ce39b26120da98996de3cd3d17d19c1221a82684b2dba475eb1a89d1ff7122e6fb52f13e09fb5b7c03f10ee16d56ebad4571041fbb22c6a9afe0a1e8e35438b9bf3ab2c1b28de96b34492de0acd00ec696c969d9207efcb3896c5a8563d1ba2246c217d9f0c492634ecd78cc5bde70e82c3e26f6d208094f3bdf3b5a14af05362c46141bc9cb38d8f1701de16d36ad62a1d39412a9c9c06da88f01512cfe13cb5899c4ce955f918323ec53c03e2965ab48263ff5b65bbaf7ea84d44892c80a1b940b39922cb9c7a9800888c2ac625d6d23ac43b3c655e46bfcdb2fd033cd2ed803d9f4a0d264b5cc6fa02ad652232b5744566f5d5fb7fc7b9f0570f1457a781bbe77e3b70c0e2649e43df265cfbd5a6c38927f627481a343cbf4ff168adae746636658932ce2f0388d996ac2a9bbe0bf6bc7581fa624237f8ff556083e7200c26ab8cf24fa6c1e363618614f716d9a2e1af85982cc414e4ce48d5a7bac6dcce4441febc7cf54f20d6657d1fba1d543a63b90c38871b1beec8ef62281a191c0b5e3cc36c0df6e7a38d03730053c4f46f422de91ccde20ce8a282f605965bb1ace689a29ab4c327af1f99491b24ff08e73bff49027ac1ad4aab135727d893ffb86346002295994d335798b99187a034f1ead9e18524639ca60f69a90ea6bdae6d1fa3cc051900817d1cef2c06ff45577933915d47a892a36cfbdedffdcd1b084b3c29ba5b284d1efc953cdb9c4a329f9087c2c557cd0256b8260786933ea588cdb5898a6a187d0bec3135db5fcc35178697c14bab4d2cf6c861f6427f1c697d3d074167494830ff0e074277177135f4b1a5ebfd7a63ada11cdfe7907620b26676867dd597a6d54eabc30a1ba724d488ffe8c7a949ea373dadd27e039c17cd4cc3ef92ccecd73dd3db2dd9de79fe89d4f1626e7dbe1f3a6081a197220235974f8f63ee590445ac78bf2c346001d397481b197c2634a2a38e5f46e441a0583fa12d545c7de80fc3e8e137839f1ce5c0a2817b46b66e43152e41f04a4e5adc78f497b54c1b9e21d8f02bbd1148c0e21194d88a59e4f6a2b5065c050c053ed6e0810c03cfdd469ca1abd6abc11ae2df6475bcce9f2916c0e4fd9da5e28165c8cbd459d03a5be723697deeb9af93b15a637c5b0779c4c7cf76c3cab9e333df9b9f2297c4370c59d29889bcc363a4a7a24e1ff1836d373c45c7f7b9fe5f528b6e18b7af9924276d3b4612405fbf0eed72f5592affad74fce0c1098a6b12fb8455d348ee364e50dc18193ac587682b2d37d5d8427f6513c74e1974cbda833962611f95fb9cef61cd65e11ec3e328ac19180f3a313ee88685c95990470702db1bc7b0668d5c1839a22af1c365c5c303fcbdad7904e6609505042ab2b0d199abb93601cb00be3eb3c4652bb8bb44634ba11ed0e736c49f68e9802ec093aeb417df1c93bd7a622cf499b9c96525f54f27293edfc8bc0c4cd258084aec56ea02e8f276de1dd903cfc2beaf0e1754365e740805fdbcbe33b2bf804bbf0f8c5862be396e08d09bfed47fddc751f3dfbae6ea3224f3185629217eb5ec7505859aa26072c3aa9a5b1b3dc95f8964c48cf6dafc9cbe0ea83e715e7bf56f92ff3847e6524c92fac30eb48a178c79ebad3bc04772c8b96c5ae0b082f99d090ee29cbcc9e84897b413df84e6af5cf74e277c0e519baa9755d03be6996ac57a07fefeab749fcccc7e85c0ae137f2c5b1705535223a900f04fb04f54066123d9679f96e4c85f1bc909706997b34c6c42d705e089af12a9430bf32e6f3cfdd3fec6ad5a53666580386e604a241ea4ad95f9ededc70fb5ea7664aed7f78861963c4829ab0a36f08068df0c1c198a35e21bf02b23d9891ecdb3185f9820b93d05d3de4f089f85b95080fa94720a4e50e1593bebcd9ea123a04b02e15bf4e3ca2193ceadf9ecbf25c013a6ff1c2607681b56974cb046b58540d6396669fa4454894486016a6f783fc46793687979cd469e5eed81b8e8da97acb00f55cb887851e7023db6671ff3260d67dc041e17bddccb3692cba50061718b97b7a6964dbb23f523ef0826dc60dd09c34f05695f0e07a50c8c475ebdb96111c226a9856d231060ed65a270664da7fed3393ee69da5e27a3c142c26ff10f696f0a24cf0cfaa7b9b86b7a3e063c31273c8d8068f1ed10963b41c899082ff807e9072a09e33aba36d072540f4efa32b7c9bddf20aef302b8b8800aeee6f0e33ad2de2e4f37cae8f390ccda0fcbd854354cf7e191dd7d37da3318c99936cea27d820683058117337d10f19ce9a81b5803f33087d4387e87c84ebf8f4a24814e963a0db630c68f95f716543ee36b22fa8fc1ef3126938bc0203975bdb38c9e38045c124519b2c394e915e6e7ada1ba71406b580aa549f08561938f61c704d09264a0080322d409672c8413b17664861d334353112c1d4161bac59e87bff00d8fd9c186c8e76c8db0377380bc58f441006f40c85345e0f85af8a4b9ba062b7d816616b2d19ef699643eb891eb1e9f0e7588e4b0481d0d77e0773e9ae33b132fe6eaac99914ed298897846ab21b350422cb4cfd16334fb2b3d8ff850d79921899a31bcc4b2c4710515405f2279692b6ce29fdb79e4c7ba7c91ad21112ae2aaccf2f865fc2ad8959d5a77c69dee6ecfad26a604680b35577b621cdcbfd0796a937a0f7a4e2e075cbe87fb8cc14a4ffd938945243acf77c969edb7a619166c8c0d0fd5b147d8c322376c2d6e2e7637abd8233013f56965ce97d1ba7cd4a3c187b3f475f0896db9f376a2ae9e7e6ef093072ca39f1009d5934a9b1c9608eeb5e1bcea1aa6d9c9a0d05289f5f335b52ea773f9d652f81e607440e7a6264b9c4ae69dbbdb6036251bdd7606db2c70202bf55dde67e3e6f4c30ac4562797e55c4b14a6173ef45f19ace8cfa742b62f2d40f120f0831f9ce0274fb312ea311b210d8ef66338e72a69ba172c313c665277ee0e0bf9877540b32c9ecf3cb91191f4e95a2ac651a9f048f1b262264a189b4f211edbe43f75f6552cb5a63395b99f36e69eb7b50fbd3e71fe6011b2328f20bf2f63ad5b43bd496eccfa6f67407b97a02130d0de6c234279f742b26d7e848018e234d0425ccd92e9b6406a119f3627b7bd064e45a546a6fafa8cf214e2bc09ce4becefc7e7b6886da6387b4ee15e2dc89381d19d66a15540c0552931a170c4fc59e8dc4cabf9e8006d3c9082c37ba0e931ccef7855f89801c124192a0d8d1de93e0557ee17237aeb6f0c52d7aa6a8e4724fc97b28e7ff719c8cc61e0ea4b596bb0d121d96a82157b26ee4a7a7cd13ddf1d5e810ab219e8745ca50ffa866afd74155e9f55b4014a704ab27214d5f61068935b8f72103a56efa9b602fdf55afca083e7dfb3ab01e21d4669a2d954e4fbc66250f5b680d9074527c53f1e7a30eb428738cb7c84d110636d299da1aa8366c9e0a90def0808600ddfd20fafb662f4f0db9e2360a9e6080a3473d46f8e9fbf62c3bf37c8b848f87b7f53aae1d82d1486776a2cf44a1d41dc5265ef1ee1928c7338ef56ff3235d48fba64c076dba2986db6371b16b19160cd1a3a32e0e7ba97317856e8a1fde3d2ab32a498cde0e1f1e88519e01f9c4b51fdd068f5c396e220df70f1f9ddeea09134dee5aef17618c98eda3c293b519e89ccad03578e76f32a1f4d7dbb3c80345bc64b8978406da2866c868158575daa2858f89d9ca620d207385ab31a6ac5fa26564db0c94d4146164c6df139f756072f19fa7e1c3a95daa747c52a0f0eb80a2b9f060610e58f84023cc7834188d62db93e84ead8195f9561f8d17dabf463bf4ada7a895ada07fd0b44c7fc92856458cb9cacaf1eb8b15857de5b0654851aaa2f09cc969a2e7819ded0819f93774eaeae2521258730ab77540b253f1f2ca1788f404c641821ae1251abe6e63a9aac8ba4a18a29dc00e26352c9ae8705874a041b681fcc97b16567d3796648a9457f90dd98fb9cb627f40899aeeed36f955633464c8ad28de88fa0a07ccc50b9e194771e476d5a34905fa3e16d00abd99feca6ef5a21d644f90f4bef73f819ca57366834c8f2ec0c3bcc0e9bc8297609ae5fe5bbe97d11c7aa233ecb454db2a1319e30e58562a09863a7df8d4b48f5b3c8674d10e7d65c5f0675ef366818e7036990cd85527d50708dc6cd632f9b488e0392cea205e384e56e844d09774760d36005ae2f959cecae20c028659ea0576eb2bfbb486552bf9d0413be08ebc63aba0f96fcb765636f7fff0363e031217756244055e3d5ea5044016186b9eb43e0bb93fdc1ebec69da15b2b504ec3244e4c542a31c89748389c99273e7643e57c1f7bf6d421e4fdbe42c164114da9728ae11518e780c4cec968e9e55189f11090bc7a9b2a81f98b269e16f8b1c13e95c234ffc276906d8ac3dc10032f81197942bc29b215d73f98454b72cb3967d57dae0bbb84b1903676d8ba9ed1dcdf720d21985a5e9a1e38c5272d55e8c541e4c3747808fd23f75c47b11a21e8c715e677762310566f119bdfb3dead609900e73d5c423839dfc93c0dbe5a0de304d29f9b0e3afebbdbb179aa3702b06409b85643a51cb995a2bf47d566629ef76d674c39fda4eb4b9dec532d36eea32eebb05ce05d5497219b29bbd7b8454ff7315a0a7c473002c51f99478c0c27993a6d38f8b68efe51f5338ac6675c4b42dbfb049a42ee76a29ab7ea93613a2428aeadbc5cf34bc5702598c5bdad67602a4302ff55c36cc2560003c9d2d6273c72a15b913c4501838792231f85c0cca8b8f81ecb80acf96bdfc77cc1a80f2783eb37395b1b02e99a30c9f3cb583c61c162c94871fe3a084263b5918a1ff0423b39df5b76ce5f431aeebd92ffe6e5ed06b64ff3cd0ea9cf12914f6f076e31eafa1f1b26dde6f0340f21dac464b3b943036fa11f913f69da02b69fbd2b586bce8217ceccc05536fca81ffc179a494ac6b1a5e450bd0d316db8359da4fa1672deec9d1bf0411bdb943a2b601cac6ab8f3275501225131eb5193eea3f9f05090dfc021bccd6453be0428692eaf40f780da6210de0a05ac6ec3c5feabd72ce394313e57456155eea0b9daa19d4e3e3fbe5b0b65648f3b78c09fecbee0f00f1bf4a56751618106126d39cf065b5f412878da68538a0cf9457a50bbc363b1e2e5236128d30846e2f0cef94d961190130ae0acc1ab29ae7c419821659005ebd8c9167f23267316f2e4881c1d5b4a355659e9582c208b684c865e1cbeb5193e8c8a0d8404fe67113163d0decfd408e0dc91310efafe4f9c3f5628ac5ec236b104fcf904b5cea909f40fed759586672ff4a131256d53f3ae3ece76ce5606ba1b505f70781c9a6b26348181a976b658940041c78cffd56431d6b992fea4909a71b88b1d97a188ef010a8f8e2e8acdcd6d36fe0feaeda4268e6cff61f802e345e029ea2e838bc71996", 0x1000}, {&(0x7f0000000280)="8b7c5aa06b1f0149f7d035c4acae8eec44a4253f8aa16c76adfe395114ec1cd615360d5909ace6d43af55e83bc8b825a9d3b9426f4a6cd34a074a2c0c205afd479a1b0f6ee7749416ea304ce0de125df459b", 0x52}, {&(0x7f0000000300)="23754504d47058e50a708c5d928a28caa8863177a96628736cb5514a721e5e14082c8125cacd7b1ca49288e2a351d938c44d57b53cfc1c3392de5c3cbe9650ca88ecb9b5fa3237aab74140c2dc28b60209415eba667e171b072f172fd84a18f24fe5e36c5f4caf697c9e1190bd1499b902364c762e6c8d3835de51933bc6266002322bc1ac1bfb4117c6eea2207e55ff2572ce210743a78141736990e5298bf8c5a1cc9201bc70e6e3a3fb0a18fffbdf2974b1e0297017d7a391b50bc11aff9b8191e03694eb2506c7892227700542ebf45a5f9e02005ca245d86d6903da06495a3dc6955bb12ce70f4a8782076a4467eb27a897b052adc56f747399", 0xfc}], 0x3, &(0x7f00000005c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r5}}}, @cred={{0x1c, 0x1, 0x2, {r4, r6, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r7, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}], 0x100, 0x4000}}, {{&(0x7f00000006c0)=@abs={0x1, 0x0, 0x4e23}, 0x6e, &(0x7f0000000800)=[{&(0x7f00000018c0)="84e7419de7807afe986786d254a3807962232a70ee8f75d457990481b7564587feecf8d7d55e2fa705364d05d7c6adf0b1449042a26bd9c0ae420277e5597cfe15c63ccdb11efb11d6ae76cf854029798d1237d06221f55d501aab6e566157835e22f8a6cc877614db7eefa8e2095cfbfc6f3030145738958da325bef7aab490c1d5f022a233d991b2dfaf699b42a9ea2ebcef84abbb2cbfb69ae8f2cd58b7daa131cfdd1cbe3bedc4efe548ecb05db6d67619d30b98045f07aa23e9fb137ff01e5176fa50b8dfcbc1f5ae3c525aba864a8a876214db4a05b3a44be0c4a3841c24483b6dffaa0a36e54e1035b23040cc7dcd0dccf4d20e87d6fe706480ca5da9df715cfe8e37cedd19b09ba40406052a402aa0ca9542ffa99df2b61fc1e7b309e3d80a94f64efd69f35ad37a0fc665745f1d9d59e404b28a7ee300060de4f6678dfc1c0aea60e716838703107e3c24a9642fa6ace21ad8c4980793377040aaf3c81267c1bc86eefba67b226a08afb5592e89ccdcc785dc2e4a526d9ee7d610fea7ff5119908b87b41f9005779665218fdedd34cc7079e746a61cb7add96b9098f31255ff8711208ed51b733a5b1d5a29f422368ce6da55c38ec755a09772f2978db8c04228fba8d88b9a4118aca83707c0d792fa0155c52b015b0e23342d1b9c2456b1dcadf5a423187008aa43ab7d83bb9d41f3ee747f5ca9efa1dddbc4302ba8e9d82df09848cd78d9d2c2b7a2851177c2d458dd038f5dad0b6c63023afac0428f9638bfb1257febfae9a1a03ab122480c671748897131bdd0d3ef834ecb040078c967c67a14ee5f61a4f64f2922d120f773ae3fd8dd32acde92db5eda6ea891a6cd8d2c5c19702c1fdae8daf1ba0c16f3835651325a4594aafcdac04e36b4452761a13af62de2659757456594c7c2433741f3846a4c35f62207ebd1126436551baae5ea6c8bbdf51cec3ced03e51eaa2be3b7207c6e71a8808aca9cb3a1b84cedc903135dee31427e6f38a9f36dee06bde9d49a6091357bb15839c36039f8ddb0fa342edae7770ab0c5aa66de174a5813db5d97d9b55bbe0063d603fb851fb48b7b5db12961444b8f378d8e387a06a1c1bde3c6f5f850779cc2ced53793a462edbc2511bdce50535b47bfbebdc0444bcaf1aa130ab7d92437dbad442456d56a837966c253bf5780cd068620c31b49d10e45bd39d402eaedcfc524c0add302437df601e60db5f7e3488296e26305aed3317063d02c05b763e33e9bcf8dfb9ee520887aa44e9c4ea068240ca025c296d1e38146a6ed81abc6659d64295cfc55be5f422b546b81d0bdfa328627f6bcc866fbc88c00c33024c589535c7d97f6ae3111ec9f3e0f7814eccaeb927a520cd04119c99189f08c3983af1c560fbaf215ee57d67a602115b1c58128ee9ee4f017d6117d6bae15218d94c25d9e15f6760b5cae13626eb8688a76313bbef45d247a41eb17623707edf00ba12fce0080d007bf6204e5df6ecd83eab8b56e760d3e537bf5e8bd7bd6e8d0ab617f7cecda59e49bfca69e9fb82dcf492de21e59c2a81272ab7667db42af684621abe24a50eec332a93c365682a17ac3bf917dc746be4ac418d5d404756b62024fab85d0dc24483e69692f14f5abf1b24c85ded680af43aa4c38d0031680591f7458a75a6b2ac30f9a4a47c0badbe00c6a5ced8d4c2d1c4c320e8104319e2dae4fd59161959f4d6e3489f5e91a335d44a0712e4e37a40da1b4673ed199550dbee9a4ace0b00652f8a3bb0488b6d2f2d0dbf581ebd77c354d304b5be4ad560cd3cd6cfa01021b453381f668d73ad317a980fe4ba2fb1520363f6534dfdd29100a01fcd871961571670aef2ac026392f6e7bdad9c26e6fd7c5a8379f4671a3493443f179cfaf061c04752a1798624ef381700ce55a0424b2f8765a3e785513ecd0ad04e51adfe15c9dee0928913d87fbf87daf391bc5579085f08d022b4ed21917858d67b45aea2038803ee691d3c5bfbe4a0e6a4290216f1230137c3acbcfd0298dfaca032c551ac624b3fb24c26d002725fa796b75b4fcd15458841bc2a5c8b0d28f1c9a614e71994234f48fba29817531e5b7f56644727ede975921efd19741823b40bc49d68a2e2a2e0c8c6de7f676e3f545d2a347d73579e961f1096198576e3f1edde72c3b948f5c9fc7657d1406065394f0bdca8241fb0ef72b085e98a0fddd7ce299a9463223dfa34357f116dbf693762a7e66578a319b781b4c5fbcdf8ad1e994b4835d2f9c2298c05c2269cb5d31905dc0c7dc2d82e0d92db9f0e3b5de0f21dcd11080600bd8c8f6e0df8a1e0b6bf8423e592348b55041307e2d939d284bafbaebcd16f38ad56d71f8d2bdd2ccf40475f3ad6615213eb9bd7aa471213dc7fb449c75bc6cb9cfb334689a18fd0bac548cab3ec8c63d8c8e3dd42bb421d20845ca5a39e7a60d8bae06bd3d6888424f566358ab128a6e17396afa1c2bd121b6b6941aaf83dbe80eac4fcde184e4c66112c16c8d80d666fb32230cc7066ace8a4fb0fc67567280c0a91cccea6e03b45ccd6b0b04796ea31160d165932b63dc97b4298f0a39b6caf0f7115f65a238112c2bb387a64bfb9a61b03c9965b0ff80a2742730d05ffa4535232eed3e3098d40cbdc110ead18d7bf707451768f636375733927905fa5e24eab71300d09a63b42ab21cf2f146c0c540b7a520f0ea8b1ae4dd8f57b0c12d37df106769ae9c0e77010d1d2ac392bf2b291cde15420600c445c4b47b0fb2dc4038d71c0b6e25e22d232100694f6d33c6f8470848546bc601f0c6336e08dcb6cf661eccc5cf2432d00c230a21eabed73310b6fdd9d654ee16a904b04e79e493c855a80c2fbadb75e08cf946b58ac1b3ef6548771e9d0a11df91653e168753a58a611ff5910cee2c6447f5c565f838abca04dc6cc1959a4192d0f62eba7d9dbca297a01afae6095c49948d1b6e0289184e2311e6e3f6a7257452e9758911d47529a6ee1e238fdb93f4c574176e0805d8e3a7272d9086969fd7f3eaa5ed4c7f61e91abe6dc22b6ffc89cb48f15dfb47f56c1b15c07199709741300d4bc638246a9f69840f9736cce01f910e5bf6ae3af3c9affa5dc7268d26e06196c2a187f22fe2c3ea78017e0d643254a3847b583d34bc1965900cee2403f14e00f310eab0790b8fef28e8d2727631fa32058c9150910711969df9586d4c7d7fc60fd738de49ada181afcf96a921e31b1ef85eb48f05d2821323b0a1db183ddbd196bb72f77b82a29ab012379b15be331d107ddf4d9132199ff320452ec134a9114ffe4c9e4636790b99aa291ddc91e8344fdacd070cd3db029b3725b82b13ff28a7caf1dd7227fa72dfc867db6d86861355020ae83b705cc9fbe62af9aa1e4476707ec57a86998dc3f418ff894bfba45ff6a967ba22724aeb2401ed8f4c943ef5d8fc96b211e6632c752514d282c46ce494f11134becc50a51d7548354c8293b83f2b55a88029eddee89cf8be62339d23a7ac72156e26ea0902e1f7b445bc65e61c8a76b6611f742c6661a652d82da36bf2e457aa76b9c3340f19d25e8d5a0e8989b54ebf87711879004b9566e54e8b9df0d269f70a72ca8096f1d6966d66a7cf9a081090cf3ea99e1f91d683083c5edb3ae529a5bd73caabe2057dfd9d0584668b195a961df750e564c34477ffcd58b4333abc24d257ed76e470ff183386959982c9be61d0617ce83255129b98b70c46de328ca5ccd76f72917cd279bafeecc529dc55b0729d90515a08d5575ef21409db69b771448a75fa8489b2401c1bf2d3197346235dba1264e60b0a4ad8e0225963da86af0035f74a106ccd3947d4ade26ea5c6865e65ba6fc9ad51dc2a11f7aa2747312d8c27501b6e6f339796561e4210f32bd3ff88e557cb2d51f43d1104564318d16c227c35349effd23e93093bef312b09c14adf3ed3cc45e5b27d1e696a8038ef4188376c754027d9bc4746e44d3e78c598d9cd5c428215ff3ac8c10a2c4bf245454de8a43aab0eb5d975573c51ec2496a2b14d62f293b578f42f983789f31d6aeb9e2e8eeddfbd3091f57f94678b6b1906d173d99c467a1920fb4ea58ef73758c33352759f99d3302ad87388a301512908286f712084c818e737ed1cfd42092ae7e28b3c8371a036968b080464072808fea7bd950fa4aea78fd1055042ade67b52f739bcad9b0cc3355dd709c92455b7f8d6faa8d61278473bc0cde703e4788e01f7ba2eb555f3a7fbfa6e876829ecc1b7c619ea45336ef97e933ab514ec77ae4d4649740959b9bd9d897b06dd53dacbe35286a9aa5ab2d2550200edcfe1435fc78d9a855c48147a313511d7a895d5e16f6b7615551239e53bf0e5698bfab18e33bfc3d65e247ac944f62ea827b7c1a8a664c2150a9c51608e0719c22e3643cf0c5c7582710c6613054f3e22c13e42dcf15db6a78e39752415df811c843af56223789bf4708e1c108787a21e3f9608c72702ddf194306f573966374fc3882af7e77cf21a542502c5b6edbc54a8169da047de3371a146b6183563e4bc379ec44cda559534797fa4c6db16679ed20b6b474bb47322ee03ada85bed6a1de7fbb0fa84d0456d3953e4f10239ac9311e03a29009e880b89bd2a4a05ed0f7862d1e05e06423c136ecfe6af159a850ece88329b0c0758ae5b156303957dabd56ab7d989cb184ad0b91627ec1929b0e25b75a7c9bd2c480a4e0999f1a5422cbbfef0c6036d96bb9284c445d07b920e277a7f25858dbca51ba28c766e0755fb85ec40882ecb2399f7b30a6e90f4974208f45a400419eb9a484eb7f1189124b8a8a2f92e846528ca2be8b512326b5cc3ce5bfdff8e3cf1d5de76858d57d94992cf6a32dc4e1510a6189ca2e649cc90a3893be81ecda834dee5ae099df7f44e244548c580893e756c20505e77470db4f84b48a3b8ca1e4e2e317af499c5c12abbf69609632f7b9e2912bcc133d881ce760d1f3925b922f34266f1016ccc29ef56aad61b0a210b8cdb0b7d2af4c0d60ef273808bd9de06c4d84bb81c7675811beac7bbb503859245c4054967d899606d8674f9bf6fd15fc6c120e3315e62ace711b3c6d403a9c1785bfd5f42b095eb13ca94e5ebb27eca868101c5f2bdc464b6472314f22d73cbd731af745b4a13f7835cd8d8c60b6c8051cd445166ddc392f48412ea37b90792b470f7916331234ea2fb613d78db9defdb3e1f3e219b4e2250c64fc90c8b334e00c83403f04ed24d08f1b9862cdb82b2d725b1b4ee027451e553d1c07eb72b291db650f2141da3463270e6848aa172e6f41dfb59c5838c278ed17c853d06db43ba439ec09705204c6349fa8c9fd18a31598fbd432b186d3fb91e39954a300c35150e968bf4c722b9e544ffc073bb52ff3efd7c5234422a24bbcb2a1245d8b4349bf851cb5ce2528be8b9bdd227fd4e390567b209f70f3386ca8a6e70aabaf3a1a533fa74c823609c597a239301fd03484983e5bb898db683fb8c2d66c549b8c0069aefd0aba1d35ee9c415ab468e3a7b23858cc4d85fdf91a258c68fb150d785ca22875a7952f6ed43d7e08791017734dabebf002b54d6667a5ef951047b4834aa6ec2e384f217e0317840ceef9219eff31b6ac1795e5528488ddef90f5aa6a64cced8bb237ef70b94ee472c0083d117c4660507e32dafc7ffda13e0b20e079bb02005b16cf8eee8ea5d2c7a85cc42cd1899b9244eeede6adf33b32d69ede79bd8712c602975e5fde2425922b417d84014fbbca029ebbb70de2b8cc441fcf9c76c5f4edbc97c4fde1c0613cdca0ff603", 0x1000}, {&(0x7f0000000740)="466421a47d47df50c6365679ed631823743dd09002fdd7a9a79563dde7e30b25fce0944d50f303513e9d22098f2355010fb04a64c4756e0cb840a44f7c9851520c472122ee49b180cbafcb207f3f29cff39c71c04c18175286fe82af9d46eb06b9b9dcffddfef69eb48a3bc3ef8fd81e01b1c1", 0x73}, {&(0x7f00000007c0)="0f43e5038345cfb5ab33b5a14856cd27558c33fc625df021a38fc60b8d56d9852065a5047bdaf87390203f98e711ec90", 0x30}], 0x3, &(0x7f00000028c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee00}}}], 0x20, 0x804}}, {{&(0x7f0000002900)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000002c00)=[{&(0x7f0000002980)="df08eba0bfc089261b639ec7cb50b403d8d9e041f6f764edd1bdcd32eee1c4af1666b3e3eee568b9145e55a3deb0998c51d3a653b146028c0b4b856e08c6426cef98c188d27d215b4f8caa104920366df4d7a298f2f57a971ac3034e3394f2c4c4f9ef22488da0a83551da6f1f90a0805190f5d9ea75ea1dafef99ad0ba80683a7c8fe0526e7f133c703d0f88ad78688fd41f22455d1405fcedc19795858228a9fb18e830c737594a78e68443b775d75092ebc434b69a57cdf", 0xb9}, {&(0x7f0000002a40)="3f923d00d3c136dd8176ad686725ea1ff9314cb016f5dc00a41bfaeb9a918fd454c3702e6a815d3df04447bd24bc106df6b0bcb21d060484a8509bb4e219639b4db7ef0815cb1fa80e5eff0e1608c84984743a92c839af92ed8d9945499a5aa95bb998e1045babf7065dfc6cd1fe6d9431c9f5a561256d4f940eea19250c4d3e8e3bdc13111d16736c4a0026dbf92aef4d3b796fd5dd7f5c6044aa8185c572a7f38c7a58220cb0", 0xa7}, {&(0x7f0000002b00)="b21ac1057f742d5e21b705c1f3b78fa7db7c71b1a1b98c25a46d4b6187f83cb7e93182d09316f4ebf8cdbafe41eb693e0b65709893d2f562812e1e611ed90860f73c40632a6a6c5caa6028c2bb1ea762748444f2f22dbaa98416b5ae7d39dc9dbcb2c3007a63f0f11208b09ee2b61cab45a104d9e537371084d2af99a579469021b3a476906dd34b6124b9a5444ccdc962486ee5fdfb9d05f4a2b6a04bd43a538451562b6f31c3a244230578290fc9b1edc619a45fa1cea5d9df974313a7067daed8de30ac463a019218c60d71317fc25664a6d87abf5792e18061b1225028646d1166e0278cedc8621519ca3759", 0xee}], 0x3, &(0x7f0000002e00)=[@cred={{0x1c, 0x1, 0x2, {r8, r9, 0xee01}}}, @rights={{0x14, 0x1, 0x1, [r1]}}, @rights={{0x1c, 0x1, 0x1, [r2, r2, r10]}}, @rights={{0x34, 0x1, 0x1, [r1, r0, r11, r12, r2, r1, r0, r13, r0]}}], 0x90, 0x4008081}}, {{&(0x7f0000002ec0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000004440)=[{&(0x7f0000002f40)="c151b349f05d56f9f4c372dd0fed519c7ecc625e78e802d29227a974cfc9a58c7df6d2299d8a347b0eaa458ce9482b65943feee3e3d055726444119273bf6aa56a6c91ef7b466e6718b1a1e8e752c199b72cc2928651bbc1427ed15baa56cf9d8a82c0600887cb0c2026e6303be70d8381fffc2633b5452a1278006b7b21e3ff97775439985d16f87e0c57498ede83120a9785c4ff9c9c982ddf0f34305632b6d0cf7a60b24f6a65dee0655c6ae3d952b68439d5491fe2a6cc9c480c365fe31a6a9d5be34de0c1a5eb525af7b0a4a7b3fabb876cb0ce771831e6e9899c117f8027c95e2f3337aa5ea3a1b2fa5aed6736b7d7aa3811e0f310489013", 0xfb}, {&(0x7f0000003040)="c092ad957bd69724d3d352a6d42ee4303c6ec7f01aa6cab1503840b654382e922052c23717d1b59a2704db44c442d19a5cd507fb5df80dafb93d4c89a48b31b5c75ae790190b8caa5ce50713bb86f2db4778e42939c378091f5bd0d8606b0cdaefacb3bf7776bc625567029119a4b221768d2b7fd68bc3d00a9f19", 0x7b}, {&(0x7f00000030c0)="cc9163f5439c8454d4a7b6211cef87f1d67d99b5888988258fc5d2d2aafa73dcc6849fc4a36898952d7c1182c051f749f4a03ae4e1489f2b3b8af996f7975d99e9d15577228b8012ed02259e3c16ea2da7ba4aceacfc6cb97b321365009efe14d29d8d4a8399f5caa243cb5b1f334ec443e7f7da553064eb2443816eeaa8423ee05d5ce6900ae0019df448e1ae6f54c18b9360f73096d4aa7057b21857679d0728087854fad17bd16757a601d265111799f581ab7bb68b416d26d82e98d4cc96b4fb0ea3a6c338b35e7e4b958c06ad9f1261f8131842a56528be17b675c4e741abced708ca97732920ce03141e42b56f595e776e79151def5fbe34e78fac06e86e58aef484fea97316f74b19495e6c493bf7f7f73532806725291a1b652e82fab1412204211978046cfe2cea844be25b2f871e56f165fb1ee7e93e540091ada2a0c43dc466a2a8a8b3079719bcddc79981f654d6761fb1b3c8332d37173d81664f8bbae24501360517912a917257e78df356af97e2095df5673edbf83a2f687f1b759dc7aad090163b8fa85f4f5004ad22b9b0126bc9e029217f9b157df67b8b216eaeca948b79228b32fd9d6ab25e4dacd0bb7210813d2ca0de70d7e8bceeae291875a875d9e5bf7918847bcf4acf26841b31c67a50fb4994528e9325a0b0621ad3f4a5a894183d8afc92695c49f3cc71e83d9ef9e43d3eda710643ca737be30279b139a590e6d38a86c30f2c53eed782ca2eb5f0e1f7b623cf3598a3edea9dfe0352de884f3d0d12820f222f21b474d6f8b9062e5adcf00d8b5afcc2e4b6ef351e34d11668dcb8803c5040e5193660c3175169443c2d6b78bea6b0f91cd1942be4eb9873fa74865b7e544a279e2e58e3e940426bce26c6cae964d48be534ae11921b52007a9e8b7178c19b4a32785c1f55597f104afccca13426a0f7adeaed7f7830d29eebdcce2c461ed82b682423466b3c5ac60145f5549e98fd3b69c631eb391b3f0b064bdf6f6b398f0b3680319520f3cbd8dac50a02fa876f241289af83e002c1d15b52983edc343c037a613bf67acb243c7c3f89943d14e9d0fa511cd4ea3c376f75b6a4958173d9a8dc8b14d883252cef31d154f962b55ce4f8e9a2d2916dd478299456346e65c19f518cfdfef6166b202eb566cb293884e616625f43ad01839ea284fdbe109f793f78540f4dac8cade633583b76b524dfedeb02a4b8710bd568edb04463c4d7f00f65650c8b23dc21caf3629ab0769aba78a999df9905ed625be3d293c5f6d4e8206e4163352fe90cb8fb998674b0cdc12df8d419205c1abf21b8f2afd3384ee37dbf476be4d34ea7f1509fa401597e3515adce73af004d825bfd065c877f2f0a5523f7b9879ad1fe4869fcb31e4b6474be73b69ba633e89d9b69f1009aab59d6c028c0ac6933f36d8bb1857318daf1b1327f0582aaa44560b0b69ecabc92da7e9b2f48fdec905c4db9ada1568305595b95978ab42d4d2c9b06d5f04b6c7ede84a7b31e6dbe0458e824747329a8c3028a614fb3a7d1f29bbfff396b3f625d9295ced9e3d447804cfa68dbe05f6ec5536aecba5b5689fee6323ecd06e2c23ac55032dbc3856fe626fcb65fa4d414985c7bdefa026410de0265700801741cbf98c9db2c1e70d8520c1cbd38d3df1f1ff397ccd65d76826a00ff7594b319c3c5f0b3db6358983072c65efc4f236b6d7ae847004f537350d06cc2a265d7fb2355a3a6d44305b15cef42b85082c3d145eb53bbf08ad2084ca3edd179e7f46e2ce0e7afa737d0d5f100b2f8e897359234c04b5084aa163f249a0692596780b1ba36b21d1a57a24bede6fe0c61f5abf369549016e86260333708cf6853058d11c2bc95649216c81fd520180475cbe0f345f4ee02973163a24ec0c52943760db105dd000504c28ba9229c17d83c3ec3007c7647638f1ff5b8be84b7bc6f1f8076e1e19ffc4678cf33daaf2ad563102100457796a9491c7d9617dcbb9f254bb5d942795afbcc6bf7005a26492393256ecd9d55447ee03d8c611dd9d6aabe423592efbe98320f20e6aa230f18e632cecf2acfe7a4de9fde0710093c472e4cad938730456af8ac665bde962df136b604008771bd101b506d239eb74a529e9b231688c71ad4732931ee083835afa6cd126d691ef9a114a77b50f1310bc0d4b1db4609ca83a3c90fcb15fa137341f8e2f80b810a396cc7cae8ca524c3a695a0968513372995ee53ae01ce8ce080618d14fb8ab5c273153d748b26f52bebeb46e67cb7a9e9fcaae2a57ddcdd6422d51dbc8ea268cf78e33850a45354d67fcc78edb087f9257f5e293d71ce2491bd1706a1175037c3a3d3493d02cfc81064b30ab545b8e7e23fb3d5d67893b9a6fc5df1499755e9d4503bb4b7f12c9f3f6e6485cff65d12a30ea1258c311d3e16ad43f869fa2a457432985ed6a7770a9094156b339b091649438d7b9626376f73038fe8df181a81f8dc049f2aeca6cbd9892623fb7781c470986eb476edefcf1c3660fd66e0235a7bbb1a14a87a870723935243a812c302e0fac7d78ff148063f76abc24d546b484b3caa318623f66cb7f36e35ebaf42ed7addf410bcf9926418ff3b9ddcd3a971ec6554e700de8688bbcda7fffc6bee48b44d6d88ba71da36f05b7a64ba4f9003beaf04f5841960267dcd487b85b7aa55d3e15a98a2856321b4c745ebd3ace72d3226ede44f40a54eb18832575f72237c0119ecf4a5bf0021d9a0bc1fb8c47a0826c25926fc10ad8fc60990956224a7a3b9d4b087ed62935660bce228620ffe248729479314e7cd0589927445e07168385f63f70fdbc9e5120d3dfa566d539f608fb3e81f336f97a53b3d05e4a43cac42179888db4eaaf249206bb2688ae8416718b886629d3d0fcd5c76e25510b01f2f440eec0c13132cb7f8a1f67f20aabfb18d3ed4b5b7f101dbcaa26b66ab85c77faefac77097441d593bdb6ece85e4265ebe7c930af2f6e5170942d8eddf45502b1f33a7e805e15219e3b62a7926c2fdcdd2d587e54db984fae2fd9688292adc483deb3001b6e3caa76a8d4b77fc4dd84dec906df91557aa36d85a77dc70f6d410b3208a9be60dcd95cf23055752ea37f5ef2657617fdcb43c0dc4e7928c706e6404d1b50e94c8ee11a2ab174d24048d772cad0d013c30b769d62d25dfd8f9f913a3d5527504e31c5eb290a9155d30a92bdc32bd3402ab91f576b048bf9c163b81904399be97d7293e91992e4f7bfa167092535d40cc7b3a31a784f1825ed7faf465b9ee1293aa7b266aa07f4c4f3af59361dc5db2b39a2dbd5c1e1d3a1a39c108f50aba6aebef6084ef5a9b3a7eb607c04ba389e25d4e0fe6e261ff6a3607a896f72d3119e8d2c5f14d7a304fddf29fdbd6394850628ce875ad3609c28be9d5eeb8aabd07d685cfc4718ecfd3dc84732784e82cc625ad4072aac8c337174f3b532be0b12af2b7bbc32613f844a01a68c07608178b6a7c968953672301b33122fc335cbf01e35a03a4878839a1d21c742d0309c7392804ff5beeeb16ad04ce5a51587735bf86af627c311c19097cc0273b4a91c991b5ba21cdd6f4f8089770bcba35c92762f312cd6b9622d00174062e5253f97a85cf3556ff48222e806423add05542db3aa36fc19aee99bd53824d0c821dad5cf30a50aff4b8eee65fbda77354f768dc398c98ac2afb46647a9db99a50cced19edda7394abb1e54f4f772b081f58d47ec88d5f87305294ca8a42876ec15f8d8fce56e89022682fcf510fed21d3eab2456b638f44e127a3b53a7d4bab531d8d607508e8618a66dcf57b59940412124c58c40a42f68c4e0987491b3d9de9cfbeeba411d7cec5060ef1c78a004587f42df01d7cbc867a4f7de049c4b0ff05d1cb31fcbb7844f6a22e0052dcbe64b208430b7e1cf2bc59e93a6312eadd0c46f1b6704b0720d64ac6a8e899f050d0da6bd52dafaf834ff0f0ec4c4ae2cf66abd08d3aa73fa74edb208dc55c71e27a232408f81f5ebd57099a6f7f5294f19ada8d6244177b5ad997c81469865d8b60a057602251e616bf629561f4339066c35f97b49ceed51e9e22eb33f5f090c94b3a94569ba1c503b23634e034bb399dab9a62b6154b33c7813eb5442729524d7bd47b3d902378c20f0ac2d77c1e9c057083b0487508f5814c90ff44edcf643e54a05dd7e84c64796d40e50207743151544fc59c2eb6021a39d8ee4d73bd50b3d4951700d8d476d7c90c3f464edfba29ec1bc14c7e62aa553233a049406c33aaafa7b08ed318827b3cc6857201ecd74fd5e99a3509311b6d655712bc618a5216203837a2d17302c2e3fc5ebefd98b4691156e27360981c461c935813ee0f7c15cb99f88f8f1e01435bc6535a3f9778f60f7d3cfb4220144e0935f4a10c7410ff42f351b311dab6ddc70aa3e11de65674b6bb321101e81c07e486578d6a9ba44d210918bf178ff1927a3455bb357a06cf1b9f241eec6ae72bf2fa586370a8eee76df0ed4de0b8ab05626fd872a7b9666ecf9981c8309a8428c6c7e9a93739b60f5a77fc8c71f191a1bc37e763cf7b890e8f118d77b5c2cff9152f3624418e9f3970bd143454bc05994a8c7db8c7c644cf75241296cf2fcd06fa313c0e974ebcebfdc6de06a7d1e1ef4669d75573bdb7df0dbd351103dbf0e405a3937afd5b70db51e825f309138411f9b36099f57cc7e00ffff71346f34ac197fa0b407677a8f9e11574e290dd44a2b05c11485db2c07b1b9d653a8c4a86563a0df453ad0c45ddb807b038696f12c0a7f26057b3f7512d3634a9d20e28fb9d288da39c7b0e37375f12806fc0b16b5bf8d57293d85d5b1af9019643665737406d9ac21b07b75c80f24fa82c2e4b283e259d59d7b6059a89c1eb6331556c13749d3a843208f534d71f2709a42a16b1338b6c29ddecc790895ee23c897b47003fc4976bed645034c439237829198ac6722f5a0f6aef8b1c4ac3d363398f988386d80fd0a4d22a03b093f061b2ba37f1dc01cd18b81d33443d7fbe1f431fad632e8045aec524e7fc0c8f5b952d6d68556e34faa5aaff2ebcd52b178454a90b2dbdfad800d45756309c7526f0cff7cddb13ddc7304a2030a18505b251e3453b3286d6609561b4916d6022cc2c38431ec5c31cbf7ff630845c2bc7424cc97984dc8827baaec1d2161d71557dd78b4500ebef89774c91352f67589e4195f0f7ecf777ff50e960a3546326b4e1911561779a8316c0f4f1b280ad5bfe2f3afa4c429921641a3e665bc704dd7d13ae750c1c49ea8f411b92285f79fb8241480b8c362e95c9d2f994799114c18277a154197cd97a642a4291576b2f50636d426483f0a6c96c87cd7003ce727a4bb11153f23a9de5f6a072785e4888d34a15c8925d3e60c8a3fb765fe1c332aeb56c3a8134b14f2f6667a96e30bde68b2e0cc34ae02d617a9d6b1ef7299e582d924f5be910bea4784a7890d89a9c15b5b097bdcfffb4d4d55519d5c798ed1caf5be3c909c5372a78f0bdc6191f6c606b2d57dd30044402cb66a03a6d51b16fbe13fd4545dcfda6c64b8854a43e99f62d715c912e8e9fd273e994c7b318de19586e032436ce4c9fb7c41099472363df7eb41a1c1d6ff7cdc4810b14920df86e5563be67a4595f133123c471f26dacf43b1226e5ac9e07a9627822ed86a05f96d91e12b108ccf792177b3c76c1e1c4ac0dc318cb17ab40bd471526355a58a3ec11a01b8960afac2edc56c66f7385ae483465aee276b005af3b207ae27b6d831f97fa43b9f2a90e61ec24cdd95317825c9190e1f5b5c834e113b6c", 0x1000}, {&(0x7f00000040c0)="e75822fbe9625ecd5b69f80d49cd60e8235d92c010578eb7194457c781b1d62f2ae5ca9a6224a7807e190e6927dc1e2faed1a8c5d6b26d8b73865aaceaf82faf11b78ffc6b2754a9869b237c61c4b299c0e968ad59841fe510aab91ae4c6cbbfa6f367b44621c88ec59b3b04b19eee7e3ba632f0acff67971f0c3761869724dd890c3f24ae0426f3c25bada8400f155265a2425a7a077d13925e73aebc28654087b2e49b8e9a22abd89f732d9b4c412b617a8251aa403a8f5c46f917854ad2a2644fb50ff543e2c8f45d3c4080a27e7eaf60fea40eb05d5a338dd92f", 0xdc}, {&(0x7f00000041c0)="b4a8a8bc5480413ee82f9e03e7c3c1d0adae77ea1e9425f4638c7a43bcf1e6242e43c43ef0942360f22a1bf554e0963ac78edd5e59cf0fab9dd7c700c18921ce9dcf4b014dd22596b161f8a4c1cfa836bbb8b7196cf7eb497b668aa34d208cdb3b63d9bcfe897fc5b89f667cc12681db3a8e79696fc89fd6567f9d44fa2d51374b7e43c28ca217152acf572f60a3f14f7581014f71302320c4db13f08ca10e177c50069cf02c3d9cbb09c79c423ee7291ba48fbe14a281056b33ef1117abe7e431686083812fcf654cedfe", 0xcb}, {&(0x7f00000042c0)="066c4fe0ea89fa64a1830ba9ffdd0c79177ecf0cf6016284edad1e5857c33a2019bd08df5c3af7b1b719e145d0a54df2cac25e04", 0x34}, {&(0x7f0000004300)="8b89f5f34b669f16768879a584ec59dec3db0b1e63c6c685169c2bd5a61b08cc49a5b1c3dcaef1f6845e04407903a0707c469d15fdff8c48b7", 0x39}, {&(0x7f0000004340)="0650f75f11651aa660c58c6d278159288253915b3bf991960c54aaab6ff65f0fecb237bea89616d962844df02ab6be53612df2ca1de4db4906b1a1e9e6765dc292b17452d2276d31c078aa8ec78e395aa81004d2e331cf2f6f261fb0127a217a5a4b9cb195d7a39fdee97c1869c03ec64ef8cb73a39ac810426473863a6d973ad2338b293cf62b8b06566b73dc5d80a51bd10ddd064eb476f7a718a7f2ffc61342489fd01b2d2f48a6e4048eaa6896ba763fe7ae62ec84cd2b8fb94ad245f746141e80e9fb", 0xc5}], 0x8, &(0x7f0000004500)=[@rights={{0x28, 0x1, 0x1, [r0, r0, r2, r0, r2, r14]}}, @cred={{0x1c, 0x1, 0x2, {r15, r5, 0xee00}}}, @rights={{0x14, 0x1, 0x1, [r16]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee01}}}], 0x80, 0x20004080}}, {{&(0x7f0000004580)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000056c0)=[{&(0x7f0000004600)="1b4117e9ccd919767ab836f65396ae33c1e0385605bcdad46d5b9815d5d9caf2d266d3ea1c74533b89cc9ac60309e585c6e7a0507edc190e49f6a5a5466b306404714a92f6b8d7334d7715ea0148e52d0301fa01dd2d24d4", 0x58}, {&(0x7f0000004680)="e6491f54ab4f517320840bf4ebd4a671dd885bbf0358c6528ae3a6dc6a8f25961218e0a8f2a4d30ede01b01698de4ce1e5c00a4d74050b2468d4f4dd2c340ade1775239cd304174ab769780be54949b7891d9b955b0cceba38b78523d95e5c73a2f676aac2207481aea5bc5b80e02077baee1064ce925b5c2da9bd909a393f044c60b61efe55cd8baf75906be93fbfbbb9b0c9f4a37f9da79a878e583eb5677e850d50e147de0e213419df215611f65ca95eb5ff58c6e5a6d675f5f94ddfb79a2eb912be6c2d5a6cf01a00eae0e51e304b99785f7b135e3569dcf41b5f679c81f06a52cac7e30cb66ea0dc5ea68f819d407713f3e0f85e58ba8996be080224395584e32ea9416edb78d142bb7c3c766cec59bb4e5c1c711a374cc5f1a51868f4f4ba535790fbde0b7d44141a0c94a0cda32c2f1c82cd58e9db4d4c01359d73ca269481e6ed056c7b7a1166c6384de2cf9cdb08797b0d922be283b907e5c63418ae4baafb73fc04a11a44eeec0580bc6910866ffffb66072afc4bcb9c0008009913c0c23b66654d760a369c6667eeeb0ed8b839f2a142cfb5203160afc068090b5906496e8db3f5ba90e525533c62025ab6026c443390155f7b06eb8eefc006af7701a1dc49b65e15c22e47f4be9a5b98bed4cd8be19bfcab9aa1c6999577aa23fbedc0d011c00ac8e95e08ba5c707450205edaa8ddb0ef2b582994dd9eea08a88d576a6b0f210f472494a20c38c0391a8cec3316f08a7ac4e134ac1628d7d48b41c624ae20c2ef5c37aca0017a71b32ff8fd9fd81e779b2e895a5e2263023a46140e5c9421f318ebe094247f4eadb35ecb6f7aafeaefff6d59a7353ac173d2c818bc2fb4c9cd81b3b0ed77a4db8c3711a0090aad543a4f8e85c121587c2b17ce81f12e61a3cf74e52bad62637b4ce9e55a1e0d3640728ab54bfe14c99a71cc0284fb540df6b88c871e983c45f9091f9329f59ca47a5b6d2d721509feb198d5a4512dad2f65a812c30438626b6e0f38c09e8766706fa5128cbe80bee12efe9fa8f874f2cce1ae27ac202b0dafd9553364d3c5e26450af67a54f78b64670c4eef95e703f2187ecd22344384aeaef9447b1064357c231f4c8e6bc1a8df0ebef5a68d49faa9f444af76adee21f8cfb46082025ba25a0c27ef859382bc8006817f3a3994bc867bc2b2e0acf66b904b2e49f7bbe27023c8c7f4bd27476522c8272e609e3ae9d23f058ca783f9b63b27737e0bf22d30a8bd03ae8b80075de1dbd44812ccbf3619ba93345ccd399db2ded1cf9bafbb00eb02efcb8503377af9701ee16107adc04fa3cadee04c6b2c54c167c4169dd8c75d01b89e46e0ad79f220b2d21ceded3f10fa17efd887260020e7e6d01e62ac02ca9956cd20f0fcfd462746ea94c8061c3cdd6bc1c1cc25ae625d0190d77603267930eb06447de81dacd18e2618d6a1c3370174569f88a74ccf3a9721ea5fccab573b29d4b4d2a2fc639abf544beb3a81fa1798f098a14030f4c8a4787193ca12b8b3b40f5910ba57bb404a669040ee58b3660cc47e1b740e62470dd67130273f4a517ab568733644457efba8d2dd32892115e9bc6706617828bedbc8d73f516c58eae6a1dacadebb18b8a4bc3228546822b4ebf971f64a077598e71737a296bcc7f124561cf2731bc3f2f28630dc957452fa346e898e558f7fe45b54de17ff48befea3c1b47acfced86c0e94559250d403e9d9c3c945c827c7baf433fc1d7e9814649f00c824993b97e3a963c83d85921d0780a52fdbd72334faf5d53036a8d9704406071681bd72381979d1c9172b43f5e4b9b3782df1bc6d3d638768a8307c1adcc8573dc0e93727d8c7f176fd2a9f91a1bc56150858d9b3449e3b5e34519198e468c1bc80bd3e53d4a29ef808e2e6e0c3a62ed16bc19545409380a3dc9f04adbe6c871965a6e8756e85ed6c58441957b92e9cf9abf7d4e33ec366c93702bebc3062782838af4076e7a48246112a0b509d8a5d4fc8ecc78add2926ce8dcaae744a94618d04a2ca1df7082150cc5f1d288aa414f3136e1356e23cfa32234cb94d1aeee195c0e37857a712bbe30901d187c47b4b45ae30fae64bea0a8672d3099117985543c0a3b9f3abf306f0ab539842a798fbddf1b25c1a69b33513834e6094fcd315a86756c09343ec56ecedf6f87d4b12e64381ae1233afed806fa1ee7855ae7300a425a7f3dd200f806545d7416622bc90de53c10a2f1eb8e66422fff528e87eed5701f5af00fa2f230354409e6ef4b62a9de22f2f8b65657a050b2f5c17dbc9ea9cb9796e66d376ac57559fc61622050c0fc0b266130c90703d29039b54819dd36faca85b7fa2930f41f3c90cc65530fb19686994091b5026161033d2cf14ace36e0a10a817c533c967ace67e14810f128d010305111bcd895177fb997e93301c882957a382429b1fc774bc7ddf36d745e807391c37236fe9fb44656abc6e664616146949a39b081c32daeaf5be7d010a5b9dd81942eac33aa997681b99770db5fe411d864f1cfff401a907915ad96b5ca3fde6c6b4fe98ca7edc26b5e7110cc7611350767e42343322b075f2622f624c566c4d7fbec163fbade3ddfbb6052384c5679f58fdc8728113e9b50cfca98d9ec92e252aa9f6df6f69ac2dc0d264fbccac4b54673905372e91305f1530d18f42189e65e69bda6adf5487c1aaa2de7932cce8f7ea4edafe076d67c8a54d00d7a2dbc8765b900dddf133a39561f12f16fd7bd696cd68abba582ef8f7249665337e5f3f9f65f8bb45d79606b1942c7108288833f98873970939d1a58153529d3039b8700e4ae245b95456da2368be31aeff7e9747aff3dbb5b60b155fc5f846bbb61ef2ebe416dcb1c6a9fd896f4052ee9e0f0e67e6105b0c13f708a480964d8bf0c95c5464688c0732175e5244bc39baf11d97763e7dc7ba8bc380adbe7a626c4275920e46c2684ccab605ca0709fa84b61cfa16312ba1d864a6f18c20a34f00f18687c3b9c2b0fd9604624a5a09cfcf35f61feec87e964dbc4d60a4ab0be55b4f5e4a5d9817269596c08fb8a4c3b3e23367c0923316ab5ac0252bd937cd5f7f47968df3270907b1622a6d6ce19dbbfe2ea4d1677755cf30e1458431f456a749a4dfaaf62222de25cf534e551ef0605fd027d8ae06497f57ad3d35ccc7b6318d17e93aaa0c9504174662e530e5d4e070afd660db7ded8554fb581fe6e9cdef2cc867e82b50cb43f1cbc64ee81edaa52a29c02faaf3b95aa62963907b9de04f657baf12037901edd46d8bf2922323d65a142a210fe6b7039bdea6c3dc318b838fa30856a3ad61ddb1a5d038eb6a1ce0a88266deebb1d1ca847fa5910b78841b04e043f513692beedc81fc0f1d1597a104ff42940e02c6b2b50f3f0bf3a523652c6333ece693c70750cc8864168bcc08bcbc52aec0dbda5b59c85746821ad425f0389fa53041048aa708b403278f3cf0e434bf35e6f0ef8b88d54bdf7a599b2b5ab74cd9fc116e252eaa8d734504695273f5f034123f375a0cc70fa25385f08fa2687e9dcf39e0a44eac24fed06fa62f744a00e124e8e7931b84ae3b8a8de6bc834c685cbc7b8bcb76989eed71d808d4eca8aefeb593fc85ac20c4bd737e841e2379696f9593aeff223c301524a81ef877edaf98aa87100b27b6031d45f4c054100294286bf23c6057e2a8da3c8a633a830e4a4ff697891783bcbb2ffee3a0d896b77542b3028e93812e7d2f004bcb62b903eebbeb23b427f2d38aa09fb35323a1de4ba708769ecde4cc5d9150afeeee6639faaa2b546b7183144d362290ee385b224e60f5834427160fce114d8c0558ec1d32b5e66bf2ecef4e0a3023dade4df26abc72efb9a1ed119d37962974295a5150c73ba27aa7bbd17c0b322e50f9b4a1c2405dcc2955cce0dd46b8572749d2b30a3900d0d5e53b7473d80e26adcf15671ad7f4082df0a8f535e21b2d2839e4db381ddfd4124accaccc1cd0e1d88f5279bc53a520e8622244ff6f19dc63ded3ee5b520765df03069c5f61fcecdad99c00512849a53f8b27518a9b43f32ad53103f6636fa65977816bb0029fb3fa799002eaf66590c1eb7ae34a0ba33ec988cb0266e720367c5d24cce4bee8cce08b20de5825f224ed1198858ed62cd3c399ad47cbff5b8dbf7a3aac9b9029c2ec336e916e80ae479d321d17535d28904f7196bab1f13a12c74dc896bba0c989d1a71a493002461f024adf6eeb5d2755cb7bd33d34feca85b66c66ec40993b3b5ce955cd853dbddf6971f88869e021bdea59a671047101299674bf854813c8d32b5d3a0b52bf7951d5508c4428770f8d63aeb35a2650cef4e7a9898e31156c74b843949c281219f1423769673a4f61e55b3af35869f065d458326ef954a1c2afccf3814c5ce91141467351b2c15e426c82fa5de611324d328cc296ac0fc625c1e9d7688d6c432feb2214fe2a0aad8467ccdf5638a1d9ac099d57979bb8077f11b2d7bc446bc67cb3d57b8f60fd09d380ebd8f31e7df0c07934f8394e98ac2c119417a9bfdbe46b4dca2fc21d4565ccb42616b67cdbbe8745c0b4c5d9c2834c1c3128185e3ff4b43ca15acef713eaa1055534602e27ebed79d33a4db2f99bfcdaa8ff1da188ebf7312474512349d1f9db1a3b5f8fb83f69024adb051cc5a002b2ac12a4d8cf6566947dbd9acf53596154e88439473e08a5d554c24fda4e68aada6ba1911e16994078a6d4bb16b5c73ed09e12a4e55363b48498fe18d5726c6db2df1fdcc598c7bc9f3a25233c300cf8e9e9afb68b756db9e72681e7c59acf74f03ba86178be153a6b2325b7dc9e1ae34e0bceb290ab4c274225eeb1ba36db9f5ce72f49129c470a846bcfac870ac1e3e2f096c60796633c4bb6ec07eed2f9e6be6a87e95436c79600046c49c3e9dcaf46f6d52b90234b8ab4c55b9b822fe529c72db6299163315f59214b980fd565b215950c2a7f45502b3089d627b3b5fcb791908612c8441a00da8b7bcfdc89166fb0645636d9a0d479b6ab5f10fcc373a4d61e88ccbe105686f63457f26869fe30423432cd56c9f9883de2365456601cd3330bf2cd79598427a82034312b2765be76563eeed60ce12ef6e50db0fbbfe0f29bc9f3d03a1301432e06917682b5a42a580c8580d146a9e15313e9aa6b5008d0143723bb094afebf9f47b2d6004c9a5ce89eef0287c98d0a2204aca3f8e1fa8f9b81906adab2b53d69799f55fc09a8b0ea87ea77b303022b52c6dab8f138cd71a0d681d312e0a7b482cf189ed93d156669a8c86dfa8673fa12c2ab18498a0fae991e6f5287e35326acd9b375962fa44ce69683fecb7b1f4afca0b8e75017548774fd51409635b20400ac194714bd998da37a939480b9148345c2b4ba5888a53803d81e0602ce0c2233e67d7c66629027dafcc4f5987ca79c57aafa21b4ebb60a716a4c90fd5b8e262998b12d6841112238676c10e55536fe0ae27de72ac71440576563e067b20412437b46210adda9cc1cc7197398698c138e090bbe4c22c866c66ab15e4046bd3e6f10aea90db4b16db898607af911e0c41451978d1468e774b88a200d292109f5a106cfc369b020db2dbe3675be3e6749d6579274c858675b9ffa64b0007acf916587fa16f90e21b75b409d1d58809a9edb04a6de29a8b21592121a312c1fc0df1d488a77a6546d5daa67c10ffdd40a955c9b8b7dbc71bf013405efa69b127593a6bc4d4ff211fb37880bb358e709720dcaeca3f81a0e9447f82b63e7cb584f1497a21c78a43527fc3ed8f3be395e48c0f78f71ff9c4b8c5", 0x1000}, {&(0x7f0000005680)="65658d9cbf0cd93430a072a9dcbaf7d94b", 0x11}], 0x3, &(0x7f0000005700), 0x0, 0x40000}}, {{&(0x7f0000005740)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000006a40)=[{&(0x7f00000057c0)="a704126d4b4ad9c4d814c3bbabd928c32f47e828fda9583d793fee07257d5fb74283a630", 0x24}, {&(0x7f0000005800)="8182839ceeb209564f5191a97f469ab55e9c47d3998f44882c2b48a175dcaf769061a0a4000daaeb7bb835ab2bca5119e6098cf81955067de253a6c72cf68157b45f6aa6c12a54f32556701de45ee8efb9658473029e68deab72ac4059aa48114a1faa919f2e3c9fc7369a0e408ef3ac2abb2b311641efbe187875532a9992e8da2a99d3dd85a81754e88bf83ac1e19eed6399", 0x93}, {&(0x7f00000058c0)="cd38b8d253f650df0b0996e7e4f4261e732ba65f2a3d3c10a9e2b7ba18c74f59abaf5182f309862fde98fd1d16814c8a1868a93215bb212b8f0ce3af19851875ac946c5127cf29c1d56531add7992bbd42e310ac315793aacc2b3af0b9d7e6fa13b0ef29584a65d7caf9b1c543b9b382d5207b05fde34d474e79a3563a66ff873d1c2dd362a4cd06bd5cc4b3efb6b347b623fbd89d8ca7ec33ed7d0b81e1c937c10dfb2992d26df73ce9536ca14d30f9e6c162a47f2d3492a1963cb8ca3e8cb2343f8301a286f96ee5ef9f4a8391b548b8e1aff201238be24c48e3dc304aecaf9a39d66338a8c01c715b080ea2a0ce7cba160b9dfeea0c293fdde9e2d792ea35044986e4cba76eefcab273756d43d990973b1b535e74507100af1c57e4f2e6fba71ba60ca60017e28db9e295fbc11d1912638a189ee41dc78ed801af4570bcb84fed67cb52c834a1e6c49b53463962c740eb150c3b6b4fac7813e2e632c9e2a5de0a6d53ffb01a8039080ecfd546e65ff133a9b4e11dbfbee8305e3f6202c68db606c0edafd99e45209936082ea652ff82436a6423edcff8ee5fa114aa50ce4bb2dbcaffeb0ad7f2c6f7c692cf3d8da260853c4312bbd016507962925ba973ad589d03e64bb0d6e4262318c055b9f09609849114ab21cd9424438be61e1d20fad1e40cf89476ff3b59dcaa83813157969b16f10b0f6472113557659506bbf5d4fa9d6ab2652ae18e81f33d548899bd3e282a9ff602af0556df8317f8e970e2426e42adc793b71e98824acfc7e88320239b6e51401ae68e0b8fb7740dc612020cd72adb006bc87dafb068039a8322f664cfa3d5f3c48c8a153fd428343f25863000f9be69038b4b5bb60efbc6da59e755d1856998903bc1c363f1c22697259f920f42b59dbd684d9e56e64d75c0ae8cdff66038d5a9d0052d520349508e461cfe93ef9592334dafa55d1f0d79add50e102ec388fc577eb4a41d00e88069b537a0534fae958d77929a10023ea54c0f2b48cdd3a301ca99f7245dca79532654402eb23758f3b978278af0e8f77b9dd264d61dccbcaa843b633329d6e82b644a6819a4974084de7c5caefa88c407e336b4db965d41eadf99bd8324beac0792154cf27eab2c417dfb1deb467ece9b899eb89a231d35aeeb1f996be166270011d433ba5c71d28ae9da02006797c616ae948ebd23e2c8d9eb78f2bc8d994ada2d6d58fbad5688a09a337b54c7b349957818b51d6b1b4c1cdd53fc4b0ad265dfb8c1548e7078911098a81e4f7d4686bde8b53eb2f59a5a5482db30db76579592e57c66dbcc5e2b8a9edd72ecf1ed579089f00e5deaa9fcd0fe83fa27b9400540428f7aed1a9b1687ad72feedac2cd0bec6501528d3fd0490b24d357bca6e41a0a7139ce24e8ec152da68df18be94d2a7fa7954c3721d9b7ce1887115ea92fda1cc6b1d17b1e8cca7b45c2e292aef68b868f2edc89ee1da523fae93ac285c1b1fb056e40d799a438d9ab67de4a63d7e02c083ac55770fc0d7634a90412fd164ecb8aa98897e3015c925eb5af3488be9a831696bced14e7de1ee5bec3857cc2cccd1a369895a8d952d04cc0674b0caedd373ebac80254d4d8a50a0895ad8deac32055a5a236019e2dfd4fcec7f96d99a4e63d9e5effdc2b9abf727b4ad07f1534ccfeac116ddd09ac61dff8be04479f219b5dce0367fa252f01728f2b2c0192296b510210e56a812ff5c49deb035b406e0992ae54c728433a7c36f3662de6f5368e366e2246b899bc2f900eb8c6117e4024f34f2934287636a23e095311c08e461ee1677e7b207b0cfc9912ecf34f70bab76da7c4bda39e4232669004c2a136fd26da09241184c7667cad249a3ae6602672edc9fb2e01c9b0d2692dbdbae1f91a55701b25e8ccf0379508ad5802a264d02be1d81b6563ada9a05891e83c931b74ef39982106511d25b5e4a09e6c01cd30f02867fffe69c7874c6c24f2b335dd280e9bfdd2a2c528fca866d8cd0d9a77a36d0bef47e4196c8ceff6e68cee5bdd9ac2396944161cf4ca285f9271c45cd0801e15ea8e5f9a5aa853e10bc18798f5420860dd78e8c0743ed23937dba6609af9aaf6b2e8c29a3c5f0535994e8839b94bfa2eeb956c40394cc50e2beb3a8652c3b1d2118d7bec6dd1d9d29dc61e3f7aaca7fb510bff16e12014896cc871f73986e52fc736191aed8ae52c3b8e70f3a384e3af1c897e7f1c192952557ad78fad1352283c8653e1cdd29b4ff4d1312f007a97acbbeee40a26f757505aa996f81c78b1eabc4247c284be502c29fbf1227c71a6ddb35339e22e28ff633b2497332e92040e16221f13f64120af2f73243dc3c8466e793f8992929122a43a6f668b16093049515553aed2af21516afbba23e9333e0694873ae575feeb95961b15acf660a158f646590deb13bfd72327fe87299d9a54ae802738ed2af5245856de36763c70b328efddbd991938eadee5698123bb5f50a8f8fa60c8399ed3fd82bc11de9f79887deb6187eae8bf3048fc5a5fcb2fc26de38f5a7598cfa1d280a20dbd6105fa660eb8f61445e292a3a498ca1420a4e7e730aca0aa3ad34f73074ee9b151ad179203e3a1b1a988501cb87351ccb62b49c04a734eed9ded6299f8baf2bf6f357de6d5f0175a40c93560a37c6a254e2b5a19693762b2cf815763091d4b71ed23df5ea270f6d834608f8f77fdc5ce20457c0d3227b6d68a72ee45396a69845436a6005997415343750a4051e5515f957b794b6f4a2243d677a7cea96dae1ef405150611f9eae89f16577d8dd2be2111d59a4989fca77623a0f860759f39bbc819c602449ce9eeb029871a270b8cbb6e74c87e95dbb4f6aa417011a7dc7fd13480c74ea88d1256fda8207550ebee89bc90d8e2cc93f86652034851a20d593adfe935bcec4319a52decb0ad94184af2ff7976d7eee7897a53181bda7f4af9650ddbabaac00336dcf94008d8b19f78a30d5bcc03276033eb0e6e967bf7f309b49077fd6b4db66faccf77c0006016fa6d6c1720cee2c517088a554292291573a3ac235b115f8242504da4ad82e17a00eef6d0d2fbd810e012e15db2d50a4f8d62e1826da893860984b915b2e3167a897a0efbe361ff3da1684aaac564f9f8ed6dc85766359da71ca7022693d1baaa901a809ae2d8694bdc19019e8982921505dcc325e3a07b99109db7444027d399ccfd841e884f31de3f8c8fd2ddc1898d96a2b49ebe29c3e97a87a5cb92eb1d51f0590462976bc5d3728845d94db6806688d4fa6df9cf392591e38a01da7b79a09472c87a4dd43d037bf6d0131108f4725eb12f603cafb7bca512938bf2eabfd8dc1d206d2c4493ca640f069814066b7db1473276d439ee98f81ed3cdaedf685242d4d452d5e0faa83bf86c614a40a26dc46b73567d2f87e8e95ceb2d08f73af3399458841a8bbdf02c8d239d17b4d1db3b628c1e62e6df89d2999b316a7b585c0c439e3b42b82bd273a7621bbfbbae0929f8af7ce2c9620687e231800b78d1641e2b991260cf6d0069ec285989aceef62c8d4e6ff9855950140079912df18067d326224f2315c8327f9b10959979823d2e21342a2ed1d6832173018a232fb6daed2c1e2dea1d63e63d50ff5d9a4c48bb60f45e2978a955b0bc11ae9d56cb993e97aaba84a773bbd09bb5848262500ab36c6c9c9e7c827126f22c20672d3804712dbd35f7eec27b6cb34f5403ad9ac72faa10adf7abc4d1b61161b7270d64e249f76de80b69a0cdd70f4a122c2f50ecc18d12be0a0fc234045d59d52d1eee68ed60b3d4fa3e5c1701633c6506f778422468376ff43eea9ef4cd25cd3c75d9544a7e48be58119a570364d10821ad626e52109cbb9b96906e8b7f432f221c1fc29876960fc04bd58e791f12478422349e5517ca4ef433b041bf50a780d5ec7b26efd5135e216bc1a63973056e3c7628410ddab4b89fe7c6abd1b728b0c378afe40bb73d2d279c1e0f87eb6c2dc6ef2d28a448b573846377b4709d3f77932f44e6e8d5f19f4101d1f531d9bda0382ff8d09f1b6ac0ae35231cdc7e5309c4aa35e9d7a0e08b14ddbe60a0b3720efe38b6b59b23d24be411a2bbf5f4591af6eebafcb788be7335537198443e573af4ed8ddb7b2ee1dbe1ca92d66fca894bfc97a8f18134558d0bee4cc5affb3d6cf6261350b7ae90241199e618bc23f6c867f1a7511e4188a48c38c2a50303d3bdd142d1470591b0bf9281f0234f1238aea70115b4a136e4ff918c9148aa73ebb46972ce8d8703f9fc8cffd3be77c5fa3f407b5d37c7e9605e698ff1b07e4c7434a89abe4662526121c1083b4b78d31e46c8a7ce844d4168effb4803992f892934708f16b91e0456aa0fe0fe24976a1cfc293fc08146cfe1c7e443d0782e5cc59901400e4b0376146188c65b321942435ece29b050b3b53da9f3b40d558c778ee8934e2c7317f8d51aa9e6b4318223b50c322def9192c7320036017598f946f3a6a53337666dbbd6e7fbeeb9324465ecf491b076cc7315e9768355866372154f0bbe4c925c6dfdd0e4098e5f63c7dc2b949a2e3a947e64d48802e41552cc0b880a6602b8c70b2eef8050510a78b664bb0abb46352b4e7379bc29435fd722993454cdf7fde741c3f4ef0f53b095a5f45908c357ebabee4ae8f42ad916e521a43b0126c8bd71900f49548d441ee68da409b0ba81fef8b4ab10b9435b5f4ea059db4fe42b162094ea0d04869bb69f1862a84b509041bf3fa40426f5d56f6733a9b1911e6dde15a9011ed60ba4958fd8c5edab857c66fe2026d445cd4a6c94a51ea07fc4f66ace3b787dcd81ba6fef0cb53852614322382a4515fcb6c7f63c2cfe417bfb6f4683da484ac663adc5f63638ce96d5fa2a204c6f4b27f9cdd3fb0c020615248a9105d7dde687aba67252e1c52f03b5e43669fd4e78e5c1b7ef001739eb19cd40e3d5b840b0c899b5900b5725267ef920ff704fc9a224d3c1e95eb70dafaff84911a04a304f5b86bc29bf91bbe520fc66f98f1b9750f0cd4c295090d3614f3ee6a8a35a7e6efa8e10f1d309187ca2f2ce19dbb328a3d58d5fe911ea6151eb038bd8f49a756aa34b337eae60e8f8724109efb75779e99e8117d45103041711ce1f1dd8c339909126d993d73d3dd8f21c6b95960f095f7855fd5513dc5d0a963633c2d9409158954e0895e567b3557927937ab81f9468691595538f6825e9a8e1880c403c80ff15ddb4ec8da639ea7abbbbf09d53cc103f0a3425e452fbc1cfd6d4453d9650c0c3ee54fffd3d5f60d44ed4eb54cb1577ea3ba90bfeda60bdc241d2675e62c26922dc75f97133dc362d47bcd2869cfa50e6ca1f5e89524adbdba230875d2c26a19e038cf43db4940231607f1584c623a6d2b11d2f9bd4d0b7937d8eb3578cfc8b5a7f2cbfe13cbc67fb6d697b17083f0671dea373cdde8112d2a5bb95150aa2a8f43697c89f86a56c1d2138b71778143183f6ab1aa137953db67a4ee1218b5c68e8e231bd8a555c1cea79b9cb0c23ac2dcb7a3dbeb58b783f0fbad9dd35a5585deccc74e3e2b7d60db528253c8774514a9bf0ad47363f42f349a43dc40f921ca8ff04dc347be210714a4923373bf440a7e4fb4d2c202757daf8e899d1999e62922bd3afc3feaa4097474c4dd275547aa69e1e5af9a50bedcf25caa99316cd38e309949b8818a402c31f15740a919378cf1feac2b2166f8558218c9e30bccee7b0b494801584c210d1a1d5cc50c3cc674eab6be778004fcab926cbf184fd52bcbb52792370015103e47d87b3f8a0a5f5d24d126e2cd72b2", 0x1000}, {&(0x7f00000068c0)="c513f3367b8f7a0277ebecae305000bf3a52be4deacd78cc62ff68fad2645546642e90157c76fb830105aac6d1cc62a2d63de0cb8dc4d26350b924e5db6ab5bc1125d258a82a6abdc6c00c7cdf81b21ddd4a5906", 0x54}, {&(0x7f0000006940)="b3ca586de4419c10140f129c6fcef0495df6f4ba81f27c5445383f52b898bab2a06e92d378009324f59670c0a6594d1cebe8f24595e0059f7547e3fe092fdeaf81938c3904bb4d6dd8b1f171306ad0fe4a6a43051686166f6c94561a2bb25a1f8aef82af04e9dbba49711c44712929e9cc4d740585059f013ed7d2b9478ef16caf8ba64787083e0b8861b09f7248d3b3cf05bd4b55fa09ef80bb7564002da96e70b75224a720ed6364fef125b730f18beb30bef60a351451dd22ed7dc2db8afb1bf76080115ced0a23c8b02581b804aeb09ca5dab23bf67a5c38bc7bf34f", 0xde}], 0x5, &(0x7f0000006ac0)=[@rights={{0x10}}], 0x10, 0x4}}, {{&(0x7f0000006b00)=@abs={0x0, 0x0, 0x4e23}, 0x6e, &(0x7f0000007e00)=[{&(0x7f0000006b80)="965548104755f4029ef58e0dac0dd62e9a043075d648305451d73a036e24a0818a1db2e62fb30ea10834d5cee3f9344b57077e2649d6bc46382939ce3ac0924632fa683504bfc7d34bfc19d0004c131b3d538f0daa76d0d2b8b76597a0f02b7968b745148f091fdda7fa45bbb0b249d44f3cf5414b6c4d3641d10bfd7f4d7de5db887ca8789e9e0f777b03b42d8adcf16ca071b0afa7accb91eb917db2cf658905e61cb2b4997ea00e0a32ed669a19ece1388e5d43b82939dd64d352debba5a70a5d", 0xc2}, {&(0x7f0000006c80)="37175124ee5fb82c1ac12eba6c3ee7d3441d50aca21f760a67e9d5ecebf6a2ba6c281ca9d75dcd633a758792a6be051135445610d33719237f7b1916150db81392a7fcc17f6d0eb0e69f47f912e4c212be40e1b0ce458fcde660957c85d2a8413847984f233b904ad25330960694ae6a31b9b967da003871b9371517e4cb50899a8fe02b1493ee8b4be5649edb24d3f2cb80d3502a7eccedde976d2122e66ec175cf3db6fa5376acaf03c8e936a78f7186fdd95f9a769a00e9f29b8de8f7a8e2b9ff20ae416430f315be3e8c34d698946190e4a55780a22a756ddbfa9b7e20324333e5d04d39ff5563c0d7ba70b71bfc", 0xf0}, {&(0x7f0000006d80)="b69420a159faf034189493ade7323ad83dc73f8e9c50d7ea1934862def3dac1cd7298d61f704ec7a75485daa965fea64d8b4e02aff6c17884835bfaa72c02a7471a4e14d4d1eafd987c0502d75c98869fe30c09a7d2170a0d0470798a444519e06f793beebd2d31d0ad49ce0de0a1ec1f30cbc4fb395968517358a8059dddd224b8a1ab6ed828fce23e6f037bbc02b7cb239250474ffc6ecc5ab263715b8e95cd0d66fb57ae00e7bb9e6eb6c29e7cd1ec8df382ad9c40d32725197bdb6f1c92014f1fdc29b4c308de69b1dd7428028060e35695993353549a20b0e407e9e00374ae8109ff0855bf138f569db585b34dbe1be2781ef3e90ab2e8aa929960215ec9d35e4d9f081dd298e008546241cfd847efc50d154a2cf3f710dd6df1a35697baa49c5bfdb1935b27295dcffcb8e4c2397dadebbebd4f5581be7f90d52422a7c50693193fa64f61d6f65e77443f160387ea39147e7670b9112151d93e69e0265ced1839f5523c97bfe1469f7d0bb95758c66b0c0d8ad018d8949bc6f24fb5bca631383c7b79d6b83710ed744ec82fa87f9a9587df162092275e2eb016c00173c4fe9524884ef18dc02d3633acabdb4aadb3de51f9b1a28088d9ea8fe2a69c97eed0e6f05113a5a89b744fad9f5fe8a184450824795ee7cc7947b4a8708fb3990e00e6dda13ca4a4c2e6e5c98db0d7f2b4cded79cebd6e9facc5dc95ca6d92f85065f8a155f86acb2cd15f86a9e6640cbe6f1ac9089f349c660368ff6cd22f55c1ef59650bb463e5c2768b6b9981a5618bd0e79b8d49957c2e6492e70605c72f27a177c093159aff0f499072c508628fac9896986f7ba3bb08d2681e493207517cd9f45eb42d79c2325a6c305c520773f2d0e5f1fef2c0de2af854159357cd86d9f19c8a7c5c52876d6e7f352ce0f60337f433e8234cf25b3b830b84a0517f66b2bdaf6a69ba03c90e2c50f61caf56a156d1083d069f590e9230a6c077e1b5bbba16efbd0f26f3733dc2a7e8709492bbd8ba95e229ba7bf71be41c554b6d320d5a35e0770c925bca0ad84b306483477ebc90513e1dad9b710f855a67251c4695002bb93527cf952a3e0b8c9f24473f7f8653f8a64d96cb81f0465d2af5c688755bde4b447ddb98130627d0dd176f5280492874e2a86f190e16836a934589952a000cdad6a3e308a90b7502487cb953853b0a2996b47c204de7a9dc39e2f6672c630ae635908d4e4c9b1126ca50090a4f4131d71998d12170d31fc85de3a2f9c2188f2cd3b3924df80788b840de781a070477ed5b37cba1d435dd0f44595f98bb09db1d738e72aec9519a0a5fc3b664f750fbd22dd0da6324fa60058f500cc5a38e7f0afd9f333718efd8d023cc31491e822512fd47ef3fcc17c9899a28264148ed00f7015bf062b30809686b1d6b05b1a7ffbdcf8d2d3473182998620066070c7d41c4d98d1f2b6bae35b8f7030066f458de7db2e6ca501df4189b7f45c0ef2e406d4449b9f33f57064fe3ec7ade681cba53f2b4935598115acad0516cdb0860813ff6162f5d13dbe4fc4cf8442d7c49e7ed86a9940282a57ee80adca60fa1b3eb33bb7f8f2f474d1f3730862046f3158b8dc3eec4d529b44ff40093ac8f6edc3e9aa6e2de0cc6f3f75e7b51ef890475a3002aaa2d590e7c5c9a7fce56959731a88c487060c2bab12dd65fc45cbbdf4543a0638cef222db7abb9a84381db17d1d1554a865b5a9ca57f66a061e16780c8746470f9e06001620133fe637b99b813101eaf9f499a66ad3a86feb68b1e39794d20e91a4e4e5bae36139e18bc9cc1c86f575a178990f632af792fb50157175ddd858edb68d61dcdac44f630e8c8cc210b389b8ce9da4d8de0abf11c03e87cb7d8e64f7e0435ea5b4b22952f49bcdbc242aadb0f56ef2fa2f9a5fb34e4b459c4f0e26e13c3cade13513f5372049ac28c75a112ae79292cb0b10b872f13e57f3817e3ea09164aa137836c62264ca8bc4ef3c6c231ca6ce82fec20d128030f1fb7f4a38d4f41e70add7751e1c0eb74a314728de6ce9d360219e16ee2ec87ba1ad24bd2130cec9100a26b0a0032cab3e8424029d6040a3d7ec8c8c284085c9c2a5913a86ee3678ccb7fb547b48404db399947a138248c3e5818eac984dceb9590bc9afd79e75afb7e9c8a39db580623c943545a94678cfd07ad38306361cf119412b431753ae1fcb5f0a2b4f87fb1d70a95cce8831e0eafeefda67c68c27221d74e562e35e36c29a665477a462942552159b8cfe7ab7e66e8f264b38db29a5f6e83b1f06a583c8b6247cc9207fbaaebb0d706c5bbfb32ff267c591a3666727b8f1b5ae3540978c94720c95c7acceff303cf955c2d93fc45a1a7777748ed04b4c576a2f68b5597627e32fabc723e030b6fa0e7fac9bcac65c2ffd9c45a35ee57ce8af209277ad11fdb80f1ec3b98ec396e06160751922c33de2fe99b803089cdd973c01418aad923f1beccd9662087cb7858f3ba8b4cf02e70b2ce2099a41d1940dca9725b658d9aaaa4a35a97db6272c74948d36deaba15ac9c72da0e3813e5691ce3b7e78399bab4a356d083f53cd6f365680ec7d0958e9235574316e9e732af8efa4dd2637e1001f031a4efe45c3a44e76d05bd7f8dd2c6d0fc4630c1477bde7e8a91ea35e9ca151f15f2cc4a4f5c27dddb1c07f3e922ce01b0fc471942b2399e6c959124c140bcc2c15f5b1f2f3e9c9a53f6704de3c64737ccca688ea56cbaeca5d4da21c9e6a0fcca0ca69299e07708217df5e244b70577d8e5f156e25cc06083561e8d777a79243f3d365ec0a0e4343c14a790a40dd09c8c791b5f8e6b6175e9f00067eccd6b619c4bc1dce9460ccd4a8255d307c153dddf9df1692fa7737d0c8109dd23fafbf814bfe18518c02dc35a8030e89b8fa49c1396371f858d55131b5bd05a7bbab80796831cc8ec0c37b936c9cbe1b108737d75dad8e0916730070dd6f22f5d50fe477d0bd72bb510dbf72032625ea1d638ecc9b671cb79ecf6dfe4d6ed186afc7f511ad7ae46f510bf22c92ef97732f1774d5f89549b6bd37eec1f5046f9e09f1370c6b96bf8e05d480686efb3afc4b04e634f3a11cd41067ce2b889e9ec678a3b8eb202f3aae3d6c663ac9d704898b2d905a057bf30f1f86428f88bd6227da6fed209e0e19c42a333d7263f41417c9de6d24e6593b8ca5fbc4f27802e2059fb3ea9f8f7460b52823af6d217940cbd9e74e2e8e538c681aeccbfc3c5e7ed3585f44e9a51dbf4748f485b77c63da632f6fad5bc6ef5dfdcd75b2fcc820b12087700a6884f2f2348f3936b2a7cfdfc8e59eaa1a326b3a686c57ff7cd5f64a515a2b74046ec0b60ac09b033c19913e24dde13cc7816bfe43505ed75697ba7ffe95c7aec475191e44e41363fa342c1973e1798e22b190fb23c5e2da2eab0a29321e7a4de2de38e91625f38e6ead720dc8872d47ab3dd93433b9fdc4244595331cfc6b070082f1e7df785fde9e8760967bf2e6b27d4f2d43b26b255410596d36a12add1fef87c920e8d19d63f35007fbf554bb8bfce40fa345e7c41907cc76402922960a8219974c2e766ea003597ce01f691330b746185fb680d7ed0dbac495de35e8b3eef8f87c2a708a6ee3db9b83ca44afddd348bb8ce5144efdba18643b520bc1b2a02f2cdb670c11b87447412af1fe9138d2f93aebd8eaa797d6b2ad4dac9f9cb76ed65039cfc9392ba8a13227c66698096de48552b082f86bcb4db7647ed49c6387c699bc8a462c948059a64337f9de89e728e6127619354418b83ad4d51f62f3f6547e1fbce6407647cb692639ba5652aa906c41ff8a25646cdecf68e3d86ba41868af948640786fc84b6b7f507f8b37a2679fa886aee65d2b0c58e64542e7f75ff7339c94be5e52e7fff0e066b89fb21785d07fdd20c05e58f8c0a605e78bc7f693a49ea819682c870207728c2d08cafd24366b5c96449929a1f74359c222ff51e681c69d22e711b6e5bf4e210cc643b96aeb079cb0b21af7a13ec35d3931d07be1627fb61852c161ae2d7b043078e598fa6ab98162d00330762a47caa4e830c8465aef13fb9a14b94ff3792793edca46c163b953bb9535680720067f25ab81ae8aa3cc9ba507afe82f271a93ed81f7297723978afff8295ca10bc6caa804460dda1a1ba1af50e8ebe79703b974c97c26618fe78b13e6bd4b3e15ba80652e5ac06b4677029b8e627efe1cf345c41433268367f3e5e418ce0782ca6914182e4fadbb86e48c9912e1925826cbc8bfd10aa9f0f3566277988dfdf12822fee10582904f6ee6a4d1ae2e0bcb06882d322e3705cfc7d857a5de8c721c36246ef734094dee0c16e4bc74e320893873353f749f4e7f293432c277f76a241b6078fe2bb6ca828ad4d93914a66bfe9d331719c42eb58b84a7667b31598ed7ef569e9a3bfbff30469830ef03d35dad5749b1ddd5a077e0a6116febe9a7cf7de5d1af0f47dcc23e3e8d29dd297911038aea4a2f037eca691789e1eaa2c2c74c6a792655a0222656528ba39bb76c709eccfcc55be1d64048c5f3054847279b993543a2562c3753769a5eb1c53e8f5f7e65db21196e9b3f45bef7f9fa0659af35629bc8e1279cfabf40f23911799928adc50263f3a1157794cdfcfec2de2a99cc2cfceb2ab2c75f280faf3da3ede058661762dc5433b012fd916d71ca351a0ddb33ea6099d746a0b5644d19cf15a6d8dd9c86a4f92fd771ec9792c5cb19a0f079707db61d37d497811bbef290145ef75b04811818b0423b9a0268f599986c9f2ba5a21aa3e1bbc159292cc1cfa8b95c964a7c84a7653920359ca32f74892c7554f0b00bd3dc2686f9a7433356fda08c8ecb06dd75530047ae45db3edcf8201a4fc7e8247cff7f602de9cda133ac0471f5c86faa9f3a381c9010d4ed739fe8e29be596a1ae8caa1d9b37d9ef9f5d00778e125cd7902072f9b72fee1947faa6977f84e501cc9a1d26e312599c01aaef89d750f6884ad951131d57caf1323ad63a337490aa673aaff091af6fac4eaaca1230cfe670b4a7d829ee109bc22460391e6d2198c70baff41b58759308ff2db41829cf78af75ccfdc907a94932319e1c29a247b19fd249dd92f7d1232bd7acabf81645383d8728e0c58ad87e368b1d124f1f7d03621e9a4191a981a739a3c8a337a6b4d2d548dae68a893181d29f5b1ad592bab974b8b1b2282c4c8060a358aab6adc826f080732ba5b02dd906ab4e67c4a7ef4404082c1d83d2831f155282631fb782df01e9aab40ae5d44f49d38d3adb614c670ae69a3969632940c3f147fc963ae49674764e048bc3aca7c4d8472c7e5c5fc16c1e6de2b0384558cc43c00caa226ace14044b65873e6cf4443fc4723cf451774b296cffa866a8fa66a99eab74062f39ec8c3d7bf7b585b5b8f47645634774723c9cecfb4fc44dc16000b3c0b7834843b136a6ee3ad9506f8e0f59dcea7bc4975ad611c7bd09f9e25baccbc6c9efa5aaa440e8c07160f4cfe60a02457824133063ef4e8809d53a1621236b757967825a5896cead3f2317f3726fbed57f6b91156bad8f7b9cefd74a51e74634f771195fb2a48f229fbfbe114b79444823262b7e4427550badc16f1f691b0945c2c829075db54b67b8853a78b8adbf228fc45a9d82e2a46b0e3412951ef18cead026ddd00f6f1fba6f050759b54c1ebd5186eaf069509dd5fd3207e0fec179d3bb4a633d37bf1eb591f6138334c464db5b710a8e4b1484ac92700c0d28f04335670563fc1686628890b83a966e05928147599dd1e778365b162e8f745515f4edd3", 0x1000}, {&(0x7f0000007d80)="025c0ee5b27eb267095ba74b3f3b694a104c54cdd14c33ec7038fee3b43b724de1332824e514d4eb15968206fa3d7f5011dff0d85220d0a01e90d41476470b83e50bc878771970fc3a825c", 0x4b}], 0x4, &(0x7f0000007e80)=[@cred={{0x1c, 0x1, 0x2, {r4, r5}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee00}}}, @rights={{0x24, 0x1, 0x1, [r1, r17, r2, r1, r18]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5}}}, @rights={{0x1c, 0x1, 0x1, [r19, r1, r0]}}], 0xa8, 0x8004}}, {{0x0, 0x0, &(0x7f000000a000)=[{&(0x7f0000007f40)="15efc9b2550993b1a0fb0036c17affc7f0acb0e40feb6ac0accea9ea3bb729f22013e8800b532b7eb102dccc59370cc4487b247d8b22def4fc9f34ca8b77eeb82c03f9cc2e3aa157c922c12f7bd55c0c3a86aff67f499cdbfa4287653f23fe3076de6d480715b756c803366783e30225e621de0e167f1c0e184f1c91ea833d343973dca5ed0e1c949df26f3f6178", 0x8e}, {&(0x7f0000008000)="02599d24d4f1543923a8f7e2494f77e08c86e276c40624596bcfad95537947f27efb0a18d35d36ae2d26d1cacf8bc286113c754cb05d7046ee581be96cf7b34f599b4e385ef995a5a9fcda67178455dc4f238901797222d4befa6da6ed5409a2b2779b4f66b9c622e30970a14a027ceb88b6eb4c7261a1ac314b07af4bc82ab981748570f782fa0b0921a16a996e04f6a6ffd3339e163a79a2c22745b1614717d718599f25fb2cb553817655c89fc57dd0d897e68b9d440ae3843f9c3b4cce7e3b67e46faaf5d6d592d6f858173c1c7bce902e5c3a9e23cc3ee832797c81aeb049a47e0c3b780d972f369d1b967557385dcc6bdb3bae02dc25c8cf70123d15d5c0ebcf02c2118ff6c59f1150a0c6e1102cc5ecbb91b4d0271bfdb7eb2fb1dc4af06f41a46d84469ac867913de1bcd3e4099d838daa71e70bdfe8e20d55c87c69384dc38b2ecaffc82f531d92efee548e021b0e8c61b5c63fa23513aca31dd963fbad443f31ead140dcc63a68ec387ac55bb9fd8948315249a9a058b01e85a6c1d9bd04c4aacf13ee4069a1f3d9c583acbb51dadade58c90720c1cc03b4d0385df4409c0a74fb54a02ceba88928c80e9bf541be02f44112efdebc8285214caafb43d77391751637dfcf9f2a9947c962be47edbdee035c95d1d5c9987257df2dc82cb39f19c6fa4bf2752824eaa26fa1c299c9b0d237ba290f74fc8fd5275abed35cda588eba30aba211f053465eca6918cf8d1909a41adbb2b7f41ae6adac0068c0234535586dae12af803ddaea6028bab83876565bc34c49d12fea754006a2f4c317b402aab989372771d2664a8a01ef46dd3bd2a8c367f1376fb495f54d4156e8a040e37eb5efeae899175cfc1bfd66eadd24084655433c108a5c8de2342bf9c166d2d1e67fe07f707b6a273ce10bb7fd4f09b5211a1602465c520eeb8c14dc47098a39d473fd7304fa28e48809ac87817e41167d97cf139a999195de316ac112909025bbbb63f5f84f5dcc879d3832b12f24e6f317c2692ad74e76617183e432cd5eb470fd679ff3773432e15bd6d4c8ffe73d2f0074eeb49d6dd64be98778daf269d5cd636f7c91ea515874edd88eabf4c1e32cfe3b8f27508c8f932d48a580b3fdc4ff7aef6d88b04c3223c928cff8ee7f7fd47b39ecfbe1faba5c9910715f3744b2ce685b7c806f4a81eb973f00c12ef3f0b1ac96aefff7e1e5a35f5fc2da16e0fde338b92d9048bf555cd65e6e1cd45bd1974edc0ae89f84470ed865edd52b81918c8cf00ba3a2c83d49b519349744f47c42cd8bbf685288c0efb84b4305c1da387f21e2455563ef41fc3f1b17f85696ebef9bb39fa917a7cc01616da835c6be73004fd8abd9d4027ece59f7c67a1019d9570e040c9551007c512e5e12f0b6f3ceb5070503d2cccb283ce51e3f350ca83c0c3a058860c07be80219b9cb14640a5c3984a1a0d95c27969e5a857cba99f55e17e0534e39136f72a51ee81b661f8637f963dd0a2e52ead459f6eb90933ea10020954cc5372122a412509254829707467a1f6daf141fcc9c29fd5baa3b0f0a7c6cf878cf78104472dca96a9cdb19bdae205d4fa79f533d40af0c782f50f4208417beea7121ab2f6d025c3e1e926209b43f46618b8b257a6fa736931e264b82dd55d9d6c23b8e170844ecaadc4c4b479e36cb05b2303ab8d24af1a89121419483fbe6121ab50af9e48f3e9603b7383914a0de869b38c074a4e7ffc7578559ce04c8438be8363561e5b3f953b82c4ce10059378d748800783c5e5de55c61fa6b15e9cb094b5f24dbb6b550c16549bf457cb00aae098389a404f2afa03b49567b88f309973b3a6bf206c7081b980c4abeb0197ad54b9e97dff610de519a4b8b057a4ebccc08e27dc84c45ddda1e5df1f212fe8834403c6b5304ffe9245a5469d53647ef0d214d82f51786b176369ca5e5d8e0dc7409b9c969b85fb43816919cfce9ac5bec76f82be3d16df8b1f9e24db81fff072c72216ede52362a2c370c11e1dd2f79325c6d85dfd608b8a8f2556bfd5145b0555631860f02a083ec6b0b3f3bf84bcc7a04a9121f5c32aab482a7023e78bbb414a362755f55bbedadb3d7130ad1637de86e3abd08827d4ded9d8307ea5b90b94f2e8afe73ab9acc2eeafc21c923cee4f6f5c710c068f4a9a0eb1e3c32ea36cb115eaf4f66d8f6d067babce0aa341763aa57202a55a6656057d60107ca7bd7cbb5710f3ceac160becf95a8262416be1d876b9d9e1430246ff041503f5aba8153bf9fd5ada50431631c0f2a062eff4d03119500b854561db8b439bf5fc50ebbe88fe77e751893720b2d30cf9103fcb8332bae1c283c0ed748d09f3617d3f54d8a661ac8e3ff3cc759884537844ed9b44ccc64944412f0a2c87faece13585cfd9c28477b392109d46462e3025fe86a49bd7fcd6675f8ebac0b1fc884c8ed5b559b741958e752cf68abd90476c7865637a526364e91e261b1fa0b3f15e1c74917c5097493b52cafb9106b13e805e43f35667b8bbca206b79745a97a3544bfffefd396273080e30b5a9c3ac22cb2bf2ed149e6c669a834f6168dd31ff2f6f54b139dffe31895ce3ca52db8dc2fd00c92bf62b4eba771229dbbb377fce289267d4d4c920f6c15db728d0acb523bbc5443868a9f1fc4b0645a55f3c966cf7082de0afc0c941ea5bc4705bef3fd903ae8b9185323ca52717be09c8da33a7b8e421d768e7fc945fb1e4ced28e21c2a3b9221b11a9348989130451aa177e6362a964c61e98c09cfe74812d9c722d418a34a9f4b0bbe76c97051a8ba4b9bfe88e737c946bec63897d02d57a0e422b06480b29ab742116c453cb9c0c81aa4bc522fe7865d641847a42de4188cde311751f7bff6c24a2177f9738c35676103c17ee65629f08abdac11cca5a63a84b0b4ba713d0acc6ad8403a50dc16b50e1fe2c46b2a7ae3e6ec483bc8fdefe4c440ab32fcc858c9fc0c39a120174c8992cb3be31a2c9c4c0f913ee91a1dbc6ae99d741dabadf461444259202c7ed01b2ef5ea8e6ea78b0f97e94e6e2b885a2bbdae86467aa59a42b4532dbe744b22a75c94f5055ab6c084cdc433073c32db9f6b224c619fa4013bfef91d68592c42bcdd69335b84f6137571b968f63d9562a936b1a1ec6429e44f737e8cc1243d65ca8e00e134deb20587daf23cebe203c3b3e87b6a908ce3f0fd716ce360d3df059f718458c83876ff241e5dda5045798019505fdc40a0089cd229c8ac8bb71d86e00b8c002aa953bc93ea919db01b7d6099c94505206750a0e76f58242109b751b15a4d357da7cd93a5408331278d2045c18f1396aa4bc26220091c3bda1d2aafeaef03558cf945df1dd270a0efcdd0d15df5a5b341c7596103c79a4ee1efd1b5021364d3b8dcab9dc8f76c73d90438d86c56f1229c4e9c79d4f55e52f7a2219fe7fedca07c3bc5b83841986cc28db1f0505f39c3c511ed752339b86ba529daed8e166705d5881cf24ecd0d653243cbba4c882fa5efafc77c9e14917d195a98ea55729d80126f4234e05c079d327e6b4323be8c901c1c98ca58aaa544ae66c39f515722f7f9d3a95a7e7d3bb1df7e20cc2d828dae935b5a2819b766f56b808a932291379606d7ec6a855f329893d4c9f74d03e0a5a374489ec752615b0d02c338ab35c8cdec5830975eaea7f0f44f615836d30bb2b6304184d5b7acc96de04de860664b65c591d11fc63851c068bf36b68d1c07544c4bf007293f4969137b03373360b6249a98f14aecc074ab1a86960b3ab3c81e44f5e77707a038a6c3b51aede78c38b58ec7c97f5fc13e9a1bf492056328a90e9c45948fb4447b398ddada3b831b9b1dddf1ae3984b478a185c2e644d72c8a84b0837d7d3690262e92b113e6e84272be13963bd714886300a8a2b5c3a804f8e3fab70a85e93389c6127da112a0b8b391ead2dd1c971cdeac81e042620b74a383969ed4e458779aa30a0bc1dca997b5e833896fcf9ad3205d3301e233cad43324fa30bb3cdd04ecf4bcd13994854ee51484567ad079b98f4c3e11c6772027a05959d0482d09ca9108c4d35401c9ba2bd3ea2c2f66a23292fc0ee6616739c0438b05cd2da4498ee4dd78864517c21a1d32fe9f0883f2bbdcfa8591dc9d017866d78cea38d8b944e596b5d2d7476687e846c511888db78bd16034e4fadc7dfb23e301502ae9887bfc19b745a47df5677d3608e224b26124e46a075dab373029cfdce53752b1001e0259915777e8583bff43615c204575a2030033ca8a803cda650f0caf0f10cc73db5aa3c9429dc284a4751ec07042c681a04c54c36ef172d97b78a3d8470055508883d55bced1797279790cb39022b166ddcce7d6c50e64d93ca498fb060de42b503ab7cbec38a6675445933d649b98e4a21e7327145d5211df01311c21d27b1efe3db301b117ce431e992c022ac6aa447c7086e49f82835768bf2f4455491e0928fe7d987a84f870564962f238b3037c2678b1427d931aa2c0c0a7473ed0d55316eafd48c988d2ca455330bd8196876134d9e5162a0e1b71399f7607955d4564543e4081cc5b43c860e874d270f021d93c64d8eefc319afcbcca82f8ea4ef1ab8ef8e60a94de97e74e6974d7813e7275ca1ddf2763b19b2159433849d427e5ad49c7b54db8710c7590f3d6d52f6fd10a0df64f36cfa6c56c9341bab34f248741ba2d262efef10d0df7596a5111ccb4cdf20e45bba0a5fd04f72afc8930c6b3d8946f7efabdae50d797503bd4d21b95472bf8dc6d8f943c9f4b4ffd9ec24dcf22aab49b143ab41506a3cad30e39b9ef59ca7f1df664bc8b72c7b53701e7833e45e8867c48c14919393f9c489e14ef1c3e665bff5fcc45582b94507a1e2e15b8ef71e249ead603e2ffb744172117b78965a055fbc4c0e9af72239f2379ea8b611869f8ad3a17f4c8474523cca66e3890d7616225a2e9051b865466b6a277b97f05b7b4ea1e05be4cb72eba6a8a7bdc0acea61389c2a89657febfda54acaf3b83570b7fefb76fc93984671f54c0d68bb2b60d7dabb02a356d2f58914401bfb8eace23c53cf3b3c40d43a29b23df65b1496cacc28ad2652e1adce9dd90bd0312ebb7278465bda4f9e1fc4759861dcea669b7d8c73a087afa89b6e6a4004be49e8f71a1cec21f0a5118e0146a0f6de189cf6f6eff770906ac8e917e46a21b0e8f0978dda0dbecc7dd2f05ecd368c82498e91c11cf5d4256262256ded8974fabf81b6c791e27eb9772adad71288ddeb9c7fa5a0de40a5e482915c101e96038f492d588a25456addb8f2dce0b50a24e4095cfdd89f0dd4dbd9b4b3106746821ef8c41b9d5defba026e27671d0b758ee567a35a57e68f668b11f610c78d601a85b513021b8a45c524a7a609d51ef97d643369cb831f1951ac58b3b7223de27ea1e1bf007f44eff780c6efecd18cd64b4f1e0922633fbb339254ec1c90e25930d2e575dd86492cc1fb0fd1669465eb412fc7587b5d506acc96e3f1a3a407fa25b44b354b37e8e15f27a876f19aabd7a2fbab6a89c67f00591c5d0beab22cf1d5f9168cb8a9d942d76fb13ba85eae998b4e7c35dfc86225b7b91c5e2272fe85c998fd7707e8b90ca1af1a7fdbaa619ac696b41404665558bb67fd312f8fc6ec6502c89e9a50fa8f945125e2fb98cd4f89af1e4eca007c04aba6a6701f484435189d17edab86406389e97d6d48bcd4f127d100ccba814d43dc8e57e0d0ac24ad587030e69a688c753cd6c941f19508ff7195fe4b6d076b4681c3cfaec843c74688261a44f05b1df2bad0e8ed38096ef", 0x1000}, {&(0x7f0000009000)="aff351691af1e87d27f55334ce6fdb6d70e649da1e19b8f77105ef8378b00eea382503ebed5474b507fbd5439a0a634d7fcd9235c5be1b92a7c4825a954079e6f76962302a3f4028e7b7c8aaeebe39e8a485cc62a4e4b855d214614645ae8210dbd4774c276693a516f0a4627ba69bc6d27b8480ece0ff7572c22981b6052a40f88e4946c69845147a5de35fa7659107fab2e2152775960798b426620cdba8d9fbf957a487b2fff2a0efae002ef00729fb5cc940e3ac1198996dedc8ed0c7522bb9bec88c1488ec5b92db363649fbe752b7944c63bc0d53b922c7e3350298f4efd69768a93259bf682562103b1c8ecac356d5924cc6144ac6ac8f8d5dfc3a9183182d9f58e2ce8495abfd9f4ff1174216468c82f65ba18caf398c60a8dbdc655a31664b1d187cbd8597d7ec585200b601a59dda748409eef16a44dc4db704e9ae937fe84cadf83a7cd1ae760817f5f95d7fe5eb12fa24baf26b5d0950ea3af1a7acbb8c815dd0907169c2286d953dcda6db3390a19471790bb7f3602168da2114e88508b2cb54f3f8fad02a924114add94ab0dd75cb813cd839308a0d9d71f47b0961ad31ab93a66c544d6f5806b6d10a58596ecd1fdae12ac26d99798e2d7cb86f7e3ec9c5e3dd67bc19cd624879b4d933a277c52db3936c274d3cc0fce0a00b672de9a51f140a94d8a28ecf9a0561937a2781f84d838f9253be2f6fc4bb9d098d675196e87cbabae5e4d3c27689528f3eae6df2e2dc4bf1de58d29077c2180ed0c7d9d6d2e9f97d8fe5a11e015c0d4bf499e2c049ef35ff873ce0dbd9f80f017ebd09210657d98dc9be1c22d5ed2f3eaefe9deaf78ee127a4eb0d7e517a20f89286878a2182072a479ea9f05ea112000c3a6d6e1ced57a0c061f9a36e83d15ff719f265f8efd1f8c90c3d10d7adb21dee1bd15494848f703120897ea9df84147f7dc38179de1fa844c3ef75e13bfc4cc7cf9bfb7d30f0d7a268b1ca4037b48e51cd1be5265a2c5035089543d34a9aa460b77cfcc9cd3b098a44253dc4fc793dcdc80ed1598efb393a5d37b32a0d504b39888060fe7f9d85d74ed34c8a27a51128f4700d283cad9e8d33c11e60475b42d621758c5802065199b263a489d8b50c3e072a727bdf27decd148f2c8c6f24dfa31d2b85d426931c35731aa3a7063be3e36d43152a09476b61a7680d5a94bb55d700f9df35411e761d471c085d6760375c4e5770b98c149d25443b0e8593906775157cd33bae3140b6b84775ccb40b4366575775dada72ff88acca5ba4149586fdd19e767fd7d14c4d9559f425b15c559dd6588f5cdf01511eb738a164933cf3fa6a80cdb624c08b055e24d9a0e2d399084450e9dc0da26a99cad657081dde9fb1e27eb056af65e453a539fe2df46f7329696a4841bdb635e9b4e6664262d5c3feab4dda6286573479c46ad5252e8869bce6ba3c04a4ebf1d346a6d9f35b21fd7a87b19fd819677ffc26ad5f4bdcffbab104eb3b703579b27b5724bc1e4ea1e8aea5dc58261f3b5f2c09d58c0d3c6a8c0d9cb24aa9eebb6bceb442b4b90899b61b756580a8b9abab2981073cc8f7ae48cd28a7e25bf00bf46af1198291aff69ed5ba313a5cf58df2e5f7d4a2c0ada2613f4efc4fcd251bb12657c658c1ca0fffed4927a16510da0a2c6988a443b3260433f6f294de59659b0a59b83fd144f00b442bb1500e4df6a63c2c358a1b858bcf7a71e4a24fbd5f1819af4e737ab2e2dfd365b27dd723be72dc936e34eae96dc5a9b458679038982ee080b611abd0124d9efacdfeccc13df76673238045d6874531b22cb7b8723eabf8593a5c01117b403948e2277cf57c8e6a8510e711b7d983ca967ce7ea04699c88817a70a9d68160623df1b4f8bb4f04069084f9a8978c6fb0bd7725d7214875bc63bd0f996d87a4049e0db0ca377482baa1f0ccdc5e29b2aa9b728dcd464828deea9f6774143a746a32b4ffe4525195b6de95a35b56ac2496003cbd0984d44b0775246111e58d457e6ad4d07650414134410b7d5f0dbde35d306e41aabc7c3ba19476e3490f08fd153dc637d5ecdf82e2166215a7417c4c499bac7adc22db2bfa672acb1dabfd00e4f7551050710e05d061250b37760649372aeb9d485a13372fa6ff125970606b97383ad84f7174e5c44a84dbd30825b023e77a1addce224cbe9f14981382ca81afc27fe2ec78fd43fa1d22db3ed4ba73948a32be351f50165930dc1904e0479b763946d3874fe8bd31b28f98750959aad9dc764372bee138b8ed06d76c7e5b7b1b0e97d9bc2e55b23a1738d22e11c7448abeca0255a63947d9517bf4180e5ba44a6c2b4aa724c6445f8c39db4473defb89556e25356e14cceceb9ab3e20d4f5aca625852cc44d3f054ecf131ad7c619ad230a57ebf21cc1ff576f5b485e8c3d958b2c106fb4657a5ccb8c13fdaf747d625bd101237d1e14a740ba903b59de052067ad609606a15b113cf19fb68b00d2c4fe14712456c32246d96b9e6d54765e3925edc40ae22544fa0a84b42c57f84a45299dd0921b50c7f3f8c8a6610c30379ebe5378858d4746e78e419202daf5245039474cfa03b9116e338909d9e21b49dd956f819e1e3922ca9b775b6f272b5f14fce90e70c2f7ed94df7c43b789746e8a597db4314a0a657abec7d5cade3c39d3d47b6405bfaca5e0b647339f39961f09d6ce646f7d22493dcea389dbbf182163ab744055f00b14ed5ca0cd3b1c5d441397b0a66d084622b6d71d5434b6e7cf01831982f7404ba905401ec9a9f25207628dfe16fb623b476b1f643c212afab6e64db408e5d437f36806c104554156bb208eabd04748967a81919af045a61b286523327ef8e34a52d2648302be0f6058b8724b1bd7296bd7665690418f4933ac31400a076fd2a6e05d4179b9441de7c99d218c1a2f5f137fe95a19221a79d1f4d15985013ab23f06ed5a3af36dc2aa0374c68b4c540ae3161880843c813947aebc8b4e43064051ab1c398d9df3762de5d397aa6d05881053313c2c397339e477ac01fb2a91136fd54e6b08490df7865735fbd2a92c9493eb46b3e3fb20ebe1d067ac01b1b6657e87dd4a8f34d7df39d085f818adb4537cdf2f1656281ed1b70ae1dafbe7d15c1bd2b18645931b1a6e987fea01b545dd9c6e4b78ce41abc6669ab34f2d24705ef7b8dc83dfa6bd06e43e5da8025e7a31387deb2ca4764b92310e507770ff26289565483b3dd910199bc8cf8aa95f4246234fa1a2779dda84280e7cc3c9370df4ea6b2ed6b1d89f1b86c9375e507abcd54c02abea547768b99a4a256e23af984d73bbbfa3eb547bf97ef15c7eb91b07dc0e5143def9f17f6e946f028838afa3182d29f0831ff191f79780d4a95059262b844bb6a1edb9f5ebc0720e70683c88554f044bf9e38e45a3708235007a81b8d85956648ffe0a23021bd7907c47ccade5a6011fd0bcbddd3b34be66d72ea0d0ef88492355a6ff5b023f4dcd6e06886fab4010beac590db1933f41645f8f2c32dcc29590785c0f725854416cfb4e0674f1827d2592ac6e7b42e5707181b39da193a83486800cba66d62848e9d1e9574f864525d98cfadf93a8d744a6011dde076d4b4b3d668f7290f7e99c4ce90da0a08eef2990c06c21902c8c8b9c7bdac59a483b568a9313d3039ad26d16eb66364ef3c8b5086deea9b8bf74bbd8d13f6af82ec8f54d5a9330fa9721e4148efdb2ff522fe1b5c7e1b701dfc64de14da40fd90eef947f80c4ae3ac4a4c9c0deb0bf2ccc0ce18c2f5a32d7ddb55d949b8b979c9cd5395f651cfd07179841dbbb01b534f7df46b09f836bcdce068f9d1cc99cd790fcbe82a30d3ef624eaf04ab41a34e0219ba582b0f29a7588e712e577dc0b1080ff724debe6cd47950aec435f892040a1b2100199c16a06f9b08a3d3fa57a9b07ddc930098dcc595c124d1c73336a02201b867f16b4e39a9fa6ad40017f9eaacb66753b30b418b04ab2b91fa2e8ef3b223ffcd282f9ab8cfe9068874d9e527d3cfd3aa15f0117e7f7c73fdbca9f4b23910118d96a39d0d9e8d82f15184ae1688771b4e2ade4393ceacec43249d066938e9311318a34a44a965ff16ebdee3ff822a1d0832d64bc3c29ed4311ab85b203c4791e6a315964c9c18ddb2711596d9579ceb7cce1a6c2eb7127dbccb74a23e5105895dc4aba98f78608f4c095d7089142482225aa9002da623e0918d5cc6cb404baec41a1959e368ea5be999a0e4060646d5146706ae88635fbf6834db4656160854b6c30792c71b81840ecbde042f99af816a0ffa4e56c22d04107840b52604109fd99edc66b088d1b78244b1303598a693edb261510e6d9ed731b468aee49ebdc2422953d65abf08b1618aecf1bccfdb8d0752f916c03e140fd64d554134400ba22cc4ecd7d59cd6739c18f6741deccc68899e6249fa60131c201ff6f9967aead4a41dc8c77e1132e74c55028d8d3dedad077c0fcc49c3333abda7c5e8ecde84b3290f04ef5e9c8b58abeeb701ffb3f1a04b17d045dfe9b2ea3c78e48a8615e2ed725ba8b5f81602d00ac0b214352ad06a5ef707ab52f196a1d83bd47d75ac123fee9076191e1a1b01f57bce4a57aa3983a6d064fd062da57659d4a5a3a0d6c46fa81a1a341c53b938f910fe55101369ae4b0882e804309faf4fbee721439d8e1a9b5236b71b473b16d5dd020e3e35dff8ba1a62cbc90f37e652d3f7ff325f42fecefcec08a99ceecd2eaf8529b6f495d2910086285fbb6ddce76a2cf0fbc7e0eb96936999865b36b7d3c0950ef1b21c9eb2b5a1a27d96c1a84acf4f96a61fa52469e4ecabe3e4dae71c4809d1a65ce24e97bf5f43dc194cc4d17ae4063fa13daac53716b89c32c9cbaf979cfdccbca22ed91920735ecb10890545d3694b1fc57f46eeb4ede46d7ca44bc126fb02fd212108409f3c8494f2560497d72b5d1e6be37302415c54b8f0ad9db71fd3ff9fc91a72b4ec49ebbcfe4c1cbdc6ee869f4c1026cc76b1991f0b57f77f7b7efb4db2412dc28648fb2ed165f47dc32ca615acb762e6a2025074bfdba89d3f7045782fb0653462545d3f27dc942293dd5e79d49412d3645624a35b218ff3c80df589246765d59b646c094c56f209f5db3fcf494c95a8934d33d13f6ca8c57e4c8aab594aeb80d1c06ba35f57302485cb130241f63d9f2c623b4f4e53fcfbff852f5cb392e6cb45eb0982cc0cab84e6fe70541f3467eb85aa02e4dabe5fce6ca0c61b3e44fdce4dd6f6a042a01904d8c13dd85edef3b5f0f610253eac920bb9f15607784f1da928bf5df7799cc432babd75fdd1d53ed1180d5dc78c4ebb171554693abb736f72de8b1f87f3be5553affca27452f22977a43ad677159a2ad3bfb6a2784f87380712469b4c0947d0ebf9f184de3b249cc6787e6efe0edc36b691f857868ffbbe7a91f503f71688078d0aa52c57b81e50eb5d707e68312d48627cfb8787fd8b7477c06a83ca29b4f625a778d4b9e8dc717009a7c825324567e05f05fe996d84452880a77dd3707f2b8fe6c1da18b5e4ad1237fd07a6ca35243232b3f0e30b0c00905ae5616335a137f4eb31d4725b728fc2d3ef77f84ffabc20aeade30626b034ee13e55560b1b2187cbf6878a6f6c4d86757d4223d682670ae03da735a671184a5686e9cd1dd796bd19250010b3bd233fa451ca428ee95d4ce04511a583c69849e24e426aa9c5db99f1d550596e580aa763b6acde9539345e2516ca954a33c1b59fb0305f6f01aeffe4cfad667f78675ecf4f6b417f04508f7d70cfea60f33753c030d136144037", 0x1000}], 0x3}}], 0x8, 0x11) getuid() 00:28:46 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x400000000000000) [ 1726.288465][ T5607] virtio-fs: tag <(null)> not found 00:28:46 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1a000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:46 executing program 1: r0 = fsopen(0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:46 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1b000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:46 executing program 3: r0 = fsopen(0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:46 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000180)='./binderfs/binder-control\x00', 0x0, 0x0) ioctl$BINDER_CTL_ADD(r0, 0xc1086201, &(0x7f00000001c0)={'binder0\x00'}) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) getpid() getpid() getuid() getuid() socket$unix(0x1, 0x2, 0x0) 00:28:46 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x500000000000000) 00:28:46 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1c000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:46 executing program 4: r0 = fsopen(0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:46 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000180)='./binderfs/binder-control\x00', 0x0, 0x0) ioctl$BINDER_CTL_ADD(r0, 0xc1086201, &(0x7f00000001c0)={'binder0\x00'}) (async) socket$unix(0x1, 0x5, 0x0) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) getpid() getpid() (async) getuid() (async) getuid() (async) socket$unix(0x1, 0x2, 0x0) 00:28:46 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x9416) getpid() getpid() getuid() getuid() 00:28:46 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x1d000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:46 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r1 = socket$unix(0x1, 0x5, 0x0) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = getpid() r4 = getpid() r5 = getuid() r6 = getpid() socketpair$unix(0x1, 0x2, 0x0, &(0x7f0000000000)={0xffffffffffffffff}) sendmsg$unix(r7, &(0x7f0000000980)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000880)=[@cred={{0x1c, 0x1, 0x2, {r6, 0xee01, 0xee01}}}, @rights={{0x14, 0x1, 0x1, [0xffffffffffffffff]}}], 0x38}, 0x0) r8 = fsmount(0xffffffffffffffff, 0x1, 0x0) r9 = accept$alg(0xffffffffffffffff, 0x0, 0x0) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000001c00)={{0x1, 0xffffffffffffffff}, &(0x7f0000001b80), &(0x7f0000001bc0)='%+9llu \x00'}, 0x20) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r15 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r16 = getuid() socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r18 = geteuid() r19 = getpid() r20 = syz_open_dev$tty1(0xc, 0x4, 0x1) ioctl$TIOCGSID(r8, 0x5429, &(0x7f0000001ec0)=0x0) sendmmsg$unix(r12, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000001f00)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b285b7401db87df3cabe102c6b1ca9722b2a487695cfb319901bbeb68f3aa8032d19e0a6676712fb853c8376c2e00bf82d513e9811ac95b1017861353c2916802da46d8166d1e59f88e2d503785f7e5c1d762ff8b24358fc236", 0x146}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000001dc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r14, r11, 0xffffffffffffffff, r11, r15]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r16, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r19, r5}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r17, r18}}}, @cred={{0x1c, 0x1, 0x2, {r21, 0xee01, 0xee01}}}], 0xc8, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @rights={{0x20, 0x1, 0x1, [r13, r9, r20, r1]}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24048090}}], 0x6, 0x0) sendmmsg$unix(r1, &(0x7f0000001d00)=[{{&(0x7f0000000000)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000440)=[{&(0x7f0000000080)="7b694cd5a7796b94a03c1f1e255ba37b3878184985af66926ecb54094a0b107a3bdeb35da756b66c1bcb49e12ad79d54e9f60358a2abbbc74b5398c94451a3827c8f158ec2c264fe37033dbe35ef84d221b00896c982eb38f35082396f20c500b384493822699ff93025eb7052fae9c42fe3da4b487c3b54cfac07b4e148d3960c49fafff1a99523b19d226d975dc8ae67ea1b2fa17ef14038e97e7aba277880bb04316df218ac5f37bda5ad02b5a31c4a582c6f30", 0xb5}, {&(0x7f0000000140)="970e1cb97ed177617f2e575d1ed4523648bab970f5b0cfc3ee822b5b6a617002c93f2182a06768ff4fd6d4cbbb0d4b397192b5", 0x33}, {&(0x7f0000000180)="fdf7a39c2718adb3c68e7aaa3d54b3141d2ff652b09a72253b7fa77dab097e4e96fa92b2703f8da6e588362b9a6aef49da73f5a41b6a8da8dba6a09e7c14134fb967916e38c6ddf60ad4b3408371de571695d46df5af1bad96", 0x59}, {&(0x7f0000000200)="be9c18268ff9d888ab37158db8eb88a7708b45aa6cc7f6f8326fa1aabf926d9ac19b84c5a7afced05ce0f63b8f9f0ae4e30eb0f64bb407bdcc379829c24cfa5272a0fdc3971600b50fd699e27fe67d20f62c14fb14748dfd234f8ec695bcd6831c3f2f5ad7a8e02a06df6cd6e859c5b7350c0cd2040e5eca5d7cffae41b61ef802530d3e014897946e8f9dfcf344a27f5ae94ff21f3aa338cd704ddf7214512062ceb773bbfcd61a1bb27bb01d", 0xad}, {&(0x7f00000002c0)="15fa8e6bef968f95bf395fc8fdf1f7ea5ba7442fa46b9990051bcd24f68d385c8ab2b9f1fda7bd6ca401053daaa928d3bd4cfbe7b22d36ecd901e70ea1ee83f80f590b62b7b3f359ba9216dd24158540dc97f491cb1521186115f93b6b7d6cbcd984289bdb0a62255d52b080225927a8c595e6e826a43f0e41b71f083c2d5ec2a40ba69540e28d822e0336aca7aa6b53c50de54e4b35aa0a166fe353f00d9f931cb7f1a14514a51544d198", 0xab}, {&(0x7f0000000380)="e9bbde2b1c99ce35249593edd179fe890c59bd8c60f55b8a2c296328490bd0197fb1288806f69633f13816a5f83a02106325ade39be1d50e9292945f383320a4e1c4ecf24617a08625be686eee3e1fabfed09a1df52ff4760008870dcb271055428b5918af31883aefbb05f00b46fb95ff13a4da78bb912399f1ed24f01f5b3e47f2cc7f6388b4eb2ae9c70e63754073e84a565eb363e602040bb8950e324812813908c8b309595a501b726746dc6c9ec0c3958fac11", 0xb6}], 0x6, 0x0, 0x0, 0x8104}}, {{&(0x7f00000004c0)=@abs={0x1, 0x0, 0x4e20}, 0x6e, &(0x7f0000000740)=[{&(0x7f0000000540)="480cf940815beaeec5a465fbe31fa39fb9311c79ede8bb5e2f34dd809b1e9c504d5a5f51e9695d594623d68860c42d57053598910c6c23cdf4415207a42a5b880a5964d8abdafd62bfd87e52aa9df1be6faecfbb87fd63d6c6aaa633f7d5c3a1682d42b7199ec5d8f101ed5a8b2d84490ef73483faf52f1b10ca77cdd7d4a107e7f4e4b481d50c789866eab440e6044c38049e4cfc198d0167b119790429b807808a20e364fdbea2db1c7182c35823f75915eed68a92934a180dd26e60bc1d1fbd33462cb7", 0xc5}, {&(0x7f0000000640)="339634a39fe9f01054d13043ce63881c6ad9bf507c5b526ded81f931c2f36f84cb7f49cf6752da62d56a29a440322c420e0e9b91ef9208235b60de01e932d5dc34e21a1214f02b0b585fe91d56069cc686b9363111043805b5b36d87f65aa594e258cfb98afe44f9a3fedb81ab101222ef2d0bfc2da0d26f0f253dd4be6a2ec8d99bd692e931ac8a8c40060b83172dc370a526c98855", 0x96}, {&(0x7f0000000700)="5302bfc827907f8db74cec479ceb0872012b236e0359a4040d9abb90b0c31736e071330d074eeca770c0d35704", 0x2d}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r6, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0x0, r5}}}], 0x40, 0x50}}, {{&(0x7f00000007c0)=@abs={0x1, 0x0, 0x4e20}, 0x6e, &(0x7f0000001b00)=[{&(0x7f00000008c0)="a28eb50582abce56e9ef7abbabd4bdf9766f57082f235296d89bacbee99bb081bce7a045541c3f16f382b80b4cef2f600e1e0b4dde2e58e91f72126cabac0af6a1bd428b5c611c391b62c4c4d3a491c5abcf2f5d44be852045657d", 0x5b}, {&(0x7f0000000940)="9721b30d078b164061ae297f287c8a5b3ea03c3386216fc4214a92b491eba9210670a23fcaf8d443d4b8388815505047ff37d6461d1c90b785e7f64b41acfdae303c68014fdc59aa68be47c84407c8d11f2d81ea1e46965aa551f39acf22615a5389e4bca03f4381ff97a3cd488545b9703a919aa28fbfe933ca128969df2fe406f26a1755fd0edf4271efd201941cfca877a1b1d5f4304d38c7efe11e20086b562e7c8d53962c64a017d74ca8afce01549b8fd2f174b559aaaf9f5f4caf611a24abea70e3c6c52b0c291be875c42377956cce95b26b8e28863b34c10d499a1060164a291b3a9b2d7aecea4a320d4c56dd86a9614601569ca83495ee79e8b1d5add1c53f34ec3fdfff67840968e9efce3ea5c19dc013b39b73cd614aa12a38260f63654b4c9f70577b55b774cf07924533c47df0970dfcdcd6e9f8429b4b029a47fe5e23be4e1d9686314b401549f5e832a89b54f2c8bd6d8f0249896e9d583022b5d2dbac1c232740558a0dd7700eac8d3c19a36875f216a347a7b93d1494d01f50ce7abdf1d656a5e8d6e7655ff2e547c94112042bbc93e60485077a897aba2d6aafa13f38e17b76f135137c1e483af25b1bda76ac33b2d94455eb686c17e0aa2d96716413865b17a1722f0e4cb60cef5591110b65f9d387c7a18f466ac484d67751c593f71256972625b79f94d7efff54a3028d06c02cc408ceac0da6ecf0b17f155f67c8ac25f392230fa925bc6fcf24401d4764109275d80330c752859f4ffa8d62e3152ff3b167501955ee1968861d8270982acdc9409ac7c9e17ab49d3bd582c07cf46f1915f14cbc076eef9460f3ce8e81020da3b6cb97c931ffe28d09e6b5a2df09081f1cb8827105f8f896d15fe959fd1be04471d2091bf88192151c2f7be49b3db8c7cfdbd0bd0a8e5b9408f597eb964da95b3e9c341793630d2321b0ca90ee5dd625512dcdca0e435a99e7af54cee8d69386cc6a8c0535de799cc4d617bc1aba2e324a7d119e70ad26871b3cc7380023fdba262a2adb0f3d30bdfe5a6de60e19453b447699a22363676cfc772827b9c3444b1d860744d8a7002e1d7ce2f698ba6a60439b54765cdca8418c405106d31c2157d592f170be0b3fe11aba9523e30d901cbb91c178f5ebc91b0c1e903fcd723eb62ee6285e63efff7a7fd1479dab335848b52be4fad72dd09124e436c8c2d391184eabe0c5a5b9439d652632037bc0e20c0a3bd400b220ce2133cb8476b693a57ced7d02249b3e1f08fa37aa35f96d735b9a2846d8228bc6c6d99a2013f298198fb1c71e6b080d29a3936619553ab4e32f8d8adc4961894b40cbbc04bc59accb65289ab73b6767747c4eba8a01ea0bfdd3dc56776476448a06d8de7a80310cd1d17af2364fb2057e45d8e9b6842ad67aa3a2fd118f3a022277863f6c7112cbac2b40ca7d142b9e1d94d6ab19685f9a3ebb3c539a846320e2f28a395658a3cb6d43a59987bde373bd536e22b2b61a5a7839f947c2ffbc74172e37accba189dce017e45c2008e3ba6fa0925de32437e0c8d5dc549dc7f33bdfffce22d3d4a46b5f1c5084d7f2a1c41e821139449d916ff884aa71ba8aea622812c2a5437a678b5358e50d6c9c5238c01e1bc02d678be68656e33d975812b4a6a6267ca84260c736b3f661f22753d729019b5c5b9c0762b22f81fd8425d9f8a1899661f59a5847579afb24530e4844165427815fe3c662cbe71eaa0cbc3460fd45db3813881dcb506fa6e57270e5085aa63653c216beead4a0ad4aaf41dd0e9a863b6fd54dd2d331b93fd615a5f24825442750d05bcfe9f0fe590d948c57863705432b79013ef2210b25c958d062dcbca6c9a2f57013d1feb7c0b737f86b0afbd88f355bb2d918149e632323d13ded321688a593dabbce17258e7eff446b4b266c18ba859de77b2be65ce498756ef36afd7bb1196b8e0ecca3b8a5fa34ce4a05aac3634017b9b2c6ce169ce17995294343cc58c92687f7a80ee9a68662871c53cc3cacfffadac99cb43c0c9cfb47122f25776d1a5376b80fecdfda5cc95524684ee6d1241d6f997a00da61c329df1c9607afc864705b3a5462d6d024062e16654a1fabf716776bdeb86520cea7fb5368f87c866d0cf59026eb97edbe95fcabce34415666b4afd7798568cbb5b84e06b12ded55c41002b603bf431c1f762e3e89f1012c0d29079b4f18981048fcf3cb0386bfce2c640d3cc6fd629d2dcf15ece395844322277de4a379a9f706eafde94fd3c246ba868854bbb69b75b35ed36064b855f516a5a567fa58ef0c922be289dffd173122e53726376839f668fb9650e6f436d20abd37fc52713d4062dea78c9d74cd75195b7131ecbf16d89b9eeead587bef98f95302680f8c6b0aaf6a1cf7bce62a20d037dba29f7908dad48f362711fe484a42bebed6d55f70b7cdfd8d73a67a882a6fb68c681d257295ebae8b072f59adf243ccfcae923a70426c773eef34810a4d1383fcfdc2c2f6fd78e15ca1f98f7424c9ced32c441c4d8915116c26ad69909cf26e166107fb83ecf18fb0f090478fe2f01d575951ffe3962a68ba4592b0726110d38bd3826dc2a5cd7c5cf111dfbfd356492f5eab6ce2d40d12d76f9219e84759e011528450c35938e3f898708b4398938c5aba0efb8db01aa6a6f9a48c092650e68f956653658498eab21a8e2329913361d759325c3192c7cad853f7af962b431d8f36b313297538cde8a2642caa2827f37385d050e339aad88b20c3e1208372b4211a8d9d0a6a4e1ab54363d5a0afbeb8b20688fc5da1504283a0248daf62b6d1cd8a7a8d046beb9b47931d6adec8412968533633c1530e81d8cf4bfcb17619cac3a4bd7ce0f6a4f39492bb712d0a48ed4d99b2fecf92f8c0f38bc771fd8f773f86e21923f7a9536a35d574933bb4b962079162e15c4715d0ab46f83f10f4534b57acbb27cf089c597e1f8fb25b6a77873f6a8ce88d914cc8d0c5ea8d652cdf4cd956494259a1690abea6a3d46f4ffc08b88bd8ed1b791187eaa9c55ee008a895435e795117ad2748e430eaa6b817505881deaeff63ad808cb6b5a14ef255f6409d9e5694f95f61dcb6bbe87cf2b852c00232b05d0b2f5f36b73876277265afb0482f3c6bc597977e38f0d819d63729171c1b727ca99d9631a395417b06c394f5deea6ec2f2c59de6936a955a5fb91feecd20fcb7cd674a9228e407eadff724b35b55134d0c5df2fc2e314a3b94aac91c33d23128868f61c0eb5178e592fe41238fec12bad9a3222fe7058910bc3f4e36165b4289dac5a6e1dddc4ffcbdab6a46ca25b7d47748f6a0d8db9fa53254c04ffaeae62c693b80db48b3a19a006fee7384012d83e50672a869c9545fb22efbb3781d7396cc6fb2d5306b10337062be136abf865cd90c08c2777af15541d02500ef945d5de84215e18e77e05d07c8ebdc2b044b71a4302adc7ec7956021499f269e480a39680772394b49d5d1c34c996413868373b7431358ce28c4b9bf05f2096b562aca1cdf3e3e77dd7de23f271db8ad826fd8004287484b32b86a0c3a246c4f1046a055ba8f763a8b192b85a2d7ae258133252d063433358c10b21307eb97392ba5f687b6357d1cf1126bbe065feb37c0703b93d4aa16044e26199550b7f42918e96f5c29e6a6d00b93387272101bc090e947f5c561ee74fb462ec5ffd4d3990159792885cace6f23d7511b4c1f84f4b866c7230e2f146e0425cea5485c3b0c3600fc16ddc66633313a51ee296e829591416b3f1fd9e8aba5fe4be82ac4c6a4cca1e3a4eeddb4086fba024a941c7cacabbb2c66859b94011e2a3cc87e502a46b8ecd802f0f8faa16855fa8d0195787ce88a64dad21230e0ef1d4421c05452efeb749eedc8cbfe029486a5c2679b3567e5a742774328236e628335f98f7f9f965b81d28681cf3dfeed80abbf037ac10aa5a540ffb796463122070ce4d863d9e1d5de3e548d3ceb6bacb38217c2b833c776976a8e1f356e6df50be9a2e81d7b07a93d19d92a064dab45ca2ddc294695de98dbdc35f91079106565f3f29161dd3cbd48b7ae3f5889d27ce52625e9b32f29ac4f65d61b81f800ba5f6bfce568525725e0ed5784ce1a65e7e9b8cde66bec281ff1d99f380dae6bbd41f6913c5f85ddd71b1b30e600b7c4faa1bcf653718feb5ca1182b9a8fb590a815ed560be4176793d72567e660e448a2b82aabb6f2d42b743bd2df749962c4f6c8072988b40e96f07e57d38da8439cf0238ea6dfd3bf1276484dabf68a4278f3eac776e7678bc9c44290848c25b69b4fc08afee46baba3e1cc2603b15d4e05934da2f45e29190915df0d4871f0e18f07f4cb597e3d77606449958268e50b143066738809cc2389f240b34acc5dac4cead74a10bfa0cafe76823e5c6072527f64afd767019695a625393096c56f4a042a90eb96c940875e70e276bde2738e16e32ff82d1181ec5bcea3d61c0c989eea12251e725bd19f12c76174c2019bd3698b007e0884db9504118f4d0386dfddb9fff6b369a1582016c71f11e191d475b20a2866bccb270a9144d5ca9ef669ffd967fa3139758fd9cfce5afa3396facfe86d715d7207f2fe0b2b7051b0a6dacc93ff48b631970a4aa0d75bbe77ad7553e89a3b0cb86aa6788cb359ec3b88ab93b233f6345c9264143685f1a815202126d4c58b88094e0eb26490bace16fcabffb75a8edba389d7a6019479bae05ba2d0d1ad4b158baed02c62d403375e4d66e79f9ad528da243f34093135b7f673f0cc0fd020d1da905f3726804908bf2b9e713f000f6fdae1c3a03b4b6638e72c0ba16d0c3775c4f9a3215c3dae3703facedf3aca6646bbe6444862f23932f5517b0693f5fcc7068fd7d84950e13b9ec84f471833b8dda84f1aba2ddfedc44713fff0421ff2e845cb7f94dd5e7fcd513a1aa7f197304e4f6ffb50b7e0aaabe51fee69b938d3f616a1c8e7feb0bbb27f02b160b6d2253f44def31b1efd05108abd54468ad5f6d9f3911de822bbf38ca478643a4e8957325f29fa92d42dcc8586cb3e3c3cc7214c3f41aacba70eaa588f760569c98df0363ffb233a269f629a43af5f70f57f62b76357986e9680ec4d48889e3ec1f43bf2f45e6de2a3aa21ae2de990dfdb89624b0f42b008f94bc1b9441b2a8b1bcf5b1fb5463b0219f3a64db91c79a4048da204ce03fbccb51b59afaad80e9aa861d790f301710385a6d3319fe805e20fa03def50870ace3ba2a584ff53fec7f16b400c87a75ad54fcb8e3d9c111acfacc70a1fde3ee5c1e067108389b65d041ad30b88b28652dd04c0d53db0c4c9318b21905e5fd05b561f5bb6c6b77ebed97cde564551e741daae60257db41ad51aebb6512ec42073f82cad4a08966a9b0e4580895f7a4cb133b041cdd51e79ed164476ef13fe87348bd6a9f86a7befca17a336f85ac2cf1a7a9a0209deef8945eed63beb633e32fa2f6597f113401e7cf36f3d35934f6ea13846f23e63fbbedb86d9fc00f35d9bd7f996fc965d37d97c4312d1bdbfde7356e78eeaf4d7da6542ff51a658646ec4f89f93d4c2bd49651b31ff9368458da5a9fd612fd5390a4da9df47f476633ec42ed4d789dd317f06d0a57de5c917148fbc1300705a9cf7ef1f0d97123c1b9ce6de815699d44e693a8259cfbb36e401a2de200c47a0f7b09458ca115249cf8255a7ae1ab3f7d8377fa303fbf1674b45a419be48598dac18338452cf78a65cd528c08a2b909bc3460361ed4ea494f97ddedba1d81a71921b965ade2e3868ce2af9e08f922f0e1e103821dfae75c53f", 0x1000}, {&(0x7f0000001940)="ec6b8155de8a6a59834f772c0ffc8ab6192a55b488e0a0eafbc0fc7a2f41324d182c6aa87cd0946506c8651629246c57c3f43a7ee3db5ff8290f312f601d572dab2763821d287515b90b341efd00883ffef53943834de0ce0c8c42dd976092fe1f3ff56e056da9d611b0614acc9d2513407d4d7d1ec1a88b6ad483e66bc2aca7d17ef9380b1a0fe5ab2ee030d85ac498b55fc1299eb8227b3bddf231f261d8e9f874fde68e6be3df87e772c004e8fd6f943c66f64acfbe2f002ce8e33cd0f3582d54ede597c13d62b2a438e1b8d34f735461eede8cff17fa", 0xd8}, {&(0x7f0000001a40)="842e0430fe1056b9cd37d460d30db5", 0xf}, {&(0x7f0000001a80)="38639fbb08e89efb49799b75b139c01b547dd7e54f676cbee2d969ac003fd42510de4ff84722480d5c4843d662a5bc7df73d65657a50d368a2b6bbd6b2112b3e50b42a7ac577723fe76ca3ab8751c59075d69f0705c10f9c5dc8763da0ff442de752e2035420cefd0df38cb9ef048ee0cc1b03", 0x73}], 0x5, &(0x7f0000001c40)=[@cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}, @rights={{0x24, 0x1, 0x1, [r1, r2, r2, r2, r2]}}, @rights={{0x24, 0x1, 0x1, [r8, r2, r0, r9, r2]}}, @rights={{0x1c, 0x1, 0x1, [r1, r1, r10]}}, @cred={{0x1c, 0x1, 0x2, {r3, r16, 0xee01}}}], 0xb0, 0x8000}}], 0x3, 0x0) getuid() 00:28:46 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x600000000000000) 00:28:46 executing program 0: fsopen(&(0x7f0000000040)='virtiofs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:46 executing program 1: fsopen(&(0x7f0000000000)='pvfs2\x00', 0x0) fspick(0xffffffffffffffff, &(0x7f0000000040)='./file0\x00', 0x0) r0 = fspick(0xffffffffffffff9c, &(0x7f0000000080)='./file0\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:46 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {{}, {}, {}, {0x7, 0x0, 0xb, 0x0, 0x0, 0x0, 0x20000000}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:46 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0xfffffffffffffc3b, &(0x7f0000000140)=[{&(0x7f0000000040)="95046313d0c0e47db7d371ac6326e012e1d52283e9030789741556208756f4554230cf34a3bf83332fe15c59bccbe2dc933562f92d6d8bf59790ac6d1eef8711b7a9fbee65bc0bc2a376e430d9eea1372f99458eccd82521e820518df6970a2157c195ad2aafd81e5db05c05d55c99f0968050c51984afe6c9ffa63fb9544974f9609a8048bcd11459bb6f3e362e587e876c1b14d3f788c4c5ff1df335303f0732070d513765dac4539d93a1cda379ffd03a962bb81670fbe70b57f001f6642ba523bf3e3184ceea2962dbf26d79749758d4"}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r2 = getpid() r3 = getpid() getresuid(&(0x7f0000001c80), &(0x7f0000001cc0)=0x0, &(0x7f0000001d00)) getresuid(&(0x7f0000001d40), &(0x7f0000001d80), &(0x7f0000001dc0)=0x0) sendmmsg$unix(r0, &(0x7f0000001e40)=[{{&(0x7f00000018c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000001940)=[{&(0x7f0000000840)='k', 0x1}], 0x1, &(0x7f0000001980)=[@cred={{0x1c, 0x1, 0x2, {r3, 0xffffffffffffffff}}}], 0x20, 0x240000c0}}, {{0x0, 0x0, &(0x7f0000001c40)=[{&(0x7f00000019c0)="c2b540bd1cb45fbb925e13dea2815c3243fd1c04f8a2db770c658e", 0x1b}, {&(0x7f0000001a00)="2a4271e0e2499f9b574da01efe722b275f6aa239842026ee4247c6abc37c640ff34a007237137ee3b1c1d415c98852126bfe177db2e36de5ddc5c8237f92ffc7b82d9df1d4686ac4301e2c66f62c82868c3367df32f3090f9978f4c965a57ed5c96cd183393a70554d4d3f3bfc832a99587285086486bc1d6fddc8c80520a1722231a4d9de2cb6dd07c854ecb04c1bbb54b4346f564fb1fe024c6aedaae334eb4fc23bdb966f7c6015ac243eb70d3e270a4bc3a6d4061577ebc566f915284bac5087cfe79c4cb724167fe5c0f2e4c51b9c64363fd57ce0ca4d997435", 0xdc}, {&(0x7f0000001b00)}, {&(0x7f0000001b40)="3678f79b195a232cd3851a9d86bd0e885990da42d82cb50f25b2ad758679c41d7dccff8f623321f286c5ec4803172f0aea02389449955b1f894b4b9188d2ab09e1c8d99f14fc542e2c48b259e2d4a86ab12e803b509f5ff581d2f4274f6d5d6a4d1b1592058190da50a82354a3bedd213e96087f7013dd27a15688cfd2e0d823924f502ed5d756ee61de3ff5c96e29ae9d413c827054c3bce850858be4eb6cfe98f7099a7777d7769fcb9ce40503f0d0439d5c5ca857a8673c87282034b01e786f5cf06195fc89d6475f572a3acf01a0b139697849312963311b9db7291ab0edadadf8cb305286cce6f302677f8304659d7bcc7422d47bc33bcc4b43", 0xfc}], 0x4, &(0x7f0000001e00)=[@cred={{0x1c, 0x1, 0x2, {r2, r4, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5}}}], 0x40, 0x40}}], 0x2, 0x20000006) rt_tgsigqueueinfo(r2, r3, 0x0, &(0x7f0000000580)={0x0, 0x0, 0x100}) r6 = geteuid() sendmsg$unix(r0, &(0x7f0000000800)={&(0x7f0000000180)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000780)=[{&(0x7f0000000200)="684a7b92a41ff9d1a43b1acfad814056b0f1e7a80b18cef1e2f23e097836cc430731e80f45fa7b56a2a66ece6a11a9baa4f40bf890fe7670deabf271844e3da9875693b0ee7fe4aff259352634faede7a095cafa1ec1d8492995ce161f5a7224cfdb421914a47af4bbae567f67e3c99d31d6d4d10b5f22ea9123dca1cd46d7853a3261061b8b8661a8e46286d21546d8752a59afa46a6e91e7de0228", 0x9c}, {&(0x7f00000002c0)="0b4dde9fd5dcfc5eddae1e2f2d69a03847f7de24c5fbd04af646a1cda49765b1b0d97e416632cd5fe241", 0x2a}, {&(0x7f0000000300)="8962bf0cdf8ecb6f60a7710d762988f1a283a7e245243198cc0f39f64561e155e919a3ca6c5fdec6908ffaa31a4b00f272e2a6276c0160a686ba6e87bac03c8af50c7ecd75900a530f9273f4439a75d50317d64707788fa51a3e8bde8a3c1d2c", 0x60}, {&(0x7f0000000380)="40914a0a9bee14d1094945ebd233f1284da0955aa73f37c8adffde10df521a7672c488e9d53025d95b1eb493ed142ea8b55a19c98bd450c08151ac90e93b15a294cca270477c28743ad838662af11447f07e43b8aaf38f4925f11b97f2f7edcd2e7a907492d6502f46e4dc1443e32cd112bb52be8e129e97dc1ddd7593d5dd8919935e9649d0ca621be871c1a612997926e92e0ebd0b8a9d221c688aedf5245b49bbfad602fcdff80ebb", 0xaa}, {&(0x7f0000000440)="bd8e371ca6d0244cbf25c0f6bc5df28797e23cfc2cba1e91324d58f6c759bb9cc42750d222eae74be1a463f12cdf11ff0f6617d766733468b3d8f2d4627e13da1e96da2247756144ea3a4b8a653f82802cf319cdab331eb36459700fad74657f08b20f23b529a3df1ae1d9c311faa5bf55b4719642e86a67b1877537d67349785176b6dfbc3f1a8df42af77d963983406e558459a741c2d5a3a342a3738fb5081f0759aac6b5b1b641f87e17e14706d13c9bceece5893b49a96a0027d1e38c1c41349755840d4063", 0xc8}, {&(0x7f00000008c0)="21a01a02483dd11372a7854cdc50eae7c171ecf495ee470258f5d577abee4bbf1eca3ff12889dc37aafdbd00191d2d742f45f9095733cdaa957b1c01f28e8340a0f31de74ddd0366ad46ff80c9b62db7a8d4bf8562f2c1ee89dbcb8ffa8e902a0714f687243cfc6e47b147b0e31c2c9ba993f984a864c016f156a8da673512ba33f279de9c3ddea3bab013e4c1eb5b47c1b3ed302503d4d55ff2ddcbeac39e4fb5718ec94e99ab7c7b8394f0336f1585f30e12ec17a2fbdbde97a38ff0221f0e30f713fe3e84cf442c6fc6922920681e3b170c0e8658459c8bd137e7dd132e82ccdf90bb75a9ca48ebf4d1519fca1302c85ccd1a3ac264e86f8140317b8071170006e2c2df36b8a65bde6de27189ea171ed534071777e6876d8c7111212d106b386d01d62faa90dccefa661df4d3638697573069e23dd1cdb5eeb568f3a96653654adceb004501fb2426f0aed742bec9247c0861cb79af162200048d92674e09d4864a525a407bdb81e8721560c5f1ab1378d59c99beb6225876bf05bb6db94eb7121ee779ccc00ad1eb2cabd6462d2175946d429b2b75fa498ad45fa3e6ac51041fa674c70740d00219106ffd381b94170f2c2a8e76e97e9f666adfd2d8d206b1d32011fc7dbad24252cc0b42c814f471ae0aed1cd55ff18ba1d10b7d636bf3837a9a8c3b79721247bff0ea4c36833e2f404510c327b9edfe1497b06b81a3d613c2028c6fee34fa2b57eba7e202f6a8ed74505a1355ec97639af8fbd69b8068ae585974756cfb402faf4e46b4c46fb4dfc8230af537043aa284ac1c5eed527a67df5e02a57d0bbeb23f2afe115da98e2449d7f92d34dc5780316654c09342df693e7f509f9c8d22bda84de1dd7d4e107bc07adf40a242d3d157601d03ac841075526afe2675f5611f1677d923e9e2e993b01b751b47765685eeed1f395987ce15937bbb1f9f922241480a9e16d19d5526c4a4086689f6c9accda90afd8e6322a80e9650678be5312fe6174dfbfe737c9aab58c876bd94ab28ea2c3341558248e6eb8da72bfa7919a29f8692f6d8ab2da670aabf900f4fd104248adbaeea8638afbd0a94aa6fa435709204ca20800697e525af192f6bca1292cee7dcc90211279f01728fdfe39a19072a2d221b3fb9bb80ba67dfcdbb09871266962c59e3473b946f8398b4c07fef2f73d54ce772ef1ad734f998c40a657a3ff370a67f8275a995070bb72fe5db3835ee14d7ddaaaa60b59d4597c78624faae2eb96e0eeee8867e7d97a98ad17e11ce387455f166321c67a6bc4c3bf842ab50d57e6a828a3d4734709c1cbb1d439a7c759cc561f723ac4a1df54735cfb5b42245d69c0564db72db7016ff48e26577b21ee30acfef3a769b7daa2f0f1ead3da8eeecd2bbb54c2b8eb892d19536c766965596037a6f77f8dd532d29d2a0f5c3a377fe2979a32aaf4c4de572c7d0e051ea15938813c97e9d48dbe8980a94ee439bdb71d7a3f287b86b0537ab97e93e0d655fda29a4def683d5aaee8f653ed525bad835aa8587323c361a47a2593fce50460251e363dd21767cda8047c06dad7e04c62441cd8b4803ddd9c16986f90590d22b9a53a9752a631de5611f0c1b7d3fd22ee9ea735d1258b90f12bb04c3c653aafd746e337da3c84fa1878ce18f0c21957a55d34243e19afbc5a862ca46b6f5eb7ad48fc82b1862910c69ca519774c3cce3ca5c58a4d78deac96e94e4c2f1735c6f26b41953755633555057a28bb537cf123ce7b654ffb0c8057bd1a50580cd531fc70fbc9d9fdfdcaa2eae190b8c4c548475d05d013d8c8b580e4dc964c2360bae1515fa16f85074f1554bfcbd9a8d1f9420911777ec26abde2c5023efc51d2ab2286dd5166e2264c3cdae682b9467f892921b7f2727dc6e6351198ade5877c5406b2c8330b4d9ace0d9d7b2a811d375d1ddffcde916d033866e2ab764a21e10099144c2d4222d298cf3f1d5510c975fb9076455bb9cca6490754003a63584690050052a2c19931320efe0770d4ad8ba4129ce517ab83838d23a3b4d607e3b956ac63c46325eb9cef55d6d6c30c1f6e85c2153113ed5ae699dc398af2ccfaab1e0baa70703048d2b3061361a08bda0a215de35a302daf3d3caffdb29209c846f44672d4ee87066568552b6cae8b1d0a3d18a4a99f9e5d35079c6e9cb5ca609c2f049d6c880ad88aa80a0431d7f19e57ce24a1a7ad589d4c08af19e2d62344cac039eef8151ed88a4560e3c6ed12d6e0c1785e4222142fddaf79562e52f912bbd8e950d8bf620042f9b8c24ba7fa1fcefc46ae4d0ade001269153a1a893fafce8095af1c94c875eac1c1f78bea6a83ef981e03437498d8e3361378ada145e71f639e3ac434b3e7274531fda7fd8ccf05d9fabd68d5c684fad5ac34820455e91db0b248cac73f03412b81afaa8fade8202baac94c88e353b92905fe46771c1942182afcac93ca2f294cabccfb42d6ca2d99bcb2e041a5cbe9c3eec65f9e15c679d178311b247cb29945a57e06a354231ffc518aa7130826c2bffe71cfd80dd9c4b77605375b25b76dfb94317803f1f73ec8c4da17b9f0d9852191584050ff27555a397646eef2cbc2910e2932ae399035fe8ea54448001b68e2028b341ee9f808ed4264281c1011ea708a5ae6e727ce5af325ef20ec983ac5d3ee856897cb6adaa04d965fbc8fd54f37cdd505e43211990089b7168b673333dd96c436884c3d3f122eeb54ba66cf151000fa0b66e7b5604895562d36a27acd6c66da9b8d7509231c8cb2e0742361530873276f13118d363b90bf276c34324bb80e2d25f4b19ef8b94fe6e69ebbc5fc8868135357309204e29e7048666060682090bc076e9b2d5a22aa0f55a1ead6fde8b1e637244ba16a9c7e6d58b6a7f03fcce96a441e871d89cc5025edcd0e95162886d534cbf76c5f86cca72914a60c413a252f34450ead053a73355a47f2d38c9b350308cc2af948b33d83a6749e2008628eb0bb4c6670947c716393db4e7583be0d7f29cf2bf4252191ec03b669551986ef2700a157b3ed6f8092673bc512e2dfd8106d5a262328ff41f533b849dbedfe41a0bb2938fcd1c2ef81cc2553f178a4b15db1595f3d0ea770711d78ea346e8e763f4cf479939e59996e058f46f574863eb27295ad33f569bac9cdce8435ca78328db09eb2e01ddbdaad53efe21f040daf9a178088913706bc935b8f679efc0ff105c4994e5005a32d52a4fc4d97a5aa2e38a9843d90f754e002e52c36d5bcaa1d43fa9d0d69522cd5492487b9f92bb17c93f9f206b7492485acaec120b7ce2c26dd4c50d549848a73794ae333df48fd0720620b0f3c36b1fd53870003dca76b5184a11e82a4dab0f458a813273775bdd095541f7f866cf60718869bf21135823fdd573798f0abab36c947297dc750bdb8672b0298867cb249984c68054a44742818715c818b7d527598e436740c6cfc27845ee4c97efb507823c9dc40ccd38b3ff56cab41681a511b988d3d8ad4b4f8d95d2664c2d496b438ca5a9dca0cd9531b974a4df077d443001a764da3d36d885d3871e2a05be09782e138bb2315fc9ad1d61253cdb7a365f2114dc24e7cbf0d0a0c7cac5ebe989af94a9957771af40c84c9c62e9d60ad8407d087088d89de2fc17c571c1a4394e2dc0ca558ca05a129eea288ab80ff60891b7c3b5a971d30264f7a247ce2ea3c130acbdfab43ac0fe9a90e9f0d28696ee0c152f561e701c5695198b8eccffd31492e4d0f501357046f1824a3eb3cf952cd1be4e585311839cc5764a94ea3f87778525fedfd35cef185dcfcac39bf9dc0eff7a3e04f3b91d5797720e4ac45ca3ebad61c0d72568c7efbacd2c0b5656dfb6cc40bff1851d48ae0fa69b4989f7689ac4d2f4ee682a45ba41fe9fe7e18b0441746aff61fb0acdcb17378e170611977264843c2b01981c1bf390861aadf721ec0295f169cbc9bcaade8932f7180dc6f13d70f8a2806ad0b253c4700e64fed0822090ed9d05cfb9122bef75f736292ca685dc59f0b3e580210bd721b09fc281f6f902fec8777ce1340c3f623113cb8062d1c76fb92e9832d313c041362cce88d5975afc82d32f9562e225282dcae2fbe4ee3a68f2c4793abe2ad5c25b656121a169242c4c9593f0880f30216aa92761de879bab6063cff345099a3b58e982882fae19d4321dc8ebbf8a64f07bb65e65fbe337ea3f42be30d370e184f7741208b23138e3f34b9dc3a5f8100f2304bd253f2754aa7f1298756bace41915d894b11b0758448ccbdd352140b4afc96f495629dbf93080b69f510f4c9f4c562fbee7beccf2c7a46cc867df93387fd2b19cad296f3e32b57598026faa3554e70d9ab2d8aac543d150cbba5e47252bbc3cad69130fc639c93dd265eacca09454feb46d5ca72d41332deabd05f592f54cb55dccfa250c8eb0fad6468f526d649f0c8e01df83e6a393a1fa880273da8a06211557386a40f073f3836f5dae7f871203d5bb3c61f7aaf1cb625c37d6ef4784ff9fb04e61337ab8edefa68e7238bad97aa02ff2954a5154d137d2573cabe20a868c44b8dc6c14886dddc6864f7455c3014c4d7111138e08af5e14c6ecfb7050f648279a96012ee428bbe3b427434ec2f29aeaf6fd75adfd9195343be1f6790454600b3741053818e8bbf593f00b784ac34c5855a31afd243ad4f6e9825be910b3d3aa95208dcaed0c33b822acc044d62890537a6a2f30b7d3bd471215467ededf763476aa9fb4acfc9586fb5ce70d0d4740efde7d0f2834d40a9d7c215ab89b669627d1e0e45943cc4a06b1967bada344d74ace8f1675597ab681241885657dacd3803f507c8923fe71acdcc7e4e9ef04f19878c858ddf543ac969b4d2476bd3b307b4019bae75ed64b84680ef7a0e6468195aa5160fcf235ce24b2ced569cea5fa0201b9040480ed030be6177de20ed80f62db13d8eee8b70240a8ab78a11fa5fc303bac79065381ea1c3e9baf56d5b217e990bfc34ec3781943f2e548aef8e954787192a8d511cd84db3dd61d4e6387dc8fba50153755e0d37b80f8e0862ea928483c8d8d593fe26aa767ca07525e22bc86dbce0380158ad0d8b133bce349944dcf7acf14af42504928f3f62f5c577290923ccfef2b7f8a32b6c6b00bb07d579d6acbd7ae8fded4251aea1c6eb1e8594e55f3c4a889c50c342999e9154e5e40f200b90dc2cceaf27d252c5a6d6c73175fab2a04944ed9efb8d0f302f2a0269d6e05a83e01221bc7a897623ce1797572a98c236d5bee1549945179fa04445544137bbbb6df11f6c97b5d771f3274765da7d986e75d97a57a75067ec1a28404243047cd1b034ce16b2760ef28338ba5f460399233566ad1b5de21f32ee233567463654bc43a7c3bb21ea8b62ba022289819470dbce3fca7489fdc18e9b827d5f0cb956b26649ea532a75acaf54da3887bf69a594f1c96099ef73194623b5748a749571dddfce8a9d16e8400f4827ed6e37e44ccf07c1240d884850aea7033b0b77ec7d4420707133b80c5a7d2285673c9aff30de5d5d3c9b3cb0f9d45057e65bb03e375808f9596f9c456d6592eec77ee936c6be93e7aa0a8398f3a73d84b4be94cd555cab9404a6e75b42cda0f00e7678d4c83b76736f6266525b8bcf47ce00b997b6d08c69ebb604f45e7f5a7a8882d8ee10bd12c398bac4ba076fc85b02049dde62903718096172652cd62ab9288f6bda80f25f9f98f8e01c3ad149ef1c7d570a1524c1bb5411d394630ebcec97ad70a0db7abc02cb109d7d0768b946ec7229b1b0012c163e9ab5385", 0x1000}, {&(0x7f00000005c0)="c700e453a7a844144d56a773500543976c4985ac3246d3556f18f3bb36cc43b9110a5f90da17e02bbf60c215606948a27a682a5bf0d674c58ddfc4b4c25b2839d3ba8d463818e97a5ce7e7b601bd885857ff9c11eb819a3f15b5345b32d9a49764a12242dfb17900a52066d2768a923d7780c5bd0e3c7f9881cae9d12d8669cf30bedc7c89a444f639cb7d253b65388b8c5785d158243247460a1c414f41276edcd62e608c4016928cabef8ae92ffd4406e97b2e48efcd9e172b09d177fae369f8c87ff508ebbd514a6f4ee7", 0xcc}, {&(0x7f00000006c0)="03d12ff7fc65e7114815c1372563008f8d13637062940ead3b30ffd56eb341342f240d8d53c5c396bcb6f52c674ef0a9ff0bcb6efb21035eaa67de5fb1d247971ad0ec302c20c3ab90f6b87e07637838cf3c61aabe3e659d3e2725f615ae64af760f866dc5dc9e515d67727f6acb8b096aca13827169adba716dd67fa1ccac13500ec48622b25f7d881f8f8f4181f60b2ce1d2f5366b3a9e3149b4ad65ef9415fce00beb795fbcad61a9768eebd26a0c11135aff098afeb892d8594f6edb17f6", 0xc0}], 0x8, &(0x7f0000001ec0)=ANY=[@ANYBLOB="18000000000000000100000001000000", @ANYRES32=r1, @ANYRES32=r0, @ANYBLOB="1c000000000000000100000002000000", @ANYRES32=r2, @ANYRES32=r6, @ANYRES32, @ANYBLOB="00d6f556efa167d95440c709008f"], 0x38, 0x20000805}, 0x4005) 00:28:46 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:46 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x9802) getpid() getpid() getuid() getuid() 00:28:46 executing program 4: fsopen(&(0x7f0000000040)='virtiofs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:46 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x700000000000000) 00:28:46 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x2}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:46 executing program 3: fsopen(&(0x7f0000000040)='virtiofs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:46 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0xfffffffffffffc3b, &(0x7f0000000140)=[{&(0x7f0000000040)="95046313d0c0e47db7d371ac6326e012e1d52283e9030789741556208756f4554230cf34a3bf83332fe15c59bccbe2dc933562f92d6d8bf59790ac6d1eef8711b7a9fbee65bc0bc2a376e430d9eea1372f99458eccd82521e820518df6970a2157c195ad2aafd81e5db05c05d55c99f0968050c51984afe6c9ffa63fb9544974f9609a8048bcd11459bb6f3e362e587e876c1b14d3f788c4c5ff1df335303f0732070d513765dac4539d93a1cda379ffd03a962bb81670fbe70b57f001f6642ba523bf3e3184ceea2962dbf26d79749758d4"}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) r2 = getpid() r3 = getpid() getresuid(&(0x7f0000001c80), &(0x7f0000001cc0)=0x0, &(0x7f0000001d00)) getresuid(&(0x7f0000001d40), &(0x7f0000001d80), &(0x7f0000001dc0)=0x0) sendmmsg$unix(r0, &(0x7f0000001e40)=[{{&(0x7f00000018c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000001940)=[{&(0x7f0000000840)='k', 0x1}], 0x1, &(0x7f0000001980)=[@cred={{0x1c, 0x1, 0x2, {r3, 0xffffffffffffffff}}}], 0x20, 0x240000c0}}, {{0x0, 0x0, &(0x7f0000001c40)=[{&(0x7f00000019c0)="c2b540bd1cb45fbb925e13dea2815c3243fd1c04f8a2db770c658e", 0x1b}, {&(0x7f0000001a00)="2a4271e0e2499f9b574da01efe722b275f6aa239842026ee4247c6abc37c640ff34a007237137ee3b1c1d415c98852126bfe177db2e36de5ddc5c8237f92ffc7b82d9df1d4686ac4301e2c66f62c82868c3367df32f3090f9978f4c965a57ed5c96cd183393a70554d4d3f3bfc832a99587285086486bc1d6fddc8c80520a1722231a4d9de2cb6dd07c854ecb04c1bbb54b4346f564fb1fe024c6aedaae334eb4fc23bdb966f7c6015ac243eb70d3e270a4bc3a6d4061577ebc566f915284bac5087cfe79c4cb724167fe5c0f2e4c51b9c64363fd57ce0ca4d997435", 0xdc}, {&(0x7f0000001b00)}, {&(0x7f0000001b40)="3678f79b195a232cd3851a9d86bd0e885990da42d82cb50f25b2ad758679c41d7dccff8f623321f286c5ec4803172f0aea02389449955b1f894b4b9188d2ab09e1c8d99f14fc542e2c48b259e2d4a86ab12e803b509f5ff581d2f4274f6d5d6a4d1b1592058190da50a82354a3bedd213e96087f7013dd27a15688cfd2e0d823924f502ed5d756ee61de3ff5c96e29ae9d413c827054c3bce850858be4eb6cfe98f7099a7777d7769fcb9ce40503f0d0439d5c5ca857a8673c87282034b01e786f5cf06195fc89d6475f572a3acf01a0b139697849312963311b9db7291ab0edadadf8cb305286cce6f302677f8304659d7bcc7422d47bc33bcc4b43", 0xfc}], 0x4, &(0x7f0000001e00)=[@cred={{0x1c, 0x1, 0x2, {r2, r4, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5}}}], 0x40, 0x40}}], 0x2, 0x20000006) rt_tgsigqueueinfo(r2, r3, 0x0, &(0x7f0000000580)={0x0, 0x0, 0x100}) r6 = geteuid() sendmsg$unix(r0, &(0x7f0000000800)={&(0x7f0000000180)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000780)=[{&(0x7f0000000200)="684a7b92a41ff9d1a43b1acfad814056b0f1e7a80b18cef1e2f23e097836cc430731e80f45fa7b56a2a66ece6a11a9baa4f40bf890fe7670deabf271844e3da9875693b0ee7fe4aff259352634faede7a095cafa1ec1d8492995ce161f5a7224cfdb421914a47af4bbae567f67e3c99d31d6d4d10b5f22ea9123dca1cd46d7853a3261061b8b8661a8e46286d21546d8752a59afa46a6e91e7de0228", 0x9c}, {&(0x7f00000002c0)="0b4dde9fd5dcfc5eddae1e2f2d69a03847f7de24c5fbd04af646a1cda49765b1b0d97e416632cd5fe241", 0x2a}, {&(0x7f0000000300)="8962bf0cdf8ecb6f60a7710d762988f1a283a7e245243198cc0f39f64561e155e919a3ca6c5fdec6908ffaa31a4b00f272e2a6276c0160a686ba6e87bac03c8af50c7ecd75900a530f9273f4439a75d50317d64707788fa51a3e8bde8a3c1d2c", 0x60}, {&(0x7f0000000380)="40914a0a9bee14d1094945ebd233f1284da0955aa73f37c8adffde10df521a7672c488e9d53025d95b1eb493ed142ea8b55a19c98bd450c08151ac90e93b15a294cca270477c28743ad838662af11447f07e43b8aaf38f4925f11b97f2f7edcd2e7a907492d6502f46e4dc1443e32cd112bb52be8e129e97dc1ddd7593d5dd8919935e9649d0ca621be871c1a612997926e92e0ebd0b8a9d221c688aedf5245b49bbfad602fcdff80ebb", 0xaa}, {&(0x7f0000000440)="bd8e371ca6d0244cbf25c0f6bc5df28797e23cfc2cba1e91324d58f6c759bb9cc42750d222eae74be1a463f12cdf11ff0f6617d766733468b3d8f2d4627e13da1e96da2247756144ea3a4b8a653f82802cf319cdab331eb36459700fad74657f08b20f23b529a3df1ae1d9c311faa5bf55b4719642e86a67b1877537d67349785176b6dfbc3f1a8df42af77d963983406e558459a741c2d5a3a342a3738fb5081f0759aac6b5b1b641f87e17e14706d13c9bceece5893b49a96a0027d1e38c1c41349755840d4063", 0xc8}, {&(0x7f00000008c0)="21a01a02483dd11372a7854cdc50eae7c171ecf495ee470258f5d577abee4bbf1eca3ff12889dc37aafdbd00191d2d742f45f9095733cdaa957b1c01f28e8340a0f31de74ddd0366ad46ff80c9b62db7a8d4bf8562f2c1ee89dbcb8ffa8e902a0714f687243cfc6e47b147b0e31c2c9ba993f984a864c016f156a8da673512ba33f279de9c3ddea3bab013e4c1eb5b47c1b3ed302503d4d55ff2ddcbeac39e4fb5718ec94e99ab7c7b8394f0336f1585f30e12ec17a2fbdbde97a38ff0221f0e30f713fe3e84cf442c6fc6922920681e3b170c0e8658459c8bd137e7dd132e82ccdf90bb75a9ca48ebf4d1519fca1302c85ccd1a3ac264e86f8140317b8071170006e2c2df36b8a65bde6de27189ea171ed534071777e6876d8c7111212d106b386d01d62faa90dccefa661df4d3638697573069e23dd1cdb5eeb568f3a96653654adceb004501fb2426f0aed742bec9247c0861cb79af162200048d92674e09d4864a525a407bdb81e8721560c5f1ab1378d59c99beb6225876bf05bb6db94eb7121ee779ccc00ad1eb2cabd6462d2175946d429b2b75fa498ad45fa3e6ac51041fa674c70740d00219106ffd381b94170f2c2a8e76e97e9f666adfd2d8d206b1d32011fc7dbad24252cc0b42c814f471ae0aed1cd55ff18ba1d10b7d636bf3837a9a8c3b79721247bff0ea4c36833e2f404510c327b9edfe1497b06b81a3d613c2028c6fee34fa2b57eba7e202f6a8ed74505a1355ec97639af8fbd69b8068ae585974756cfb402faf4e46b4c46fb4dfc8230af537043aa284ac1c5eed527a67df5e02a57d0bbeb23f2afe115da98e2449d7f92d34dc5780316654c09342df693e7f509f9c8d22bda84de1dd7d4e107bc07adf40a242d3d157601d03ac841075526afe2675f5611f1677d923e9e2e993b01b751b47765685eeed1f395987ce15937bbb1f9f922241480a9e16d19d5526c4a4086689f6c9accda90afd8e6322a80e9650678be5312fe6174dfbfe737c9aab58c876bd94ab28ea2c3341558248e6eb8da72bfa7919a29f8692f6d8ab2da670aabf900f4fd104248adbaeea8638afbd0a94aa6fa435709204ca20800697e525af192f6bca1292cee7dcc90211279f01728fdfe39a19072a2d221b3fb9bb80ba67dfcdbb09871266962c59e3473b946f8398b4c07fef2f73d54ce772ef1ad734f998c40a657a3ff370a67f8275a995070bb72fe5db3835ee14d7ddaaaa60b59d4597c78624faae2eb96e0eeee8867e7d97a98ad17e11ce387455f166321c67a6bc4c3bf842ab50d57e6a828a3d4734709c1cbb1d439a7c759cc561f723ac4a1df54735cfb5b42245d69c0564db72db7016ff48e26577b21ee30acfef3a769b7daa2f0f1ead3da8eeecd2bbb54c2b8eb892d19536c766965596037a6f77f8dd532d29d2a0f5c3a377fe2979a32aaf4c4de572c7d0e051ea15938813c97e9d48dbe8980a94ee439bdb71d7a3f287b86b0537ab97e93e0d655fda29a4def683d5aaee8f653ed525bad835aa8587323c361a47a2593fce50460251e363dd21767cda8047c06dad7e04c62441cd8b4803ddd9c16986f90590d22b9a53a9752a631de5611f0c1b7d3fd22ee9ea735d1258b90f12bb04c3c653aafd746e337da3c84fa1878ce18f0c21957a55d34243e19afbc5a862ca46b6f5eb7ad48fc82b1862910c69ca519774c3cce3ca5c58a4d78deac96e94e4c2f1735c6f26b41953755633555057a28bb537cf123ce7b654ffb0c8057bd1a50580cd531fc70fbc9d9fdfdcaa2eae190b8c4c548475d05d013d8c8b580e4dc964c2360bae1515fa16f85074f1554bfcbd9a8d1f9420911777ec26abde2c5023efc51d2ab2286dd5166e2264c3cdae682b9467f892921b7f2727dc6e6351198ade5877c5406b2c8330b4d9ace0d9d7b2a811d375d1ddffcde916d033866e2ab764a21e10099144c2d4222d298cf3f1d5510c975fb9076455bb9cca6490754003a63584690050052a2c19931320efe0770d4ad8ba4129ce517ab83838d23a3b4d607e3b956ac63c46325eb9cef55d6d6c30c1f6e85c2153113ed5ae699dc398af2ccfaab1e0baa70703048d2b3061361a08bda0a215de35a302daf3d3caffdb29209c846f44672d4ee87066568552b6cae8b1d0a3d18a4a99f9e5d35079c6e9cb5ca609c2f049d6c880ad88aa80a0431d7f19e57ce24a1a7ad589d4c08af19e2d62344cac039eef8151ed88a4560e3c6ed12d6e0c1785e4222142fddaf79562e52f912bbd8e950d8bf620042f9b8c24ba7fa1fcefc46ae4d0ade001269153a1a893fafce8095af1c94c875eac1c1f78bea6a83ef981e03437498d8e3361378ada145e71f639e3ac434b3e7274531fda7fd8ccf05d9fabd68d5c684fad5ac34820455e91db0b248cac73f03412b81afaa8fade8202baac94c88e353b92905fe46771c1942182afcac93ca2f294cabccfb42d6ca2d99bcb2e041a5cbe9c3eec65f9e15c679d178311b247cb29945a57e06a354231ffc518aa7130826c2bffe71cfd80dd9c4b77605375b25b76dfb94317803f1f73ec8c4da17b9f0d9852191584050ff27555a397646eef2cbc2910e2932ae399035fe8ea54448001b68e2028b341ee9f808ed4264281c1011ea708a5ae6e727ce5af325ef20ec983ac5d3ee856897cb6adaa04d965fbc8fd54f37cdd505e43211990089b7168b673333dd96c436884c3d3f122eeb54ba66cf151000fa0b66e7b5604895562d36a27acd6c66da9b8d7509231c8cb2e0742361530873276f13118d363b90bf276c34324bb80e2d25f4b19ef8b94fe6e69ebbc5fc8868135357309204e29e7048666060682090bc076e9b2d5a22aa0f55a1ead6fde8b1e637244ba16a9c7e6d58b6a7f03fcce96a441e871d89cc5025edcd0e95162886d534cbf76c5f86cca72914a60c413a252f34450ead053a73355a47f2d38c9b350308cc2af948b33d83a6749e2008628eb0bb4c6670947c716393db4e7583be0d7f29cf2bf4252191ec03b669551986ef2700a157b3ed6f8092673bc512e2dfd8106d5a262328ff41f533b849dbedfe41a0bb2938fcd1c2ef81cc2553f178a4b15db1595f3d0ea770711d78ea346e8e763f4cf479939e59996e058f46f574863eb27295ad33f569bac9cdce8435ca78328db09eb2e01ddbdaad53efe21f040daf9a178088913706bc935b8f679efc0ff105c4994e5005a32d52a4fc4d97a5aa2e38a9843d90f754e002e52c36d5bcaa1d43fa9d0d69522cd5492487b9f92bb17c93f9f206b7492485acaec120b7ce2c26dd4c50d549848a73794ae333df48fd0720620b0f3c36b1fd53870003dca76b5184a11e82a4dab0f458a813273775bdd095541f7f866cf60718869bf21135823fdd573798f0abab36c947297dc750bdb8672b0298867cb249984c68054a44742818715c818b7d527598e436740c6cfc27845ee4c97efb507823c9dc40ccd38b3ff56cab41681a511b988d3d8ad4b4f8d95d2664c2d496b438ca5a9dca0cd9531b974a4df077d443001a764da3d36d885d3871e2a05be09782e138bb2315fc9ad1d61253cdb7a365f2114dc24e7cbf0d0a0c7cac5ebe989af94a9957771af40c84c9c62e9d60ad8407d087088d89de2fc17c571c1a4394e2dc0ca558ca05a129eea288ab80ff60891b7c3b5a971d30264f7a247ce2ea3c130acbdfab43ac0fe9a90e9f0d28696ee0c152f561e701c5695198b8eccffd31492e4d0f501357046f1824a3eb3cf952cd1be4e585311839cc5764a94ea3f87778525fedfd35cef185dcfcac39bf9dc0eff7a3e04f3b91d5797720e4ac45ca3ebad61c0d72568c7efbacd2c0b5656dfb6cc40bff1851d48ae0fa69b4989f7689ac4d2f4ee682a45ba41fe9fe7e18b0441746aff61fb0acdcb17378e170611977264843c2b01981c1bf390861aadf721ec0295f169cbc9bcaade8932f7180dc6f13d70f8a2806ad0b253c4700e64fed0822090ed9d05cfb9122bef75f736292ca685dc59f0b3e580210bd721b09fc281f6f902fec8777ce1340c3f623113cb8062d1c76fb92e9832d313c041362cce88d5975afc82d32f9562e225282dcae2fbe4ee3a68f2c4793abe2ad5c25b656121a169242c4c9593f0880f30216aa92761de879bab6063cff345099a3b58e982882fae19d4321dc8ebbf8a64f07bb65e65fbe337ea3f42be30d370e184f7741208b23138e3f34b9dc3a5f8100f2304bd253f2754aa7f1298756bace41915d894b11b0758448ccbdd352140b4afc96f495629dbf93080b69f510f4c9f4c562fbee7beccf2c7a46cc867df93387fd2b19cad296f3e32b57598026faa3554e70d9ab2d8aac543d150cbba5e47252bbc3cad69130fc639c93dd265eacca09454feb46d5ca72d41332deabd05f592f54cb55dccfa250c8eb0fad6468f526d649f0c8e01df83e6a393a1fa880273da8a06211557386a40f073f3836f5dae7f871203d5bb3c61f7aaf1cb625c37d6ef4784ff9fb04e61337ab8edefa68e7238bad97aa02ff2954a5154d137d2573cabe20a868c44b8dc6c14886dddc6864f7455c3014c4d7111138e08af5e14c6ecfb7050f648279a96012ee428bbe3b427434ec2f29aeaf6fd75adfd9195343be1f6790454600b3741053818e8bbf593f00b784ac34c5855a31afd243ad4f6e9825be910b3d3aa95208dcaed0c33b822acc044d62890537a6a2f30b7d3bd471215467ededf763476aa9fb4acfc9586fb5ce70d0d4740efde7d0f2834d40a9d7c215ab89b669627d1e0e45943cc4a06b1967bada344d74ace8f1675597ab681241885657dacd3803f507c8923fe71acdcc7e4e9ef04f19878c858ddf543ac969b4d2476bd3b307b4019bae75ed64b84680ef7a0e6468195aa5160fcf235ce24b2ced569cea5fa0201b9040480ed030be6177de20ed80f62db13d8eee8b70240a8ab78a11fa5fc303bac79065381ea1c3e9baf56d5b217e990bfc34ec3781943f2e548aef8e954787192a8d511cd84db3dd61d4e6387dc8fba50153755e0d37b80f8e0862ea928483c8d8d593fe26aa767ca07525e22bc86dbce0380158ad0d8b133bce349944dcf7acf14af42504928f3f62f5c577290923ccfef2b7f8a32b6c6b00bb07d579d6acbd7ae8fded4251aea1c6eb1e8594e55f3c4a889c50c342999e9154e5e40f200b90dc2cceaf27d252c5a6d6c73175fab2a04944ed9efb8d0f302f2a0269d6e05a83e01221bc7a897623ce1797572a98c236d5bee1549945179fa04445544137bbbb6df11f6c97b5d771f3274765da7d986e75d97a57a75067ec1a28404243047cd1b034ce16b2760ef28338ba5f460399233566ad1b5de21f32ee233567463654bc43a7c3bb21ea8b62ba022289819470dbce3fca7489fdc18e9b827d5f0cb956b26649ea532a75acaf54da3887bf69a594f1c96099ef73194623b5748a749571dddfce8a9d16e8400f4827ed6e37e44ccf07c1240d884850aea7033b0b77ec7d4420707133b80c5a7d2285673c9aff30de5d5d3c9b3cb0f9d45057e65bb03e375808f9596f9c456d6592eec77ee936c6be93e7aa0a8398f3a73d84b4be94cd555cab9404a6e75b42cda0f00e7678d4c83b76736f6266525b8bcf47ce00b997b6d08c69ebb604f45e7f5a7a8882d8ee10bd12c398bac4ba076fc85b02049dde62903718096172652cd62ab9288f6bda80f25f9f98f8e01c3ad149ef1c7d570a1524c1bb5411d394630ebcec97ad70a0db7abc02cb109d7d0768b946ec7229b1b0012c163e9ab5385", 0x1000}, {&(0x7f00000005c0)="c700e453a7a844144d56a773500543976c4985ac3246d3556f18f3bb36cc43b9110a5f90da17e02bbf60c215606948a27a682a5bf0d674c58ddfc4b4c25b2839d3ba8d463818e97a5ce7e7b601bd885857ff9c11eb819a3f15b5345b32d9a49764a12242dfb17900a52066d2768a923d7780c5bd0e3c7f9881cae9d12d8669cf30bedc7c89a444f639cb7d253b65388b8c5785d158243247460a1c414f41276edcd62e608c4016928cabef8ae92ffd4406e97b2e48efcd9e172b09d177fae369f8c87ff508ebbd514a6f4ee7", 0xcc}, {&(0x7f00000006c0)="03d12ff7fc65e7114815c1372563008f8d13637062940ead3b30ffd56eb341342f240d8d53c5c396bcb6f52c674ef0a9ff0bcb6efb21035eaa67de5fb1d247971ad0ec302c20c3ab90f6b87e07637838cf3c61aabe3e659d3e2725f615ae64af760f866dc5dc9e515d67727f6acb8b096aca13827169adba716dd67fa1ccac13500ec48622b25f7d881f8f8f4181f60b2ce1d2f5366b3a9e3149b4ad65ef9415fce00beb795fbcad61a9768eebd26a0c11135aff098afeb892d8594f6edb17f6", 0xc0}], 0x8, &(0x7f0000001ec0)=ANY=[@ANYBLOB="18000000000000000100000001000000", @ANYRES32=r1, @ANYRES32=r0, @ANYBLOB="1c000000000000000100000002000000", @ANYRES32=r2, @ANYRES32=r6, @ANYRES32, @ANYBLOB="00d6f556efa167d95440c709008f"], 0x38, 0x20000805}, 0x4005) 00:28:46 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x3}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:46 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xa501) getpid() getpid() getuid() getuid() 00:28:46 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) (async) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r1 = socket$unix(0x1, 0x5, 0x0) (async) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) r3 = getpid() (async) r4 = getpid() (async) r5 = getuid() (async) r6 = getpid() (async) socketpair$unix(0x1, 0x2, 0x0, &(0x7f0000000000)={0xffffffffffffffff}) sendmsg$unix(r7, &(0x7f0000000980)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000880)=[@cred={{0x1c, 0x1, 0x2, {r6, 0xee01, 0xee01}}}, @rights={{0x14, 0x1, 0x1, [0xffffffffffffffff]}}], 0x38}, 0x0) (async) r8 = fsmount(0xffffffffffffffff, 0x1, 0x0) (async) r9 = accept$alg(0xffffffffffffffff, 0x0, 0x0) (async) bpf$MAP_UPDATE_CONST_STR(0x2, &(0x7f0000001c00)={{0x1, 0xffffffffffffffff}, &(0x7f0000001b80), &(0x7f0000001bc0)='%+9llu \x00'}, 0x20) (async) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) (async) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) (async) r15 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) r16 = getuid() socket$inet_tcp(0x2, 0x1, 0x0) (async) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) (async) r18 = geteuid() (async) r19 = getpid() (async) r20 = syz_open_dev$tty1(0xc, 0x4, 0x1) ioctl$TIOCGSID(r8, 0x5429, &(0x7f0000001ec0)=0x0) sendmmsg$unix(r12, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000001f00)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b285b7401db87df3cabe102c6b1ca9722b2a487695cfb319901bbeb68f3aa8032d19e0a6676712fb853c8376c2e00bf82d513e9811ac95b1017861353c2916802da46d8166d1e59f88e2d503785f7e5c1d762ff8b24358fc236", 0x146}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000001dc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r14, r11, 0xffffffffffffffff, r11, r15]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r16, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r19, r5}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r17, r18}}}, @cred={{0x1c, 0x1, 0x2, {r21, 0xee01, 0xee01}}}], 0xc8, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @rights={{0x20, 0x1, 0x1, [r13, r9, r20, r1]}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24048090}}], 0x6, 0x0) sendmmsg$unix(r1, &(0x7f0000001d00)=[{{&(0x7f0000000000)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000440)=[{&(0x7f0000000080)="7b694cd5a7796b94a03c1f1e255ba37b3878184985af66926ecb54094a0b107a3bdeb35da756b66c1bcb49e12ad79d54e9f60358a2abbbc74b5398c94451a3827c8f158ec2c264fe37033dbe35ef84d221b00896c982eb38f35082396f20c500b384493822699ff93025eb7052fae9c42fe3da4b487c3b54cfac07b4e148d3960c49fafff1a99523b19d226d975dc8ae67ea1b2fa17ef14038e97e7aba277880bb04316df218ac5f37bda5ad02b5a31c4a582c6f30", 0xb5}, {&(0x7f0000000140)="970e1cb97ed177617f2e575d1ed4523648bab970f5b0cfc3ee822b5b6a617002c93f2182a06768ff4fd6d4cbbb0d4b397192b5", 0x33}, {&(0x7f0000000180)="fdf7a39c2718adb3c68e7aaa3d54b3141d2ff652b09a72253b7fa77dab097e4e96fa92b2703f8da6e588362b9a6aef49da73f5a41b6a8da8dba6a09e7c14134fb967916e38c6ddf60ad4b3408371de571695d46df5af1bad96", 0x59}, {&(0x7f0000000200)="be9c18268ff9d888ab37158db8eb88a7708b45aa6cc7f6f8326fa1aabf926d9ac19b84c5a7afced05ce0f63b8f9f0ae4e30eb0f64bb407bdcc379829c24cfa5272a0fdc3971600b50fd699e27fe67d20f62c14fb14748dfd234f8ec695bcd6831c3f2f5ad7a8e02a06df6cd6e859c5b7350c0cd2040e5eca5d7cffae41b61ef802530d3e014897946e8f9dfcf344a27f5ae94ff21f3aa338cd704ddf7214512062ceb773bbfcd61a1bb27bb01d", 0xad}, {&(0x7f00000002c0)="15fa8e6bef968f95bf395fc8fdf1f7ea5ba7442fa46b9990051bcd24f68d385c8ab2b9f1fda7bd6ca401053daaa928d3bd4cfbe7b22d36ecd901e70ea1ee83f80f590b62b7b3f359ba9216dd24158540dc97f491cb1521186115f93b6b7d6cbcd984289bdb0a62255d52b080225927a8c595e6e826a43f0e41b71f083c2d5ec2a40ba69540e28d822e0336aca7aa6b53c50de54e4b35aa0a166fe353f00d9f931cb7f1a14514a51544d198", 0xab}, {&(0x7f0000000380)="e9bbde2b1c99ce35249593edd179fe890c59bd8c60f55b8a2c296328490bd0197fb1288806f69633f13816a5f83a02106325ade39be1d50e9292945f383320a4e1c4ecf24617a08625be686eee3e1fabfed09a1df52ff4760008870dcb271055428b5918af31883aefbb05f00b46fb95ff13a4da78bb912399f1ed24f01f5b3e47f2cc7f6388b4eb2ae9c70e63754073e84a565eb363e602040bb8950e324812813908c8b309595a501b726746dc6c9ec0c3958fac11", 0xb6}], 0x6, 0x0, 0x0, 0x8104}}, {{&(0x7f00000004c0)=@abs={0x1, 0x0, 0x4e20}, 0x6e, &(0x7f0000000740)=[{&(0x7f0000000540)="480cf940815beaeec5a465fbe31fa39fb9311c79ede8bb5e2f34dd809b1e9c504d5a5f51e9695d594623d68860c42d57053598910c6c23cdf4415207a42a5b880a5964d8abdafd62bfd87e52aa9df1be6faecfbb87fd63d6c6aaa633f7d5c3a1682d42b7199ec5d8f101ed5a8b2d84490ef73483faf52f1b10ca77cdd7d4a107e7f4e4b481d50c789866eab440e6044c38049e4cfc198d0167b119790429b807808a20e364fdbea2db1c7182c35823f75915eed68a92934a180dd26e60bc1d1fbd33462cb7", 0xc5}, {&(0x7f0000000640)="339634a39fe9f01054d13043ce63881c6ad9bf507c5b526ded81f931c2f36f84cb7f49cf6752da62d56a29a440322c420e0e9b91ef9208235b60de01e932d5dc34e21a1214f02b0b585fe91d56069cc686b9363111043805b5b36d87f65aa594e258cfb98afe44f9a3fedb81ab101222ef2d0bfc2da0d26f0f253dd4be6a2ec8d99bd692e931ac8a8c40060b83172dc370a526c98855", 0x96}, {&(0x7f0000000700)="5302bfc827907f8db74cec479ceb0872012b236e0359a4040d9abb90b0c31736e071330d074eeca770c0d35704", 0x2d}], 0x3, &(0x7f0000000780)=[@cred={{0x1c, 0x1, 0x2, {r6, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0x0, r5}}}], 0x40, 0x50}}, {{&(0x7f00000007c0)=@abs={0x1, 0x0, 0x4e20}, 0x6e, &(0x7f0000001b00)=[{&(0x7f00000008c0)="a28eb50582abce56e9ef7abbabd4bdf9766f57082f235296d89bacbee99bb081bce7a045541c3f16f382b80b4cef2f600e1e0b4dde2e58e91f72126cabac0af6a1bd428b5c611c391b62c4c4d3a491c5abcf2f5d44be852045657d", 0x5b}, {&(0x7f0000000940)="9721b30d078b164061ae297f287c8a5b3ea03c3386216fc4214a92b491eba9210670a23fcaf8d443d4b8388815505047ff37d6461d1c90b785e7f64b41acfdae303c68014fdc59aa68be47c84407c8d11f2d81ea1e46965aa551f39acf22615a5389e4bca03f4381ff97a3cd488545b9703a919aa28fbfe933ca128969df2fe406f26a1755fd0edf4271efd201941cfca877a1b1d5f4304d38c7efe11e20086b562e7c8d53962c64a017d74ca8afce01549b8fd2f174b559aaaf9f5f4caf611a24abea70e3c6c52b0c291be875c42377956cce95b26b8e28863b34c10d499a1060164a291b3a9b2d7aecea4a320d4c56dd86a9614601569ca83495ee79e8b1d5add1c53f34ec3fdfff67840968e9efce3ea5c19dc013b39b73cd614aa12a38260f63654b4c9f70577b55b774cf07924533c47df0970dfcdcd6e9f8429b4b029a47fe5e23be4e1d9686314b401549f5e832a89b54f2c8bd6d8f0249896e9d583022b5d2dbac1c232740558a0dd7700eac8d3c19a36875f216a347a7b93d1494d01f50ce7abdf1d656a5e8d6e7655ff2e547c94112042bbc93e60485077a897aba2d6aafa13f38e17b76f135137c1e483af25b1bda76ac33b2d94455eb686c17e0aa2d96716413865b17a1722f0e4cb60cef5591110b65f9d387c7a18f466ac484d67751c593f71256972625b79f94d7efff54a3028d06c02cc408ceac0da6ecf0b17f155f67c8ac25f392230fa925bc6fcf24401d4764109275d80330c752859f4ffa8d62e3152ff3b167501955ee1968861d8270982acdc9409ac7c9e17ab49d3bd582c07cf46f1915f14cbc076eef9460f3ce8e81020da3b6cb97c931ffe28d09e6b5a2df09081f1cb8827105f8f896d15fe959fd1be04471d2091bf88192151c2f7be49b3db8c7cfdbd0bd0a8e5b9408f597eb964da95b3e9c341793630d2321b0ca90ee5dd625512dcdca0e435a99e7af54cee8d69386cc6a8c0535de799cc4d617bc1aba2e324a7d119e70ad26871b3cc7380023fdba262a2adb0f3d30bdfe5a6de60e19453b447699a22363676cfc772827b9c3444b1d860744d8a7002e1d7ce2f698ba6a60439b54765cdca8418c405106d31c2157d592f170be0b3fe11aba9523e30d901cbb91c178f5ebc91b0c1e903fcd723eb62ee6285e63efff7a7fd1479dab335848b52be4fad72dd09124e436c8c2d391184eabe0c5a5b9439d652632037bc0e20c0a3bd400b220ce2133cb8476b693a57ced7d02249b3e1f08fa37aa35f96d735b9a2846d8228bc6c6d99a2013f298198fb1c71e6b080d29a3936619553ab4e32f8d8adc4961894b40cbbc04bc59accb65289ab73b6767747c4eba8a01ea0bfdd3dc56776476448a06d8de7a80310cd1d17af2364fb2057e45d8e9b6842ad67aa3a2fd118f3a022277863f6c7112cbac2b40ca7d142b9e1d94d6ab19685f9a3ebb3c539a846320e2f28a395658a3cb6d43a59987bde373bd536e22b2b61a5a7839f947c2ffbc74172e37accba189dce017e45c2008e3ba6fa0925de32437e0c8d5dc549dc7f33bdfffce22d3d4a46b5f1c5084d7f2a1c41e821139449d916ff884aa71ba8aea622812c2a5437a678b5358e50d6c9c5238c01e1bc02d678be68656e33d975812b4a6a6267ca84260c736b3f661f22753d729019b5c5b9c0762b22f81fd8425d9f8a1899661f59a5847579afb24530e4844165427815fe3c662cbe71eaa0cbc3460fd45db3813881dcb506fa6e57270e5085aa63653c216beead4a0ad4aaf41dd0e9a863b6fd54dd2d331b93fd615a5f24825442750d05bcfe9f0fe590d948c57863705432b79013ef2210b25c958d062dcbca6c9a2f57013d1feb7c0b737f86b0afbd88f355bb2d918149e632323d13ded321688a593dabbce17258e7eff446b4b266c18ba859de77b2be65ce498756ef36afd7bb1196b8e0ecca3b8a5fa34ce4a05aac3634017b9b2c6ce169ce17995294343cc58c92687f7a80ee9a68662871c53cc3cacfffadac99cb43c0c9cfb47122f25776d1a5376b80fecdfda5cc95524684ee6d1241d6f997a00da61c329df1c9607afc864705b3a5462d6d024062e16654a1fabf716776bdeb86520cea7fb5368f87c866d0cf59026eb97edbe95fcabce34415666b4afd7798568cbb5b84e06b12ded55c41002b603bf431c1f762e3e89f1012c0d29079b4f18981048fcf3cb0386bfce2c640d3cc6fd629d2dcf15ece395844322277de4a379a9f706eafde94fd3c246ba868854bbb69b75b35ed36064b855f516a5a567fa58ef0c922be289dffd173122e53726376839f668fb9650e6f436d20abd37fc52713d4062dea78c9d74cd75195b7131ecbf16d89b9eeead587bef98f95302680f8c6b0aaf6a1cf7bce62a20d037dba29f7908dad48f362711fe484a42bebed6d55f70b7cdfd8d73a67a882a6fb68c681d257295ebae8b072f59adf243ccfcae923a70426c773eef34810a4d1383fcfdc2c2f6fd78e15ca1f98f7424c9ced32c441c4d8915116c26ad69909cf26e166107fb83ecf18fb0f090478fe2f01d575951ffe3962a68ba4592b0726110d38bd3826dc2a5cd7c5cf111dfbfd356492f5eab6ce2d40d12d76f9219e84759e011528450c35938e3f898708b4398938c5aba0efb8db01aa6a6f9a48c092650e68f956653658498eab21a8e2329913361d759325c3192c7cad853f7af962b431d8f36b313297538cde8a2642caa2827f37385d050e339aad88b20c3e1208372b4211a8d9d0a6a4e1ab54363d5a0afbeb8b20688fc5da1504283a0248daf62b6d1cd8a7a8d046beb9b47931d6adec8412968533633c1530e81d8cf4bfcb17619cac3a4bd7ce0f6a4f39492bb712d0a48ed4d99b2fecf92f8c0f38bc771fd8f773f86e21923f7a9536a35d574933bb4b962079162e15c4715d0ab46f83f10f4534b57acbb27cf089c597e1f8fb25b6a77873f6a8ce88d914cc8d0c5ea8d652cdf4cd956494259a1690abea6a3d46f4ffc08b88bd8ed1b791187eaa9c55ee008a895435e795117ad2748e430eaa6b817505881deaeff63ad808cb6b5a14ef255f6409d9e5694f95f61dcb6bbe87cf2b852c00232b05d0b2f5f36b73876277265afb0482f3c6bc597977e38f0d819d63729171c1b727ca99d9631a395417b06c394f5deea6ec2f2c59de6936a955a5fb91feecd20fcb7cd674a9228e407eadff724b35b55134d0c5df2fc2e314a3b94aac91c33d23128868f61c0eb5178e592fe41238fec12bad9a3222fe7058910bc3f4e36165b4289dac5a6e1dddc4ffcbdab6a46ca25b7d47748f6a0d8db9fa53254c04ffaeae62c693b80db48b3a19a006fee7384012d83e50672a869c9545fb22efbb3781d7396cc6fb2d5306b10337062be136abf865cd90c08c2777af15541d02500ef945d5de84215e18e77e05d07c8ebdc2b044b71a4302adc7ec7956021499f269e480a39680772394b49d5d1c34c996413868373b7431358ce28c4b9bf05f2096b562aca1cdf3e3e77dd7de23f271db8ad826fd8004287484b32b86a0c3a246c4f1046a055ba8f763a8b192b85a2d7ae258133252d063433358c10b21307eb97392ba5f687b6357d1cf1126bbe065feb37c0703b93d4aa16044e26199550b7f42918e96f5c29e6a6d00b93387272101bc090e947f5c561ee74fb462ec5ffd4d3990159792885cace6f23d7511b4c1f84f4b866c7230e2f146e0425cea5485c3b0c3600fc16ddc66633313a51ee296e829591416b3f1fd9e8aba5fe4be82ac4c6a4cca1e3a4eeddb4086fba024a941c7cacabbb2c66859b94011e2a3cc87e502a46b8ecd802f0f8faa16855fa8d0195787ce88a64dad21230e0ef1d4421c05452efeb749eedc8cbfe029486a5c2679b3567e5a742774328236e628335f98f7f9f965b81d28681cf3dfeed80abbf037ac10aa5a540ffb796463122070ce4d863d9e1d5de3e548d3ceb6bacb38217c2b833c776976a8e1f356e6df50be9a2e81d7b07a93d19d92a064dab45ca2ddc294695de98dbdc35f91079106565f3f29161dd3cbd48b7ae3f5889d27ce52625e9b32f29ac4f65d61b81f800ba5f6bfce568525725e0ed5784ce1a65e7e9b8cde66bec281ff1d99f380dae6bbd41f6913c5f85ddd71b1b30e600b7c4faa1bcf653718feb5ca1182b9a8fb590a815ed560be4176793d72567e660e448a2b82aabb6f2d42b743bd2df749962c4f6c8072988b40e96f07e57d38da8439cf0238ea6dfd3bf1276484dabf68a4278f3eac776e7678bc9c44290848c25b69b4fc08afee46baba3e1cc2603b15d4e05934da2f45e29190915df0d4871f0e18f07f4cb597e3d77606449958268e50b143066738809cc2389f240b34acc5dac4cead74a10bfa0cafe76823e5c6072527f64afd767019695a625393096c56f4a042a90eb96c940875e70e276bde2738e16e32ff82d1181ec5bcea3d61c0c989eea12251e725bd19f12c76174c2019bd3698b007e0884db9504118f4d0386dfddb9fff6b369a1582016c71f11e191d475b20a2866bccb270a9144d5ca9ef669ffd967fa3139758fd9cfce5afa3396facfe86d715d7207f2fe0b2b7051b0a6dacc93ff48b631970a4aa0d75bbe77ad7553e89a3b0cb86aa6788cb359ec3b88ab93b233f6345c9264143685f1a815202126d4c58b88094e0eb26490bace16fcabffb75a8edba389d7a6019479bae05ba2d0d1ad4b158baed02c62d403375e4d66e79f9ad528da243f34093135b7f673f0cc0fd020d1da905f3726804908bf2b9e713f000f6fdae1c3a03b4b6638e72c0ba16d0c3775c4f9a3215c3dae3703facedf3aca6646bbe6444862f23932f5517b0693f5fcc7068fd7d84950e13b9ec84f471833b8dda84f1aba2ddfedc44713fff0421ff2e845cb7f94dd5e7fcd513a1aa7f197304e4f6ffb50b7e0aaabe51fee69b938d3f616a1c8e7feb0bbb27f02b160b6d2253f44def31b1efd05108abd54468ad5f6d9f3911de822bbf38ca478643a4e8957325f29fa92d42dcc8586cb3e3c3cc7214c3f41aacba70eaa588f760569c98df0363ffb233a269f629a43af5f70f57f62b76357986e9680ec4d48889e3ec1f43bf2f45e6de2a3aa21ae2de990dfdb89624b0f42b008f94bc1b9441b2a8b1bcf5b1fb5463b0219f3a64db91c79a4048da204ce03fbccb51b59afaad80e9aa861d790f301710385a6d3319fe805e20fa03def50870ace3ba2a584ff53fec7f16b400c87a75ad54fcb8e3d9c111acfacc70a1fde3ee5c1e067108389b65d041ad30b88b28652dd04c0d53db0c4c9318b21905e5fd05b561f5bb6c6b77ebed97cde564551e741daae60257db41ad51aebb6512ec42073f82cad4a08966a9b0e4580895f7a4cb133b041cdd51e79ed164476ef13fe87348bd6a9f86a7befca17a336f85ac2cf1a7a9a0209deef8945eed63beb633e32fa2f6597f113401e7cf36f3d35934f6ea13846f23e63fbbedb86d9fc00f35d9bd7f996fc965d37d97c4312d1bdbfde7356e78eeaf4d7da6542ff51a658646ec4f89f93d4c2bd49651b31ff9368458da5a9fd612fd5390a4da9df47f476633ec42ed4d789dd317f06d0a57de5c917148fbc1300705a9cf7ef1f0d97123c1b9ce6de815699d44e693a8259cfbb36e401a2de200c47a0f7b09458ca115249cf8255a7ae1ab3f7d8377fa303fbf1674b45a419be48598dac18338452cf78a65cd528c08a2b909bc3460361ed4ea494f97ddedba1d81a71921b965ade2e3868ce2af9e08f922f0e1e103821dfae75c53f", 0x1000}, {&(0x7f0000001940)="ec6b8155de8a6a59834f772c0ffc8ab6192a55b488e0a0eafbc0fc7a2f41324d182c6aa87cd0946506c8651629246c57c3f43a7ee3db5ff8290f312f601d572dab2763821d287515b90b341efd00883ffef53943834de0ce0c8c42dd976092fe1f3ff56e056da9d611b0614acc9d2513407d4d7d1ec1a88b6ad483e66bc2aca7d17ef9380b1a0fe5ab2ee030d85ac498b55fc1299eb8227b3bddf231f261d8e9f874fde68e6be3df87e772c004e8fd6f943c66f64acfbe2f002ce8e33cd0f3582d54ede597c13d62b2a438e1b8d34f735461eede8cff17fa", 0xd8}, {&(0x7f0000001a40)="842e0430fe1056b9cd37d460d30db5", 0xf}, {&(0x7f0000001a80)="38639fbb08e89efb49799b75b139c01b547dd7e54f676cbee2d969ac003fd42510de4ff84722480d5c4843d662a5bc7df73d65657a50d368a2b6bbd6b2112b3e50b42a7ac577723fe76ca3ab8751c59075d69f0705c10f9c5dc8763da0ff442de752e2035420cefd0df38cb9ef048ee0cc1b03", 0x73}], 0x5, &(0x7f0000001c40)=[@cred={{0x1c, 0x1, 0x2, {r4, r5, 0xee00}}}, @rights={{0x24, 0x1, 0x1, [r1, r2, r2, r2, r2]}}, @rights={{0x24, 0x1, 0x1, [r8, r2, r0, r9, r2]}}, @rights={{0x1c, 0x1, 0x1, [r1, r1, r10]}}, @cred={{0x1c, 0x1, 0x2, {r3, r16, 0xee01}}}], 0xb0, 0x8000}}], 0x3, 0x0) (async) getuid() 00:28:46 executing program 0: fsopen(&(0x7f0000000000)='pvfs2\x00', 0x0) (async) fspick(0xffffffffffffffff, &(0x7f0000000040)='./file0\x00', 0x0) r0 = fspick(0xffffffffffffff9c, &(0x7f0000000080)='./file0\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:47 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x4}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:47 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x800000000000000) [ 1726.664333][ T5651] workqueue: Failed to create a rescuer kthread for wq "nfc9_nci_rx_wq": -EINTR 00:28:47 executing program 0: openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000000)='./binderfs2/binder-control\x00', 0x800, 0x0) bpf$MAP_CREATE(0x0, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) 00:28:47 executing program 1: bpf$MAP_CREATE(0x9, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) 00:28:47 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x5}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:47 executing program 2: r0 = fspick(0xffffffffffffff9c, &(0x7f0000000000)='./file0\x00', 0x0) fspick(0xffffffffffffffff, &(0x7f0000000040)='./file0\x00', 0x1) fspick(0xffffffffffffff9c, &(0x7f0000000080)='./file0\x00', 0x1) r1 = socket$unix(0x1, 0x2, 0x0) sendto$unix(r1, &(0x7f0000000100)="02c284ec6e2b18a433dff579b83311ce5a1ccfdd4583b509abe5c6edbd5700a4ce502f2029689ac788abdb0076897c9da3393b5666f9212304063a8a25850f949cd05679ca647ab13bfcb95886871451a6aa73ec3355d7e200a562f60208dabb1a93680de0c76e4f3390a6cec6769a55b7349abc801bbb89011862b4bf1a615de4bfa61476f6361cb2f21205c00d", 0x8e, 0x8000, &(0x7f00000001c0)=@file={0x1, './file0\x00'}, 0x6e) r2 = fsopen(&(0x7f00000000c0)='pvfs2\x00', 0x1) fsconfig$FSCONFIG_CMD_CREATE(r2, 0x6, 0x0, 0x0, 0x0) accept4$unix(r1, &(0x7f0000000240)=@abs, &(0x7f00000002c0)=0x6e, 0x80800) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:47 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xb500) getpid() getpid() getuid() getuid() 00:28:47 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x6}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) [ 1727.066316][ T5720] workqueue: Failed to create a rescuer kthread for wq "nfc2_nci_cmd_wq": -EINTR [ 1727.205096][ T5742] ERROR: device name not specified. 00:28:47 executing program 0: r0 = socket$nl_generic(0x10, 0x3, 0x10) ioctl$sock_ifreq(r0, 0x8934, &(0x7f0000000080)={'gre0\x00', @ifru_ivalue=0x33a2}) r1 = socket$can_j1939(0x1d, 0x2, 0x7) r2 = socket$can_j1939(0x1d, 0x2, 0x7) getsockopt$SO_J1939_PROMISC(r1, 0x6b, 0x2, 0x0, 0x0) ioctl$ifreq_SIOCGIFINDEX_vcan(r1, 0x8933, &(0x7f0000000000)={'vcan0\x00', 0x0}) syz_genetlink_get_family_id$mptcp(&(0x7f00000000c0), r0) bind$can_j1939(r2, &(0x7f0000000040)={0x1d, r3, 0x2, {0x1, 0xff, 0x3}, 0xff}, 0x18) socket$nl_generic(0x10, 0x3, 0x10) (async) ioctl$sock_ifreq(r0, 0x8934, &(0x7f0000000080)={'gre0\x00', @ifru_ivalue=0x33a2}) (async) socket$can_j1939(0x1d, 0x2, 0x7) (async) socket$can_j1939(0x1d, 0x2, 0x7) (async) getsockopt$SO_J1939_PROMISC(r1, 0x6b, 0x2, 0x0, 0x0) (async) ioctl$ifreq_SIOCGIFINDEX_vcan(r1, 0x8933, &(0x7f0000000000)={'vcan0\x00'}) (async) syz_genetlink_get_family_id$mptcp(&(0x7f00000000c0), r0) (async) bind$can_j1939(r2, &(0x7f0000000040)={0x1d, r3, 0x2, {0x1, 0xff, 0x3}, 0xff}, 0x18) (async) 00:28:47 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f00000002c0)=0x0) r3 = getuid() r4 = socket$nl_generic(0x10, 0x3, 0x10) r5 = syz_genetlink_get_family_id$ethtool(&(0x7f0000000180), 0xffffffffffffffff) sendmsg$ETHTOOL_MSG_COALESCE_GET(r4, &(0x7f0000000980)={0x0, 0x0, &(0x7f0000000940)={&(0x7f0000000680)={0x2c, r5, 0x1, 0x0, 0x0, {}, [@HEADER={0x18, 0x1, 0x0, 0x1, [@ETHTOOL_A_HEADER_DEV_NAME={0x14, 0x2, 'vxcan1\x00'}]}]}, 0x2c}}, 0x0) r6 = syz_init_net_socket$x25(0x9, 0x5, 0x0) r7 = socket$inet6_tcp(0xa, 0x1, 0x0) setsockopt$inet6_tcp_buf(r7, 0x6, 0x0, 0x0, 0x0) r8 = getpid() socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r12 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r13 = getuid() r14 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r16 = geteuid() r17 = getpid() sendmmsg$unix(r10, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r11, r9, 0xffffffffffffffff, r9, r12]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r13, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r10, r14]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r15, r16}}}, @cred={{0x1c, 0x1, 0x2, {r17, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) r18 = getpid() r19 = getuid() ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000b80)=0x0) r21 = geteuid() socketpair(0x0, 0x1, 0x9, &(0x7f0000001e00)={0xffffffffffffffff, 0xffffffffffffffff}) sendmmsg$unix(r0, &(0x7f0000001f00)=[{{&(0x7f0000000040)=@abs={0x1, 0x0, 0x4e23}, 0x6e, &(0x7f0000000280)=[{&(0x7f00000000c0)="4f1268934c4da557c27dbe6e29d5d50ea096d07c8d4cd889f64bacee9f002f7625d3a13c2d050daf4ef51db0640bf5d269bed4a9c25b92deb1c808d7cf93c9c6aa3445229447bd84526441f07f71aa131b32d02c5991295aa435a8221ed5ebe30340aa7c0a22bcbbbd1882b95534f4e7899d5836a9c20cc26799045c8f67a5dab116e1ae697897c9d7eb02397dd38a9874ef4e447467d875e55fddab8d84af6a473b29316f53de7bb4cbb1e41cbe7caba0445272b9699a57d747b1e6a6fa0b7eef10f3fae80877733faf96270a24b5cde75dd484be41da7ab8406d0176ffda2b49519ac98b7174c7f8627f3cf1e4254f61a5560c2f87cff98f", 0xf9}, {&(0x7f00000001c0)="aa6ae15fc2430cdd0852feac4c44a0323eb5a322f20d77dff3fa45e15578e5ee99260ee5eeac6115518b84cfd771569cbaa2b64f274b57c7dbe75215e29d9c5c97d1b1355a020f47ce9f03966675997ddbda2e1bcb132f08c13fc0c3b2999c674fefe5e03735d37fa2eb356ebf0ef80e7711876483c6fa3c6acbb2a78e884d5db16997c2ce4acd0e3ef4ac5608a40974935d1e787c10909c597ac619713ddb8dfd564559e91f5fa7", 0xa8}], 0x2, &(0x7f0000000300)=[@rights={{0x18, 0x1, 0x1, [r0, r1]}}, @cred={{0x1c, 0x1, 0x2, {r2, r3, 0xee01}}}, @rights={{0x14, 0x1, 0x1, [r1]}}, @rights={{0x2c, 0x1, 0x1, [r0, r0, r4, 0xffffffffffffffff, r0, r1, r0]}}, @rights={{0x24, 0x1, 0x1, [r6, 0xffffffffffffffff, r0, r7, r1]}}], 0xa8, 0x20040000}}, {{0x0, 0x0, &(0x7f0000000ac0)=[{&(0x7f00000003c0)="fbef377aa482a296cacc9595527cbafb72d5733c8dc76a65aacb75c03898fc3f37a06fe1196c46fc6a17bc454bd6bc3769947c4b97e77a2275f381b66f40b8dbe0f80faa497c6d1674f1bb2c7a0898ab0edc39b51711217405b30b4c7bd7b077201c6eb8f61d6e6226aed0c91797b858217cc1650b3a43ba74d4099823c14a2917e74535fa9d512cd95ce7979da2eeb59a0be51459763aa419c0cd33de7cb997882182889bde144effe49fe1a37ca6", 0xaf}, {&(0x7f0000000480)="4c3a3ea4052484c161d6e5cf18180ea03443ec922887abd649e61bb0309dd92014edcdd18497e1f6b4a4d7cce80e2ea4541c9ac7c3674d3641770bc4b7cbcdd82cc3641dc418b1e1103deca5c51703494c3229df184f852bf7a322aebe4898f9e1baceb9dc", 0x65}, {&(0x7f0000000540)="396cd6dc1dd898c175c691ef48e5", 0xe}, {&(0x7f00000005c0)="e4ef3a93fa1a3ee770710478bfdbc78dfb301374b96e80659b49229bed6f57b299c0b69c406aa635cb0f30a156ac574a071027f80994ce130cb0d54f419e53800027803372dd5298f996109f463e8db58d1af6a99c9c93c8467a30cf1f830f9cf011b4ed26e13f380461ae75dd7893d47e3477785d1ed6bb0a446dc0945ff07c2388eb2e05b5a9da2a3d10aacf9cd092c83fe41b2b261bc7aae7767af7e259b667b4a2f4535353ede7d2fc9172ea9d796a7ac8504920360954712f", 0xbb}, {&(0x7f0000000680)="2445f1ba09ccb05d9cd51b21d387b7f24a7d17df0b3ba289df1bd7f1e33331549b31fc5830f6e6e858eab99289cbd4d1d4fb2b516baf8c8224023436ad9e2962f5995d90e3821242e64a8665fee5831b47dec99d43c706420b0e72f69daacc10c20454182551a3e29dd39b0d3955e1fb98d5e6c70d659830a220d94db1c66c2342ed9d83cb236bde080ba1d31a3bf90d1313b99924c06a134d33b266dc0c095dfad30163d62d123019de5f2df553441205c50c31c98fcd", 0xb7}, {&(0x7f0000000740)="8181cfc5eacf34bf4f87cd238639578cd31b65e1c6486a3df7fa39624f212462d047ab72d17e6c9aaa5ff83c3da7e780d2e42a31eeccebf1da29ffcc0b3a9889e88424286aced134160d687deb55054b63caabd480ad8557246d73a054d2e02471f93621869c912c20938563fe08d51a597cf608b7f3b67217cc704904871c9bd74bba1077aa1aba16db810989ac9711cf05ba44801f1699712979f2b0b69703e44e286633ba5cbeea8d540e94de6ece738fbbb65cae2100f0fade3c55ec2fd0fd7a2589da0cf33bf8958656fe92115cef922af9c07b5caec479a1a3bc2c54cf9c911c003cd6c8f703f3", 0xea}, {&(0x7f00000008c0)="a8fabea5b7de9ce8d1ebf60414d8909c9395b52882c9e2c29cc59f85c2cd0afa7353bdf9b123267142193cd0777fd6b8544d633ca86aa7801ac86b65782dab40a1ba1ef8e0e6f2b05f33f9b4a00894d9d0702dfb02b34ab5fe68dc028e1725ae4b729deb97fa01a1fea350698e85c59bc1111090d08b4034712aa8e00f9c7956feac0a6281a405be572e596b04aeb18919cf8001d0f6edeb10a91d26b9c0c49c5874046ed3dc2dfa14238534de2bcb29bfb6eb82ff6ca7fca46507537af52a504d8cb224b8d6e0587d37739df6c9f79874f946c057c2ed0af8df", 0xda}, {&(0x7f0000000840)="f0c424aed7e8638745fa5aaf081e8d634512d9229446939c31545bb515ba3fa8fb", 0x21}, {&(0x7f00000009c0)="76db349b779f1b83b5919546ba3260e6b76cb44362abadaad53f0e29667b72396631c262d8aade5b71ac669328663b9c942c0bab788f49b42fb6bf9b9aebb5c160ee3657e364b4932e8205893898be8dd56c84d5edf050ab32355d8bcd1bc5f2afd05bf056f85cbbc4740dcc1ee4449f35f41eba5ce5d0d6bdf3d736b10cfff34ee42bf6b252889fae21a31972ec9a7b607d510ef476832af1d8b96341eb59719aa2", 0xa2}, {&(0x7f0000000a80)="e4", 0x1}], 0xa, &(0x7f0000000bc0)=[@cred={{0x1c, 0x1, 0x2, {r8, r13}}}, @cred={{0x1c, 0x1, 0x2, {r18, r19, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r20, r21, 0xffffffffffffffff}}}], 0x60, 0x4000}}, {{0x0, 0x0, &(0x7f0000001c40)=[{&(0x7f0000000c40)="ad142f1e94dc19ec3a17757f6d986b91f83a75ca033688b035ee3ea2ef237c66bc8aeb6e7cc3cc78e28da047ee8e0503e8bb6303970f089a7f398605dd4c9d9d93903aec4394e1283d6862188fec396b39dffd5e172f5382a075e60717ad7e4e6aeedc30abb026c48a2bd7ebce850a4a0fbef88223621c1afa4152196c408dabd49ad6bcc0af0014d355e5329c472affef59d3c4e9e3b4b97cd392e7c72151aef1d8d4857dd6bf45e6c2a158ef297a0c902e2c82523e8ee03827c2a53a089aaab4177e74143065141e6bc3ceda8cfce5255faacace01f9d400362a6436e844393abe7ff745ff76e3653df2d5529a52572a55e62d773efb2257ec6c7d518e2dfdff3275fc9eabbf98881994527d18e94541adbe35aa3c4845090f48e470363ef93fca7df827e4fb3997d57dc1ebcdac8f864382921220112a214be818e740c1e59edccbb55446746238e42697bbe061e0780b632d2347aa14df615cfc26c19666d4db36a8fb828e26b7eb938a104899c4cd683a57e47970767021eb2e81f124f218f62554dbbda116d1fd9eb697edb764067435ed26e35de176bc0694758649310a91c3676bc33ebc82b7f14b2c9c4043f9aae68a956b06c407704dbff67e0f38ed98611f7b97609aaa76d8f2b1f7c84b9b6a0860b0e343275b35ec375dc18dab93e0b26167af76c5855aa64d96f52819a509891628a508d888471140b220a1fe6f866e103c48f83f41ef958f66128ce5587dcd0b8f5f10ca3730f4861c732e4ffb35da2653fc24e3cd68018acf1281ac01ea45b25e94ada62bd6780c3820850c59cdb2c6e799ae85061b3ffc16b60d7122db7e48da9cee6f4f1dbe035e06130ef5a829f80a3d33e42b55e1ca9491538fa049a639de8433c414f68d82ae8daf554fb9d6b61ab1f153041168d9bc87f50dad7bfabc64e358214f58027c03358b4af441607882070b9b49d8d59680aef2831007f0da97bf45df3168be08f8abeaaa31d6da923abb3091f0df5b1c4400c81ff6e51b858298b15bdbfc8f45323fa39de57227c810b1f7167d4068887e87c17c0566a59b441c68f39918416e6ee8a31af1798f07ede7eaef9a29a0006a2c2d322c9b09c04fca2831660495cf9701bc40fcd6f2e23bf4742f36faffcf12185953afa5ffa8284dd12519f8cc3873134cbb6a80e28bb1d47712e41dabcaa68a77b0613ffb2a04e3bb06717299908b44b30da497d81cf399357788e3703b1c484b55bbb3ded33452823df4743a5101d38b93591d42cd1dd9bdcb9992b779505b1c3599d24be9d72cdb139eed29501a7e6c94e979304e02d946910676efbd5d64160355586e84c02c1bebfda6d5fb581714c75f47afec5d25e5fdcbc5f3cf519270b8126c2dbf0baa32a0b2db0ddd0d69f3e0151c3af74323ae83587ab32732e9a3a9111b604dac6e926ed384d5dd0fc8b9cb0ec8ce115f83d8ebaa8a636d0aae2c9dc4d8032572f3acf25fe8e9967f7a651acb8ebb8558af65d3a83ef646c75f4c58aa0537a3c57e1ccf6422325d90a0c4ec34610767589e13577fc046632d9b2cf1c5f4960b30b9adff5af55e5eeb4889e636d0503773fd22fdf60602ffa2172a0b8a71a3677b03646ca87d698caf4b98478a96da31d0593650a5f6fe0964827f8c7d9fd0a061cb80e8a1119eef99d66775034c98338286b1e22edeea400628f08f6f67c10aebf14895afc4794ec470720cab7f5178c2f4114740de716a8daba7b51005818e7090e9faf35948a25c00e7bf71e9234870fdfe36b11ab5ca640376a15c0fd99fdb9b45fedb5af7b65dbd3498eb7294e255b80adc46a91b568b014ccad3ffa265d1a5620a5469935c19f5595e739f6a3de1a2849cd2bb042d9747995c00e6133d1d0a0a40d0388329866afc32cad2ce4855bd959a10ccaab1e5511cca53dafd4c83c091ac6fc02eb54bd1b61d6b55fb84f21ba94728fbf6c98a197ab6823f909f029c2f89018bd69a005e5827e27d54de5cef339a67a84c7ca5089e47ed893f037ea27e5049722e711cb13dea0d602b8275d8de65064ddc457a1186c9b5fe769d163e8c2255c76e7585b93594274784b4a42069d0781f0096970fc52ec0895285cb453205bb7a30723d183f00d353d39f682d5c838cabd299fd78358192bcd3fa744c109d24db6bcf49d0975f1a8c5edb78212d4b74cb18541b73c18af159f7550e70a307831ca4736ea0c9ab1182eb008bb6a32ed04738e4ee89ad6ccaa05e76e814433777e09812968a28af783ade5661cb8d62bb8360b61329140e4d2a1f7ab7e82b691fe6aab9799e1c2a9221eb5ccf5cac3b37d32e3964776e7af56e7a6d779f0e082f11917543f03345443dc4e84e676fd7c8f8d19dfa4db6032d1a20195d79bf97bd9905dc1c6af5d3a9f4f197a192f9fe108fd55c0d48a303dea46b5c3c22d8385b64abc88d04f8bc990536ea9a393c822958c7108dafc5c006caaf00b8088d114ee2acb0e12f574fc49bbb23de0295d7fcba8394f4bda5bca0a3f3356ae916fe942b9b6c6f65ee38cd2aef5681262e3f7b0e1c71f8396007c7db9749cdc0b7dc10ee19dcc48d7f6d0942d2b94a1fb8e4e5f586311e1556dab192115b9174b1968718ea94e4d6481f56ede8cf73e74af5522cabfc150d2c84bf51dab9f2a6ca69bdb23abd7447c2ff80e20c7433c1c472123a3d36bbc568ef293587bc02326700a5ca2f472ebed5e21a60ac5b10fc3be0c3343f29bddd4c342c587be2f259a83f104b215708f77f81d5fb6275d7f213d064674beb2e393448aea6407458333e68bf3f45212ec3660e58d9285e77e5690133512e36e108ba7bfc2aa43986c9c33e0dfff2c76fe6d6a83e5aa90cf0eba3f5ba9a5f51b9807744d9039e3ac55d52de4a2a2ba04632a7b21dcb25680e484e1c4b7170167df52ea33744c806940d488407ea83a12b2ece874f4c903b6a369130fc1184b0ec8abc3ff65cc89efe7c25745f1567c34a5d16292d84327989edf73e4259b1e70d1a57dbdc29e85d1b84846ed0bdfcf4cdad6d46111d2d52db1ff23e34d507563135d2b45978371ae5d7c9488eb7adabd684717a7215351273a5603747ddae1ed0dc2ba08379d6956f5434b2c15c46d6b4e58cc56f0aafdcae80bc63c3700234811c05cdabf5761a5d025c33b70e21c58a4b64c66a9a0310d1046f763bed78a9999c3e5c7dd347a0fd86ed9913c5c77126144dffb9c113f3a0f5a18b36df9cec1526e903c25e740685b083608fc0777bbbae542039ca9b3fccbe5b636e1f8739e9e93585568827618ae681a981f739fe1a75b2c7a86e2e6240d19072dccefb1150f0efd552fe4c4844f748df12a8c165df43da80dab27fa7a09454212ee2aa5b81ef1ab6cc20b99d629feb8c6addc6214dfc532406cb7f69834ac3007dec2ef52e2db43422a866fb1292fc022c4db2cce3b0a2dcf62b31813a9ebbae3d14e8ccc02d5f220d8370c787253ce850317293f419fc828e6c48497cba4986df6c585ce3e06a4b1649c4a18378ea7916fe2d767c441c86afb86aa9db22746f6df444dd8bf33b8dfcf2570643380414406b35a7676f7f59bc292f589d3866b1cd52ef854e73497a5734a3054c4935b1d75737a5e394a8beff25d44bf99980e8c8daaba6b4f0f702c77a3ff3730b271aa2b340b5fd4ff78e5f7fe69f46f4e7b4faaf2e1a06d5f0aef24da7b52c030f9e71ce966af727982d2ffb0b3c34e7ed027a71f97be55788b48479ec4504f2739e4c5a83f08e28a23882f289c8e6b0bc37fedbf443167b9f8b69ced21477cccfcaea018e5923fcb50381b2e4d132a4c2f813b0fe9e56a0f09939bfc930fd5f121ac4a5d7322ad7568ca63ffdf9b47f637e42021980ae4bf4d8ab76c963cd00291a44dc60260a51277b69449b9fc175b61e4f1d545fac4ff876fee6d25080a46b02261b9f1b62fc6dada23204caaba4704cf1ef890584930f02d292be9facd1eb270d1429a36adb38b8850610518609adf822d98cb9c1e86153b4753f324e55cc38f6ea9f35f118b6619ccb8b11f8e28163081b5dda523bab7561a71fe425348efa7ac221552f54d4a96f054047aace8e457d0ebbab91e12b5b8ca40191ecb8477cc2f2edb5293978c17e1fb9b8046055e1bed3946cd9c4aef9c0a95314ae675a3d2f37d90eb9c84348f9973013e5073eab98e512df4a12d10702cdfb28c87a83d5a0d516dee13121cb4bb96669c4518041fd6e015822458d23ad60dc6099a00ccd5c0823677c0170faa42be04cae9000a70773d18e53be864923d03fae31be77e36c01ea370f4e2cefa12f670dee3c53a9160fe76048c11e7b3113f98cf0e78cc07e64837d1f80159d6e1adc3b5af658cb9365710c84d5264e0149a893c97182c3f6505008fede6f100b1e1f1b54c2c53a2c3bc72a292ca6b354f3d2ce49fcf6c2ce3dc3c549d194679f90ffbd1c8ce55c4d69476fc878ce2bde130b8d91e080659db5b39e3f399f2b8060f0fa2a6da00f576e98a2a5203c3f3cbe132cc32b7af795ff8891bbd1520e78a7499a09961c55ed5d52dfe3aa773696a04c5d16f5a6b07a2e500a06e34001020d093010b67dfe479a50fc356ae5a54437fe6b5c37f0cc9b815f8f81c47d0f153729c21155d70aa8504c39e31c8cd9c9764057197d5beaa895cdf48b08c107fc08a17252f4b68b946b9f40d800e865ae6b60656c0157174eee4111379f4dc5ba2f0cad0b5366cfc398e431c91458c114377ea93b83109ca7f48b7b5eac5090939de869098090d1e14aaffc58f7546df7de6e9fc0c0c5eb55cc572a2796da49e559be7d73ba48e8b15743475858d367fa5d30ab8e08b4f270760db24b00315332ef6491973ed8a23f296f7587a18d54a0c0923e50f7a596f9f6a489bb5c82d80bc7cf0b0e458e8313f3d9dacc6746b4b5988b6ece979196170eb28134b754e63bd7e8f72ba53a24bf36dae2eb323ecca6049ef6e8573240f034507a36ad2b64c5fdf002cbb119ae9f2fed395330bd3db9694bd82a6b930641825f050227658571f98a265f5829f3d0ba7417075d10c4d4f8316aedd6c95761ab498172f9cc331e15d524b6ed42558650476356c5c2fb26978fa54d1721c3d7e6ff6133390ffcedf4842d3a3c2a170f7882df3ae4f57460d34b1a288064fcd3cb82bf4b84e6f69d5feb5ddc1469dcfa3c2ca0eb46110b5ddee5ae65ff7ac74c35d562c76df64282d6998aa1e7d34758e2839187724f329dda822e021901cff38f326d54f6b29972bfdbe7011531d1d2c121cdb6620fd6cbdd95190705075cc0d6b822ab61cfe545b455a8fab3f3df005f7e7afa02e332fa466479d7ab0bd7d972d4191c8b5dbe2d96fec963886789af89d6f69b72ed6723d82ab42e46a9aead28c1a80aa63e35e40203556e47c08064936142af795dfb26be56a11b0508a42220c57bfa485e12d1de2841206ad410f2b2a77eeb579f106b1d80315ca7baddd6e5dfc927d3d5d7324f5aa0fb2522c9142320e27dbf13d8732d4c9eaa9627d3addd6d0ee8fbebf27c01459e0fde0cf867720c5ac033e99562fdc4585e332bb8cb3c4e643f390bcc3b6a8cf102cccf3d5f31679e77202f0a7a0180d40eb2837a3ce45ee13dd802e2cf86448caf93ebe319a28b8b64a964a978e15f9ca7c601869e2aec486490abf5308a5b09d350edb9b4031d117a7fc9c9437560e3b0aea664550d12c97b2b5615215fd8fcb7b04d4a0149e85cbba9dffe7ebfd54fc0081f6180541292531e4af82ff3e438f9bdc9bda186028e10515bf3cd7071d2ec00ddd88f171596fa036098", 0x1000}], 0x1, 0x0, 0x0, 0x4000}}, {{0x0, 0x0, &(0x7f0000001dc0)=[{&(0x7f0000001c80)="3814051670d4109952abc445432cdd2b17bc640e5b204ce375555a0dc31e33f9e5118e407d27850b8e2b634b73292209e24584d2fd54abf6d48af78e22f4746bf8bfbfeaad6b3274d8366da3b0f47e548c6708ab38227b46a089b20812a0a42901353264ca0fda09238fa6245e929ad8861d1cfeff2eba1139ca93f2bfc398c5334c62b51ba06c0c138fbeae9c62d70f4b9f0208643e3bb010a3fcc2cbcad88f769b09d31538c0084e173aee8a", 0xad}, {&(0x7f0000001d40)="70c14ba0ddff01ca84f6fe47d99f9a3bca5ef3f0a48ddb15b97f40da34b63c536d4352bd5aa7d665f1daf123a221c8f463f516d6c50cec3e001989935aa8aeefdf37a0a722186c86e0dfde370d910f98dd365ade37888d2d522e033aaf161e3c0bf7429b1539379801dc5b4d1c39c750187fdf3d380f4f928d240282ce6c29", 0x7f}], 0x2, &(0x7f0000001e40)=[@rights={{0x34, 0x1, 0x1, [r1, r0, r1, r1, r1, r1, r1, r1, r0]}}, @rights={{0x24, 0x1, 0x1, [r0, r0, r22, r0, r1]}}, @rights={{0x2c, 0x1, 0x1, [r1, r0, r0, r1, 0xffffffffffffffff, r0, r0]}}, @rights={{0x18, 0x1, 0x1, [r1, r0]}}], 0xa8}}], 0x4, 0x80c5) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:47 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x900000000000000) 00:28:47 executing program 3: r0 = fspick(0xffffffffffffff9c, &(0x7f0000000000)='./file0\x00', 0x0) fspick(0xffffffffffffffff, &(0x7f0000000040)='./file0\x00', 0x1) fspick(0xffffffffffffff9c, &(0x7f0000000080)='./file0\x00', 0x1) r1 = socket$unix(0x1, 0x2, 0x0) sendto$unix(r1, &(0x7f0000000100)="02c284ec6e2b18a433dff579b83311ce5a1ccfdd4583b509abe5c6edbd5700a4ce502f2029689ac788abdb0076897c9da3393b5666f9212304063a8a25850f949cd05679ca647ab13bfcb95886871451a6aa73ec3355d7e200a562f60208dabb1a93680de0c76e4f3390a6cec6769a55b7349abc801bbb89011862b4bf1a615de4bfa61476f6361cb2f21205c00d", 0x8e, 0x8000, &(0x7f00000001c0)=@file={0x1, './file0\x00'}, 0x6e) (async) sendto$unix(r1, &(0x7f0000000100)="02c284ec6e2b18a433dff579b83311ce5a1ccfdd4583b509abe5c6edbd5700a4ce502f2029689ac788abdb0076897c9da3393b5666f9212304063a8a25850f949cd05679ca647ab13bfcb95886871451a6aa73ec3355d7e200a562f60208dabb1a93680de0c76e4f3390a6cec6769a55b7349abc801bbb89011862b4bf1a615de4bfa61476f6361cb2f21205c00d", 0x8e, 0x8000, &(0x7f00000001c0)=@file={0x1, './file0\x00'}, 0x6e) r2 = fsopen(&(0x7f00000000c0)='pvfs2\x00', 0x1) fsconfig$FSCONFIG_CMD_CREATE(r2, 0x6, 0x0, 0x0, 0x0) accept4$unix(r1, &(0x7f0000000240)=@abs, &(0x7f00000002c0)=0x6e, 0x80800) (async) accept4$unix(r1, &(0x7f0000000240)=@abs, &(0x7f00000002c0)=0x6e, 0x80800) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) (async) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:47 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xcb16) getpid() getpid() getuid() getuid() 00:28:47 executing program 1: openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000000)='./binderfs2/binder-control\x00', 0x800, 0x0) bpf$MAP_CREATE(0x0, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) (async) bpf$MAP_CREATE(0x0, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) [ 1727.347550][ T5773] ERROR: device name not specified. 00:28:47 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x7}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:47 executing program 2: bpf$MAP_CREATE(0xa, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) 00:28:47 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xa00000000000000) 00:28:47 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r1 = socket$unix(0x1, 0x5, 0x0) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = getpid() r4 = getpid() r5 = getuid() r6 = getuid() socketpair$unix(0x1, 0x2, 0x0, &(0x7f0000000000)={0xffffffffffffffff}) sendmsg$unix(r7, &(0x7f0000007540)={0x0, 0x0, 0x0, 0x0, &(0x7f0000007500)=[@cred={{0x1c}}], 0x20}, 0x0) socketpair$unix(0x1, 0x5, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) r9 = getpid() sendmmsg$unix(r8, &(0x7f0000002ac0)=[{{&(0x7f0000000000)=@abs, 0x6e, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f00000005c0)=[@cred={{0x1c, 0x1, 0x2, {r9, 0x0, 0xee01}}}, @rights={{0x14, 0x1, 0x1, [0xffffffffffffffff]}}], 0x38}}], 0x2, 0x0) r10 = syz_open_dev$tty1(0xc, 0x4, 0x1) r11 = syz_open_dev$video4linux(&(0x7f0000000000), 0x0, 0x0) ioctl$VIDIOC_EXPBUF(r11, 0xc0405610, &(0x7f0000000180)) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r15 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r16 = getuid() r17 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r19 = geteuid() r20 = getpid() sendmmsg$unix(r13, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r14, r12, 0xffffffffffffffff, r12, r15]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r16, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r13, r17]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r18, r19}}}, @cred={{0x1c, 0x1, 0x2, {r20, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) r21 = fspick(0xffffffffffffff9c, &(0x7f0000009740)='./file0\x00', 0x1) r22 = openat$dlm_plock(0xffffffffffffff9c, &(0x7f0000009780), 0x80800, 0x0) r23 = openat$smackfs_ipv6host(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) write$smackfs_ipv6host(r23, 0x0, 0x9b) sendmmsg$unix(r1, &(0x7f0000009900)=[{{&(0x7f0000000000)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000005c0)=[{&(0x7f00000008c0)="fa37d126110b0896ca8e4cbaba406067c131b2e9e0ad5ac2bd8378a4523865d9919d2ac94f8deb2be1b52a56390465941b590cbb1a660d1b72cb56d5f1325779c64f435e8a6563595e59d518ef0ce94628446a8a823ad6db0614779a870a70108fb1599c1e47a9825efd5ad369c7e9ec5af43dbbe5ce40926e4072310392636ae16a8fe85e855bf04f4d5f9d555cd15a0168bbf4950337bcd29104410c46f2e02848da30753457c0794810bace066af24f791080321d332770e5d46a94c4854955ec5e331c714d67073c78120f18d1fc97a974a95c2cbc1d043abf8a549fff259d72628b2102d91c76e0005e970cab93f59fc8431160e38d8fa1fe373475fce4cb77bf907be2922e7403b81b4ebff75e76ae35d5a5e284b76e8e1f591504fce19e49398f7c6540b1c57f3df9ad6000085038e5ad8b0112de53ce1d970f35e043a7e1e0fd6defb9b2bbec77a9020dbf31ea9da2e1d6c6f431372e934c8c725eb6cea4121aa6dab87700983da7bcaf60bd0f5b0021d86552bd07360e00e09592dd373fbff45c1af62e2c0e7edfc84aff2d44fe12d0690964c57db055558e584345bdaf2b716e4bed0d329e572d1dfe07507123f3cf9121d4370fce9b57c2b36ee9ab468f6cf534f9ba06ad90b56d5b7550e0779e03bb86cd4899a03e34fe780c7b7cca7a12b3c2a32d4dbc8d87c2c1cd00e51f477674a31da27787f78771ebcf3dfbf87a0f1c28cd01e687302b876bfbf3ca8d4f228f0706c7be37395554ca5e9851d698627c98e52caa6d3168025036b6008fafd4d40d8fafe632c1d33e3baaeb34792ebcbcaadde90730305bee505fc3ddd7ce8d349ff7baae47081283bf72b968bcc3b3ecdf5dbfc084e94d8cecd920e8818df3c7019d75fb783d8fdf2288f15a392450afc4903472318c4805452aae3643fada5b693568f983e46f1120c3c67c885d9d36cc63c31f52c94462779e6b8a2839aeab4427f322aa38de378e0f9702cac011173aa433356aa2041eaba0c8dee901743a35c668fa79de9447d0c556bb6f60fd8916771c96f83e9390cfd30de438be1015eb7e76be51fe9aceb4a2f649bf50bcabcc6397eb8841c68ddcdd1eb70c79ad2c4f4f3ee6baf2c4ef0db550df29c16c94dd6f0d043749496e25c7f641c580f48f38b6e96669f886080176d329631ad0d7ad2a068fe79158dfe9ce9f8bfdbc3f1d9f376d8b056765d0ea352a3b9d684a6fc6ce6cbd666356d27b1ee8689e9168967196cb4ddfddff5289ce786921100a5808b262f53972ebf0e02ad9293fe6b051c2a2e40390c89c7de689d190174168c9141cf155702de67dd151ae631f6a34642bfe9af1b29f56b7c02e6d109d42e08964bc288b5f51f58d984b1ef303ec98776af8ab69446a839ab189d8079edd5e8e3063f3767958a7de47fe0a55ce40c906eedcf5cdeb4ecdecd41590bc10a92c5c03f8b114157f1b98d60513193c8fe1de8004736586df037afbc6896546786ee1260955eb7770601f645e95c87e2c7cf2612608ae426b47c27aa143c0bcd833b7326603a5f4f1ba48d5025bf370d98f74f550a18d876e0343af27ca98060cd867eaa11df5bb30efd1ccdeffb88abff851768e8da4f7bcb721311370bf00000e017b9e41397f47df508840cf0718f6e518e8b79895b70cb1269acc5f5258fca35a2facaa1533ebb7ecdda967a95ea57608c1cfca395d83b0a6ef8aba6a27d25713c1eabe1fdb002c4391e116ee9e900ae126dab7a1c9ef3117168aa55818539b5d089e6bdcf68c70c9c460c01c027838c7c4995c765ae969c7a20d7b92b36b779669b1fb479730b0458b40a0b9f6bf557183f61b5b4df4b7a9a76e07ddd94ea61c82b33add088bfdaf1bc3a47266bbb3a4a88e04cec90bb731e5bb1905287bb5936a7ebc00edd6fb179587867c7fccee1f35a92455a38ec06feacb4c1b432d855cd78c3b29ce4dfcdf21cddfd712b43928b72987d8e88020fd861cbe2879f653d6c08492c4ed412adde82755363ff58f46b76405ab32064fe1c75c693222e9ce9342ee4680270fec9bf53a585d0f971943919d37b07d98bca22208e40411a8153dd62e8c319cb3552bc5d7f258fc9068c8aa8d389efcda53b512094a38ffa8e9129389cff014b4df1242efb4e53fc8a55d00eda1fba5920a565296f0d75d6a414de3b5efcda341663b1950709afde7cfa29504a3aee7ede77173f8503dfbd648b92dab827aad42814ff5c6610ac972c510142d1d7e488b25af38f3cb4250c449e4089c057d2671eaff30ef11e746f282a46f5f4ca74fbac3cf2d26449626ac29f118e7bd3ad4bf279674fde588ab250c06604af2c109d2c8f7a5aa33254f80965f310c15fe48499f845b1ce53c55f5e9b5fcc833d5303503a47e2ce8110e72acc2acbb0af99dec4e25e4234fda0bb25dc3e3e47ccdee2a539bc2c0d25e69e8ff04e4af93f3e16c6a888022c06fc30afa8bcb64d284ed54d4873f3596f3590a91e467ecc6b7e5dc2bdcf5e111632fe4c1d1c92c1e86fdc5da79dea1420ce1e322d50aaa02d5e75c7fe7ffb93f4f7cd372031caeee4d5054e755c5d4bb986d54fa918c54247eaf143078528ae6eecd05b86e94e54c4b69de5a98da9a1f795e64d0cdaf800a130ab854ea0ff80fcdba888c0d7e65125f6690388c0cab82c453bd88ec556ba6e4d700ff4e10cddf3b557b1dace833420b2e8b69ce047cdabfd752428eba9b28fa4d0781e3bb4a1c99e003253eee35dbdf0a007a854074210db70b885334b9c920ac2c4c982c05a967defd92e26bd110ae885d5c38f1303e9308910e354f1e0e4b1bea0a7469a197e413a310b72e6c877b740169f0a53e4b396516e0ba105e28c0c51731bdaf946bd5e8e95148486c97f9be35bdf2385f13334dfd7804c010ac4696e960d0b436fe51a4fb27b472415b7f6ea4fea1f0e4bd51f891114931a9e08b00a680ad2db82db728d61da51ec5f2205903c942829dc3e697b7174bf863d6600049f19c888e805e4103c2d0ba517cafa6140ea78000cf6c38fcd670c36d32b6a1cccb52dd7d467fab225f1df1f81a8ed1bf717346600adb129e354b586baeef4eb0db5e956687a4329c6357bcd719708672b6b9d6868c6c5e2ae72ddaf3b30250f4ffab74515c68ff3aaff8b2806119da6ffb79be709bcf1407eeaac9290ce56daff4d136d6b8eab914976f8041e0171de4e2b01e9cc7e765a90341c3436ad0c41486c40c8a7aa02f4ecbdd83433358f5c8bcb0ea97a67a4895acfe04536535e60d5aa21588391fe6abc15be83dde9b70ea85229086956839a4ff0b6405353cecd82d7ec4b153815cf21ddcad042dcddfd52dbfe2c3ac86c2b9ba8491e57dd8fd84dfbf249300f2cb5889e17816ead325dac3cf3dca9a614264d8cc7bbd26630c25fc0e61bf4f4c40aa00f31fb1ba830a892291269815b3d3a6b3ec721577caa20237c3fd2d4ed8e0ca0dc930a4b718cea849b5f9b4dfe509026a12d7c7eb2429da5dc5b2e65347f5eb386593ed53199389800b56f716af258827046459653919d9bd0b469eabf14a31b3d7dd0b343b71d77ef84d1096903cfba2510aa83c2d1e9fc79dfa3b36a8ea8be8f1779f2275d03f99c6392e9832abb8a100974c63df854160e8b593ff8477da7718cf1401ee9242454bd2541b41b0f4e0107af788f9867b5feec4276ba80b495b2d78ec81a93e55c8d3bd0febb9973dac2d3ad6c4f2e51093dd544d999570130203f029de3b0dbb77171f37c74ef0435d4af2245d81bd86ab77063c1b3e5d69194b0ed63c95080ac7ecf31e1147de51f3329cff3dbb2a8b364f4a53503d89f04bd2b16436d0cd6c45b961c113e68ff816052d0705b761bf45d1447a1320e9cbd22564537691383882ddc681538c43cb5bfbd3c5a3f4e65fdb5e899d76c1d285a518fd7fe737f88a4d046c5179be907f7d0c428bfb83ec4c36135ed5b5c92d9f525f292419c2a0d5599e2486a795d2d2b8c3e52aa17269cfea0e189596976d5bb15380f8653e4f80bb1f595162138c65844041dcfa6f9d86137c228b3ed455486461525a0ea6dd682b7e8b8c0fd1510e8b1e71e3bc45bdc41ad55d37af1a07631a30716e2fb9bba2c17d74fac8e22a1c110656565ec448685d8c9f5856adf9b19e4687d0c0cd6b42a83bfe9ca455e25c757d504413286d35898b9d47a06d3afb2ea136cacc1c91584b1813f6da12f64374dcb72e3d3355ac2c6a4d62b3de3f19da8e1c17da2b92248f839bdfdb5585c2984de38ab782b73447e39d15ce933644be7aad8e49ae7d08f002f79e504a688672182d74498a39304e4a31bb295a889d49aeb0f9a3c92b35022fc73ebbbecb97b129261448429dfd9b0b8478ce9d800b374cb0384434fbeadb41052e45bc1f946015ace6317e180c8c02bd0542ac45403c624c76176b461b7f618cb91b7acff8996e098e882d45a50625b9c134f95c3a0003be93588c85578b5403f66a5bd3a963b27cd44eb4fb1c8a7610747f3fbda234e11b5744f1068eb42fe0bb9a07a1d9c2183d0fca4648cb316d636944b09f3d8aed69d1ee517e3608469a880038aae7296977fabdb22d24359f66ad8b4a25f9476505956a4beecdc44ba114b0a3219c63ccb3799997b13d9b96bfb71e38bf9e0da7242020bf8d7e213a02ae5dace52a765b79ea497339b3583a5647dcda3ab358247d634516f67feaaf0c44a1b9529ea376b5b6b00589700647a9db5fa5899ebd157c9da5807a1fdb618dd7041268dc060b11542d2a93d74864df7cf01741c5ac5d94941d2a096510f95336c3d22cc24c7fee47f903b87c2fde37428248192257c293235129029e442aac615707cb15fde52d5f05c15b7065a3e9998ed7b784fce4aae2129d34b63be4c773870ba542e276bd15c6e2771a0ad7ba1f90ef4f053b14734bb6a7ae7449bcfa723b96bf2a5cfbd02517ab16990b667e4aa3089bc5179ad7a6b47854eea2a99a2e8fb66737ff9dbb645ab2b34c5963a3f8c1a955c0e2ea9c734705da5a799a32640c0095341b5eeb37488df41609bf82a21c97a813cbac254607dacb85cc6b31305ad31eaf900f53ba979bc323e93711bf23e0d32d2dbf8e8fe1432919f3a2268531e46722eca0a033205db5c10378765b6b3a562757b3995b03ab3384d93f0230c6589db4ded9d2de1e575a57b042b47d904b707570fc7891112538bbacf61cf2004ee0cb4b7ed7db65ee269ea9c8a27afed7120361ad1e729401ab37b26ad22f56ad6c179aa9cbdc612baac1b8f8e90a1fcc4be1a42fecd5800cfef77c9790d69960e8e97444a3fc79fb4db599027e2e6c97d0bf2d1a8f28b29d03a1ed0e5ce9a676b1ff77946f255ab094df65b31e196e4c589b181bf4ff556041687b3a1f3e00638317d91ffa66ba90d00bc8dc047714ab3e2022852fda88c7c3fd5ca44782dac67fd17d43a7e3f934ff00243318bd7b80cc651496a5bd058568876e8b4bc163447c36fae13927279380c5afb35944bef08275ca10c0c0f3addbd7fc0527d4085065cc83d09ee2ce3d5f8cbf3087df0a010bf6d85b3d17f515d64406bc875537e3ea13eb26b329979b953f0256d27b978aec9de0c81b7d5ce00e02ff36a0a68c67b3180f73024c9c28d9f39b91e7c247977318742f3ee8076a90babb0c2facdd4c8f97609276e6d939892dd5b3d9131eb28240ac0868f5257c9991df4663d53638accea55b7a95ea0e0e6845d7a8eac15dfbaee4227682c28f66d0c32c26a524138f3b11d630bbe2cd7b4b40365aaf2e5069be2c4d6bb00e680419c8c24e3f0e", 0x1000}, {&(0x7f00000018c0)="1cf3ab47869f60ce64dbeeeb8b768c966c54e95c5c4d23a2aabf62a874e2b7cb02d0faff70b9d1c2eb12840d6bb9055e219654802c02edf95fe117664554259ed3b647094ee8c3e5967ea7446cb0643179a917a5bd99c22bf7f061e2b228b19c062b15d4b3b838c07c187caa350a1118d37ab5946cc0b489e128dc0360ffacd83e81fc3bda12a2117b76920be1e38f0acca8ce05f7d0acb3339ec1ad1be54b51f2460d5d3f62c140fd6f89c0f74dc10094d924880b12f43524f2a6d961fe7dcebe83e3cee44c1a435a81c062ad483c0d7bc4cd0e969545c0d31706fe982d749557afffd8ad6c511fc9ad37467270f5336afeacb1c7e0c4f7e3d10cf30f1639de03d049e4ac5c1e9e1b46f1bbd6df9204b6ab1c5de489404dddc1c494623c76032adc7a430cf02d8855c8d8d0f5cc7ed0e7f7cf4f60b1214dfd0a508470eed0de7e1cfd0e11fc0db5d563592b12847447230390d0d111e52089de01810beb82e174208c10dabd9ae75f10aedc627e8ec532abe91a5a1c40a74cc7dadfefb1400bf90da1dd28ec86633c3b6d9af3e27477e9ff42efe8b5e99f17a34b9a5c47af12db6a8b9c42565c6be3bb7dccae26db24ac305e037216a6c8fc84c4c0a4e261ad264702bb23a151f68cf66d3e634e9c5e96da0040aef24f197afb284dc960614b807e106a366f89e152bada4d960c45e9c4969806423d270c2bf8591492010755c6e1e585ea510d9aabbce42aa69257e5e66bc56f6046eefde47f1b27f383ad385f3040066cf2f1862485ab85f25dea5b26a15e3dd273c1f1b57b6efde3840890a717b2b21ddcfd618457b97ed4daacfc2beacd0c771516d57824a367fee820913ace4a167031e04afc78c98cf89782ef365d7ce85efd5f45a6c1b922bf1a4a91952cab59d32cde2bb9d0d54598f41062a24e8ecedf16e6e64553a1b3ef1881482d300990d8ac9b7f1ebee7b74bc17828bcf31048360482545c449d2ff55fc2e14fc20412bbabd723e1782efacab5fa3527dcdc6e7cd82b331f68f9ce3d1026dbe09315da525104ca900c060a30596b2ef34f83073be322bd7f1468f532e25e3830d2f3d768047e247bb5b9516254029b2fea69f197df86318c5196462924601669ca39a9bc7b89487fa5e2c8702f9128fc2d3fd79cb1b3c59d2b8444b266f2420a25790caa23928a9eecc2e06dd8b8728f53c97a92e1d4e132904bc9950eb8afcd75ce796995b3321c69e0794691ecd63c572c2f9fb07fbe0bc00d2bc96e9badcbb0a189d98c86b40d7e91aeef3eeb3188657c0c0bf57adc350a54453469d28f7e9f5d9fcbedccbc4a9f7eb3376460a44bbe01daa2bdcd245d1f40f2a3cea2ec494904ea06d1ece1650b2100ae8325f58d39dc5f2efb5223eabb581eab0bc0f342c9ba045b033b676e8d02ecb0b2880fda2283fe903bf01b6ddba76351e06c47997eecc65d974f0678fce0171b0e060b688fd565ee0fcd7aa940755d33c40228e232a58d92e9d51227fa8ffa94c582f7f0690ae89c0b940444a8f2ac29d27a12a60c9875472693058077781a51f84fb7d8a32cbc4ad130be83f80154e2742e7e5474d9b4cb9f7b7f9d17bbf15cecf49e1b92e52fbf579398b13790d3fdf3ca4c077d5d98229137a343b745ff89b3c9c14f2aed73fc9a7e3b307b7b7433f7cf93493b1899e80956fe33ef88291fe066fe794bacd5b5e1c344e0ed067d9c57f59b7396d4710b8207c99d27ac388a26bdd94ac63162e7651692a21e2ee01af814bb213992b08b0c260cf3c001f112bc6d6242d097991a41e981bb1040f20532fab540e3b25bab2bbb0eda5592f224c909283ba9c722cbdfebab4a162cd676def056cd6da19111c532aef3d73a3f132ef92a05bfa02265bd1cf57b20c8cd0bbba1628a06445c39a0d624104ab2230595006e612a14d3596ae3121ae86cabfcaf31808acb7b95c0c8f6ac733c4963a9e4105d69bd65709352d107d1345de415bc025d68543ba1aa49a0133d940000038e1cc67a4654fa4ad557d4cec7fbf8657cf6d4e7c93697978f251d28f25be94deb695ee612b5aa7abdea898e7eb91d38d890337348c4156a6913dfadc1674d0f882e32f2a0c574228a401289aa5375a625d38989eb4f0bef6d1a231033cd1d1f062a9d219d459ea505e95c164805355ae54beb029f706387e6fd34213ac4785665d6467442bd54197850182e1b4836c1c7b652713f94cdf3b7c051b76b16da599cd37d8fa9cccf8d64e91d6a4cf5e64ad3223cf3bfc3fc25a1258b827d0d798093ca7b850e04d4982abc44ffb8ae030471dd99ec4df0726aa8705c0ffbff88ba03ba990f5f3169a48d439fece89a7d27ab0df28f0f4bf85a6965f902612ef1aa98b4f2c7723cd8ad3f4de4bb30402ad37da1c3d312bc9b20dddbd74c47e24df1b533cf7713858069c84da81529473e5d7b6abdd3bfc7814af63c88a04a70150a2c77abc55977846dfe01c5173e946541d9d535af5038e706592e84cacfa9dc2970e723b3775f4ca66da94a9f8ca80014ee9974d839362ea7fc031c4c9e722516c0e4c536af7425444a2ade2227b760cf4b014adc48b3c9d72042ed40da53d25c7ea00018a3a27c320a49636fe54c30d36e58f5b6225957dcb370dcdd6b3d08a3b836640c8a59de3e0e23e09cc37012ab7a490b5e3769b0005339979992abd2cc0d9a1843b370bb525654ba723bd17667aa598806b0f0432fd73c36772728adac21d512b23031b1b6765ee019ca2f3aac4bfb3ff80bcf767983635e2f6d5345c0d3eb02eee8e5150503f0f5c0b8f3defd2235bd9bf2a845496074c1e55f8cf1226dd9d37b39305ad41cfe0dbcfe2f85d20c8adf0001fa22cd0e113e59183f97f45ea0507b034a83bd5d2efd4cc5c80f582044ceece90ef7dbcb8ad13dfb76f6ec7e4d80878de4b1eb9185929b05bcaafea6f3580dfbc5df8e582ad89f8975c6e9fd7df4eaa2bdd6541833cc979350faf830072aeb67ae91ea4c2d5a963d026b37fe0b7a93fcc0ff5dc5108725e3a197eeefa338037deee0c43c43441a603bc67881c1bd4d1c8b8229506c67b2e1c86c9a825f0b57d781647313a70b2688162b133b27e931d37f7da718f4a7836d0d7381ce5dc9d1752d5652db02e15e2743afcd77c7804dfde067a135c76316dd7057b79d6dd850df5a50bd7660fff1bf41ade2b7bb441dcd5c4dd38cff60aa7fddb49c2f6069e2c7328ac43915006febefda1e86ff51523dd9a8611af2cc6e9942121595217362389f6a6199347a452e13b9c48cefbbe2716a78bf2973a0fb8439d372573c20404e0f2ea8281a26dc0c1bc6d67552b9b1857f32dd8f1e421634efb6a0bf332d44238ac330839dfd9c4070796ae053bba14ab354ff9397cde73f3e175ad072e4b3ca2a68fb60e25c269709b758d2b4dae878a2322bef1a656cd2c10b127acc9d0601e663a36b3dac4205b424eb3a8377affa53b36ebc7617f172b9ea9a6b30de0b8b2a5b3118f5d213c076b12bb7336a542a546b6576bc9fa5ce3aee2a168226b7b6aee0bbdacc61f5c1b03d6551b5b08cc09baca71130d23bc336a9573b3fc2f50e108c2410c37788d32d89d57242129a1378e801231a2fca9e259c2619c6f85d22b3532cc5782239c2df02e50198105b484da56e518ff533380d10dbfa4a8aae8bb7b3d833eb590088755c414b2ab85970a9e34aa636d7240b91edd3605552bbaba74d41efce775a55cce314400fae308ea647205e80e9c3644fc999afaf3fa5b0418f1f8a65d4c85d352a95a750f972977b16fa81f37e61e241075d4c2bc2d5ad57318b55c8baefe198aa088f6eb63fc3d73a7bd737d43f9200ad55f3656fdf5d8c3d009c170d54f3def0e9a6e5c3dacc8b8720429f1bb5b45864a0cff037e7420b18a7a8bad6ffe095b270ebd3d0424339d73027417e42be2e4ac0c45d0d880d240ebf4f1d575b52ee49e47f247195748cb8a518b746e49647ae58323c1388c76acbc4040578560d4df314ed9913f63d895488f3288f00aa1a96a053d9e8efb5990b1cce5180b475e6efc8973f68580bd88b40d89495d8dd3c128ba2b50c7f8bfb6d1c5fa6a5c214e6af74d2b2a78f9454096734bb3d119d68e17836b371e7d0f908e055d9c518d71e1455896b91b10f2801b9932b60580231a781d1c8904881c6233e14a8cf8a7ca9187cb005dc4755e035c1a998deefde9839693cbcb2d9c98707bd18d5d315af0b6b8da8ae06b993b28bd19769ff4c920a39c05b8a998e053486e8029ec6b2be6bb03755a9d8af8a743608c20e26a8c0eba72d3c88d9d128a6dc426dd38484f7e31985433c0c08dcc193cb35660eca143c88b5155e30fdc0d0bed09aebcb065a4772658b9c903edc33cf4180a3aa4790fc7709971350caa41d452e55551e50b0e87fe8550365a98a62401d4da18c43eb1e3e2b085a34766897fc9b034f1c6806671185ceb55020f774fd90378f7143e8b10cc93aca0fca42e880cee056f5fff19f4de211470fba3476675a704b0cfe26e7fa0ea0141a1c68c30034a4ec2be8ce5e72f3ac892101f16411ca74ebcddb1c6a7c79d102c088efa221259021f234bbc1ad7978ebed418011a7663e0299386578a51a3ce853b6fb23d116366c325c82bda0e11055fc31091c2e08db8000a9f4c6c9fd62177f8507bb008cd07dfd08416bf4369bf2c89da96321b47a42131685a4380ceab21b699419c22fe243bb02a904252f9bd938fc8847f10070b6a07f5fbfe74255c383d249f5da78fb57a38adc436c63dd675694037b4e6c70d497dc7efba84345638b73ce9821903ba4fcde323403bd5079efa902213c15534934b2afb4c044c1e13e188ac3380195afe57f91dbc5f9624ec04d54977073f77056a01a8740a1bcf936c553409236acf0ff81e415af914e086c25f4f303d9c839c50924e1ea15a9b909adfc5e8171711aaee2c1ccde4f2c799155dcf80a55e0cd328f9c9d3ae42dead2be3a89f2fd3f031a70567ace2e85ce9963dffddc654197c4348e2a3f72a9a19d7df72d5a42b1ab0b7d520d427a023e5ed4653d7da924af6cdba07ad530f34ecc9224b1203a894e58c322f032e1baf16caf1c20eb5999e2904e0de2c581b934673e300f30f39f36685feb0ff73ceb2e720e477ffe9d86220063768370a9ebb7397f3ae0bebc3ff4588fbf9881ec80516b6834616357f7496e111bbbe9181b8077af79c0f1d7001e508ffa221a4ce1511f1a4b5a04b9a540f7767b09e55c0db318646b99c6c6c915c00788997095a421a3c028945ddfd83834eaa02241b8eb827fd4bdf18f0728e4a2cb703666b8caf9d90121f9efb800e35d9693ce9519193b44d43d5cc6ebabd6cd242fea7f4b2b3da50635a2e462484c01405851886f439fea0227e03180526d7f9392001eebfccaa28fe209f8cf2e7e8727e7cf87147cc3c64d7687620dca17e4289c4ef44e1be18f4da3cc185e3c2286be3bc70a6ec90ea868d2ee3622c6cc3e120badc3d6d7b666715a5e180c749fdf5d25b50c59546ab37ac9470bcf1a832a42cf6bba0c9cb454e3e25c9438ff520bdd34f8a46c9e9bbb0082f691469ab24c171541ea85f2547cfff76ac1628da19eaf4ee1646f0f6424681026d3ff1148cbd3b9e86cc2c464471ce0b905379a60f23f0231ddb5f93095715849249883d517661fa3df9131db2b82f5fa06e3505c97527f317b289a081164c45afb8568e662ef7259c7f767d989986caed0bd5b0277e7ac6f8b3b446102f847c7d9905348cee65608e2d778afc5dff62bb9f9ba1776a4757c16c7a40578", 0x1000}, {&(0x7f0000000080)="7a69e9dc6049f3c61eb38e0e19a34580dc43dafda681c5d6322138ed9fc9d696ca57625793dd380af6903c63728765ae5378b98a00cc3d4a1f6c7774345e9fb5a1c9b77f1c326b682f0f98b23f19b747f0307e875fcbbda49bae5a6846789503971e8657", 0x64}, {&(0x7f0000000100)="66d95f5d0b4ad7206d502b282a0159f9dac3a4b95992ad0f285e6099f6008abf461e509d82736be80ec45a2212ac41572a915879fde382fd9e75452cf3f0b1d7841d5d076598c9d3bca11ce2f2e6332109bbfb11a98ba543e7d20702a1eb9174997d5c7ab9235dbfc814a1c5605cdef858b6d5ab803ef431327599d21c09ff811c05305de8212a8da4a7b1c987336e0897c0ce59f78482d08632e1de5ce2e7dbdb35e1c07e079998dd1d7a4c4773f41788df37efd2e257ff0c6b8819a721d66c45215d5fddf4d46cbaf913a46497", 0xce}, {&(0x7f0000000200)="7a3a3027669a68962ea5e5391e93cd5b361c37190316679784cf473fa60402f27dacd35f00b5a11954bb4763affb19b1f2b78fe9b6ea44ad62d3410d1551f5d8d5a26e05731cc5568f16f7673d0c07bdb09a5a1fcbed0c2734d3ff7954263fcca072bb51fc3cb2f4dddede4c547bac12362d0ccde1c6de92d7749837d124aa7cf0bbf9c2651ac065033ccf2e0afc740c8ead508ea9830f044a5fa9a0513a73d9098cc19e22338393c826e9903316c7651a7a39f29c1ead376a0ffac55c0c305fd1", 0xc1}, {&(0x7f0000000300)="597a74b23d2b86e0ee5312e828f1aa8579797dd769833e76ae62185704cfead2e3cbf494628560af36e54edf804711786b4f05d0c9c78b908fce46f88166bf61d63b9c454a4c5b724aeef31198f2e9f5065cc52190e8bd5bde9310e422db21afb8f17be1016eaca8889acb20a50cbaec0246367d5c2fb8eee475d14bd32aa42fb38c800f05c52ad4eb4d5cec691106e08dd9cb19ce49a1f8347b572629280f8e7e6aa2aef0ec521cacbf8f6fdb3825fa6bb384a3245c663d17111f48c19196524676ca76b05eccae64a08a83db4dc7449a868811912c67d08804d638b04807e23bb8ab6e64e456a92a1080922aa70e34090b", 0xf2}, {&(0x7f0000000400)="ea3895d28b4058bd94781c7499bd237c9603ea7733914ebdd964fabc2d3716ce201d6d14105ccde43a9b362bbc614fdf6b242a0723fc95b37e890d2fa294083528e5bd8d2f3d04b42bc7506c9c827a699ce3959010ea4c28de4ee08c63e043f613d48fe15385b3e0f2e96d78d7e2f7036f459f42ba6f45c5e359284851cf985c4eb51f9ca022ab5894df9d430252a20676d7dac450baf3a9320899041f477922a929c2bf190ed62699396c7bf670f9bfb483c08b02d68d43604113a5e502277d382bf921aeb10d1bae9825b00ccdd85e6cbf163831c311ff517a4c3c8b242d49f16ce2005a795f2649533b6060", 0xed}, {&(0x7f0000000500)="8894a5dc8fc187d20a68d2d047dfe041645dabd06adc2004e6f90a71bcff1f26cb378c117e9e7786", 0x28}, {&(0x7f0000000540)="e3532b949814cdd50f21b0810ab69cca147d923b529d4e2769881cea7b34c844b93bc0acaf880c75647bfd0622741cab41b3355d2e1fb919d35495587fed1d636f1ad4ec8a63613fa2c170bc4b8ad1a85b01161e6f7df4c2e2f3", 0x5a}], 0x9, &(0x7f00000006c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {0x0, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [0xffffffffffffffff, r0]}}], 0x78, 0x4}}, {{&(0x7f0000000740)=@abs={0x1, 0x0, 0x4e21}, 0x6e, &(0x7f0000003a00)=[{&(0x7f00000007c0)="352785b775b70002deb8fa0b00015e2272bad043866bda538535b59fec375bec4e4ff475aea666872ae945f99e3fbb7700923cdd07d5c2394421fce46504a01ce938e40dc65a24e530dbbfcdaa820ea80de6fc9dc287a88ce3c93e8ea76a38b171964caca6a8eade3da9100ba36445388859", 0x72}, {&(0x7f00000028c0)="7dc3ffd76b3d30468ce9dd4347e3730944956066ba6dce166224360001f2ef3cd5f3ff11b00a9dc200a9e5aab72cc5ed95d5470c70ab880b155707ae7ee9480b289115ec4350a6834ae1696081", 0x4d}, {&(0x7f0000002940)="e05a03afeb48212e01b58efa94f8faae255dc7dbfd3d2921e1a9fb40cc23f8b6f9db2167ec5fb3d94976f200db93c9a49648ea5956407c414d0f596d676dd147a7a2c72787d03c3eb9af592304381f71795b8755fcb49fdfb99f94b0ebf6c59a27422ade90db42b9974b5fe939c86714717d9f156385597b2afb9474c93ebe", 0x7f}, {&(0x7f00000029c0)="637d1ecf2ea1ecce9c77bc", 0xb}, {&(0x7f0000002a00)="c7b5a7d2ae7a3930655e7107eae6c3e0ee94ae88c430a7d7b0f7ac66bbfc6e46d4675ad186a616b83d65fdad901ae83a3afbddcdaaebe408c0de1a7ee7461758f257877d4bdecc27372aceb1b0d6156c1feb4603ef01e6d1d8ead56408a8a4008130088ad09a2a5c19177f29017191871e3ef6262e559997cebf83c3c4e2421375e8093d5cb61e8c7c86c41d5d5d67dc1376aa8fec686003776ccdb3a891bf3a063a0d3839b602f00d684ff21985822c18da2df541080946551b1c485b58dec28aaf1ffefc7ce2d80a7d5522a7d2acb830fbe41991deb9700b7df8d07c94f788db83cbda1188cabda70cc997f337578662a0a4698c3c03dbd8373850015f23c49447ca4434197e7c4a5f04c4fd73e3e60eca9460ccb97b3a8f88d12cd82d6c2eba411ae33e641cdf329c46fb5a31aa15ad5b92441780445674b46590ea6e22befd743ffeca8f50e8a9dcc85263f40d717cd788623d19894e36e037b4ad4e999c14accf0f417da8500a895c6bb1b44e4deb5f6eb29bdc3d8ea252a0d8291cbe0e4a1d1661645e9b2580ab02cba738eb87a8d54ec3aa2bcd81b0b386c60f4d993cd012885a9a8a2a61ac5c1785e06d7e550f4d477ee3c2c218471517cddb6291abbd2ec77146e78b5327be7b9de611d7bf1c8e5e0a461daae48e9877ea1aa280d116f7d6e257bbff9a1907bec4f4744bf064fdebb839c5621433990ca52abcdcedcf4cf9faa9087d51eb32a59fbed2a6cb665b6e52157058382ca48209fc1ce45c41961f90146c9935da2439bc229dc79371e520aabf801eb42a928680c4c1451c924c71f5c971fb6e5108292fcea8e3aeca60f25ecdf9303cb7465384c09203d42724aa37dd7749eeaca8e7219a702e8730c334933fe9545848b6ef18681ffb22b44531f458b1947bf4a0ff30f31257f90928ebc14d48675b049a2aea12be90ddde30539aa2775204ab7e64379c24bab9281616775b805c024b81c11f2d846eff8618d817180e1fdc30c63e183f6de9afe9b455f5c6e07e5a12ab823f7f28254da826ccbac164a9276a53b5d547817f9348c087c34ab8be7707e7ba6655b3af522b7135246f3cb5707c828b252aade9f4c8fc423a03b24fd36c077a00f89d94317a85596bcef84962b581c72adba8d5155726b0b16ecedda5b2d2f4b5511926a3163cbf3233893ec5c0f42979b71b48a8bc36b5fd31cd4cd3a68a582a8ef26c78f594a426774c767b9cbc0a51f7a378119ebb8477f159b2d8767ed4bf4f079ee9c94ba772f85bc887f58eb24a491334a87e468a48749c08082f5a30e99f15d0546e4926b6509d311037cded8dde1f300db5679391b42bc689e5e332128536c1ce66f82557f404d0668ba8ed72d47cd1595ff4ca73f3569dc12fd09f6735fe7b34cd0016b9546206ff0615762a786bafce23f7e977e58ee659dbf119c523e1be4fa6d01b214d74196bd31e6c9bd4d409bf9f72f019e73155fc8593506788bb58a6238f6e89228f067692e7a56112d05db1c0d990a08879dcd8fa2d6ce2cc5c7aec0c2e08076d6e68f6ad7197e92a9d0a8aa101f5f9f7969e715ecc995436930917f4b6315e06af1e38064b2a17fd35bed2d364e7b997cf85a2201ec5076ddc00baa7be4b99e38b8e8cfcaf3af75d179b0fcee9f567f3160ff3197b0e770c8f2fc7f5ef3c53b96f0b59115989ff9282c49bb460a7f4af81dae9578b51ae1fe2f457325f9f7108fd1a7231cd05cd82b4bdf2929a8b6541495d181d7bc05592cf9f869f90e8222588d191612e01937a8df8855cb8c16302bbbd0b9930ff89e909498efe76928bee389391d73f29b6eadd9be765bc4776a1a97665ec953e9a1f42d48fd778890927060e85c3c4ac90d4716a415a7884356aa0e9192115272589371d05d78c94520532a0472bb002b3644bf0c2ac94f252c02be56fde68d7ed9cdaef35fcd7266435ad071af4e9f2c4011f88d11436f1351155828b40814b22b4916638582dcfb6b1e0323a24e849a1b043f158e05770ba67c030415ff1658c1183bbb14c8e1c29f131dbcfe843f4c6925437ecfeaaed6e200643b1e46a8418f2fefaba3059871f4125fc76a562dc495298f39312b6d790cabad691f1068d54d4209d863d390d83c00ad052a7d0c50969c7781dcc29d95dbf2d9101ec9218a0a84be266cf826f110033ccf5d887bddea81916fff6910846b7b4c279750a51afecabab01df8b143c7b7210020eda081216f19c3383a2ec7d60c4c83d55632ce28e50a537427131d99770a0ef2f983fa091390540fdc1047ae0eb8c7c3945762595229b9693c7b53fc7da97b98e4b03acc1a6998fa7ecf00e958aa98cae45517a8059c6ab6083c014c418cc1473e5e25df311432711e9cd582f8eb92c6890ca16fd2a2ffe87e95fde28d0b7ddb2f7da90291c0e5cb3418713bf8fa33a9c82972b5a9b6c2ba33009f31d3f3f49fc0e81814db0d1ce1e17121eefc6c81d956a694a2552529b525e50ba1b5947d3412bfeeb6b70306c7033a0009ad0fee9ceb547d299998f6d0d087d92cdca87a25b7591aa236de697c677771f93db1c23802a35dcc09d71485f95021044d227f4477ad8d463a99d3e9b6bf05685d88b2d7dce309ae2b5773cc46cc4110632562a6c56f6a67086d501f003d43494557c391f2e2129a9dac704b78f4df2b3747d64bb62fc6002300a0153db3e57e93e3f844ecce523ee9635710831f61815a40f6d505bd525ebeb9b2024e5b82ab4ad31ed2a9b0a6b9510171ab3a63bd38bd0d5e5dd6a442e62ededdc92793c20090d4096458ae4776a19510a34daa6bdc0c6a997238e46994ae36d816295651b48a0cb7fca48313613b4e82afbeecdc4ff2101ec05de0b1a8d17aa33f3c318b77b4c9aa7c5a6b7106f27217dfa8c5c913275a0e62dc37cd927d0858d18fa769d99107b6b0ddf8d84a91e2716f04e96b8458f51c1855d738775b43083ff3626cdcb1a6318fad4ceadf50130db78a8efb161552b33fd43d899615bd344669dc9833f983342bc84f79ff152eda4846015d567604a54c1dc0dca43bfd9be95bfd604a775edf26e0d11e80cae32ea8d6817b9caf9404d5e57dffc148bc9b913f4af57dddde165d07d3e63759122dd9256be4f9b549d6186693b9ba59a5115dbb7201bbe174c0f2a5490da51fd8481a348b803faa1c7cee14936a61e03a9f9f725e93b476a5358dc7408c6f2108ecd0422a9293314b43e64aca5244756dbfa13c946fd53b84e697dd0f40844bd15321c53a14088332a4c9ea037e12c7d84d5bf829f1c20d5af3c68c4eed67593fc29be21ed94ddd85a787a8e60da347bec452d036c867e0dade62befa899dac74f82e14ff1412613eea444f9c03a1d256da00e19a989815f3ff9e844157132f0c9e04de330974dc73b8eb84e2f1a0cd8eab31bfb7fa8b08484cb7ef7b9ce67b408395398d01ad7cf69346c47e27592aaedcd557dbe9c925626be1c8ef1c1b18cc88e26d705b58933074dbcc723aeea4a9313e1668a52a5178ebc8fa793e5766c9e09a529aad38f3c7b087226b4930a5cfed85ce2a0a1423925169bc76c253c5474b12bed3dfee057d8a4b3b19af4b18e31b2b3bc4d5ed76502deda6f183a345f718fa04e9dd1ce289d2cd4c1fe57f63299a7ab5973ab4ec8c970d68217a1fa7e87b8b0653f60e3ca6cd3caaf5d39b9e2155fbf233c3634a85724c0abc4738f1da5e833fbc86ba87ff9b2c5370935b085fb4246fa076c8ec0bab34bfc092f613fa3682c6aaadbff22466cbc651e3d42b87ee1dbc481caa7c932486e45d2d93833eeda01b2b3c43801450a28f27f2e829838076f79659c353ae764093bf4d776c3856576442f20f2be892402f428531c85d02e2bf05a454dbc76c3d2b3661501d7ef1773f63bf1b3ef6c50999d72d6a4502b34c97b2d52487d2eea16830fc1413c468c41739c6a9b8387836d7ac380b0060d87a5da451adbc1bda921f1150f85d46cc308710ce8b374c68848160bbdd586744d47c16a5be91a38c880649c50547bb33cef645c6d4c1697abb90c2936a5a4777edd99a187cb37224948d276504948b01982684d704be9e7bcd166cf0cca6a18b8b27af81b23f7e97782ba404ef2c5b71896e73a8f9fb80d585383dd35ade61562eacc87cbb1af5eae31aa3c94d13fe13c03c9585fa20241acd6ce20065425791cde6e3932d1e5e871f9d08b59f6d21b1a5760bb8caf23df2c5dc97511fde463937e20ceb17f32b7726818665a44e6661e125c87a80c884b2e1202ce72eeeb8b0b13f29feb7537baeb47536cda8bad897ecd99950421b2b19a31f8f9a26b2e0a9b0900102ec0c142a4125f3b750d7994335faa8fce4ee30f797045a0ff5a19e84d2a92c4b10042315630dcfb822a748f1976f194fa976423aaef35f7a6ac5354101f8f069fa7712c32722532668a1955a26c1b9781dea3dbd6fbe3c728de7bfb63d6cfc724d54ec8c22ceaa1046c3bf9b8074f4cc744c123bd39811136818c6d6271099289586928e3db6b24d8b3349ec5c72a2d6a961d83295e8fe3f677e69ea1603f008c08fa2602589b42af23f90ead18a6d9e2b1cb98ada803063cdef11561145fcc494c69d62b2866f9d750a04c90fe29fbced3c59f74d78755a912c3c741e294262f723ed0b973e4db06a6dcc68dbec6b7d7ecbb41e3dec682ee72c3b5f9f83336cb3f0916bad625cf7065f19481e0392003ec49e544993ab258fadea7c023baf0dd8bc4a7e12cd65ba2d749315c44658478a16c92cbf3a2b2a0d355a85995a443dc77c64f3a2efff639e3f3df9584bbfaa1222ff599ac35a6e36c97a1fa7db13573439c4cfdc811f24dfd4b3f04d5e3b87a999bd610d838a8768be189407971ab35e27c8aa421d5d265b235c95dc7cd77de9fef14e2f004a1c97c2bb37c5e68b1f8a1b967ccfcd4e02d3e27002009aeaf4b75efb7f86927a70c1a3bb2c02711c2fdd99b421a67e951f9019609ab484fef6c0259c12aae069ca6bd6602375d85f5ad7d3274d11e131b26e8f4ffca0b3af3c931b4eba78894ea2f2c1ff9a5fbf8f888bdd42d3bb4692078c18ddd0c59b3bfe5ee1e3573844b37cc54c98412a1afea9f36a4bb4e5dff95f73e60fcba1dab62fc65d6f2b9e4fa49498d6255444fe350faaf5c58da1e5f7d3d4d5cd476f32ba9658f2961803227a16268ee8fe99594eabb6364d307f8e8abb35c01cb00a50c7f83575215bc4268114c66e59869979995b8bccbf107ecd3bad2eb1afaff320e6b439a052679d9197c7fb90745e8edcd6d1e50cc218607226a2be3d2ce2097f6ea909c7d8f62afc3fcbfa93a8bd6efd0d3849a3db2e4e30a2c144813f4b91074d806bffd587b9c7920d3856ff97bb5dad847d52b3d9d95d2df50da5c2bca618af08bca015bd2600209930d8695600a0bc6fbc458b64342cb440e293acd114cd5daaa545582d63ebba5ff6ad6eef26c26fd0c4fc54a5eda7d23886ca159833b9d7dbe65d51893d5b2a9a7ba06a2de4e43dad43b3690669dbf01ef2654ef140b06bb3db6f14036a1a04eef010255cca929f6e86ea5e337a602980653fc6f7022b598d70448a86c42ab41a59104c1efbb95e90b47fe3442939e0ae8969741dccf1c1136254f96e5aa9f738e28947390a376ebe309e93f5578d0535f72989cc9a4e76e8f4e4566cebcad06d7c384cc4bd1cd6a2d2ce87a5773a110f90ae83e60baa74b6c20704e8a1c8b81b7f5d1546e2bbbaaba8249a17a506fdfd2084fccc0492cb97837e90d2ea1379c5c7b8ed98a321383924432cd57a7403048bb0da69f", 0x1000}], 0x5, 0x0, 0x0, 0x80}}, {{&(0x7f0000003a80)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003fc0)=[{&(0x7f0000003b00)="93ff6611eff8657623e1ee6fd9b708304afbf0d9a637a6f1d9919cbad0ad18410d0b8827dd19fb88ef2e0f20f9eb0c038e7812e32df4a0c119107cdf0faf80680ee407a3196883ddceecd0d80f111f0d55ce7bc097066cd7a28006", 0x5b}, {&(0x7f0000003b80)="9a1d914fe13ab7183ddce7984aecdc90c01f16c1937d29b54527ddc0eafc0a67bb4cdb7a0c6b69727c1a33bb4e0182325e8ea3d1f3214c5fb6676062d98e8572585151c17f42a5bdc084ee77dfa252d6fa4f60073712783b5dd75e6cd2f0d0877d97ff44c0a9c37b94ccee98d94d37b98db64a0945587014510c627f2e4aae71801ca31f0fd4947c1ba4ffd98346e3966fe13ff17a9d60332a97fe20723f950b69fb1491b09e120c63df633b168eef37121aa2682ba3ae84cb44c6fe3e19393a42f9bc15fc4cab073ab085e617321adbb8df3dd1633f6a20e43f961869dc31", 0xdf}, {&(0x7f0000003c80)="aee2cd94b96bb045c083e66bd17bd57d745ea95102f3e5a7653b8e295b4aaf9dcbbaa494546db8157f5d86a86c0d2274f2cd8572a155d73fcc0eed3a00e7967b5dccb3d3b4ca559416dec340c9b390b1354a1ff0a21074152186b8bd72a4e7f766a81a9c8e8e0af80cf2669b7e4007415c99bfb384385ccb6d850f217f21f20b858556cef723dbd1658600468074fd687a082efffe3b4be72c61317617dba1fc45994eab262cc257e42504fb96803eea4ad0088a2c081e2b9e09c2c1a69d3ebb4e5f62c9287cbfab4ef3f1c7aec0c713b597da5219637b42b26c8ae24f3863cb93d95238d14f", 0xe6}, {&(0x7f0000003d80)="428c17a36c4ec58b0147c9a453486af940f4bed12995101c294db45fd60abc9de1f464d15454929af4793d7e79fc65ed56f769a312d6687dda6f994a676a098854ab67e1647bf998bc934bcbfa5e92a61b175756844f30807bf4638c9cddc76ddb2ee7250ceb25f370b0e0c8548c6fa493bed2f0491b9f963dee395825ed4301f15fc7cf91b23cc5e268b1a2fa65ad781380257a185e99913ed377ccc1154364b7", 0xa1}, {&(0x7f0000003e40)="3f4108cd190380833d4ab89250830aad3bc1d9e2867398717b567ea37a63b3a8766a44ce38ba9b0add9423b9bdb4896deed28fabc912ae47ec8d0da27dac1498a0ed9ec28671da0486541d1720e3c24e2b57c65875409ecdad306ada8cb9c19cddeeaed5d832c34d235c1ee6cf8d3b882489f2e59ad5dc46bff747447c86d855410bd40197c6aee8896d09dd102089703b5068244a938c4ae1585fa003e4e7a0b52d79d1a87e1d1e15ab3108c66ee4a2235d", 0xb2}, {&(0x7f0000003f00)="083f5206abe479be4e32320f648c02e3bb5f8371db74dfd6b8602de5f39908c105694c789a4cf28bff8f4dc5c790fd408ff02d3ecd537e901bba127e2f7df76a76bc1e5b0fe40afebbec329c8d516616b174cca300e15251f271d249bb9fd581ecb74bd9d1619755914dbe3dedf723027716a8e69077055b38d558906dd0ece9d5482915d5a015205833ca75302e296b7d96e6a3b8", 0x95}], 0x6, &(0x7f0000004040)=[@rights={{0x2c, 0x1, 0x1, [r0, r2, 0xffffffffffffffff, r0, r2, r2, r1]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5}}}], 0x50, 0x2000000}}, {{&(0x7f00000040c0)=@abs={0x0, 0x0, 0x4e24}, 0x6e, &(0x7f0000006340)=[{&(0x7f0000004140)="b2ac54f987983f553cec73a48162d568c71497b2a34a0964410ae45af25d2ee5262974bb9298de01b5981506", 0x2c}, {&(0x7f0000004180)="7481ec864b88325813ba249f004091eb2fdce6fcb3e6740cc6eb33ef5494155e36bf889369ae354303d674d8565cf44e3ce51074740b5c9fb6750f7f14bbdc65396c9f8a11dd62c8939c4bed90f2b53886412bd146556aef9d99af830a360fb7f71d7599caf45580a1ec09716a4e30fd6880c390f1dff8c7b0ae3e725b79cd273dd936593c0cd4587f83a5b801f898549982af796c30c996d008ed73cb2c43c97e122ac0875daaf010160040c5", 0xad}, {&(0x7f0000004240)}, {&(0x7f0000004280)="fe692724ea2acaf9c9003924deec4519a92fec2fe89ff751e084885c5e07c26ec0f872ee0d5da8a8ff66509dd0d91bf3954468c546fa2d7bd1eb3edc727c9f59023dce5c36a79c2e0fa10be91e795f57d3384983b2252fe3c7ec74d8c544a266d2b24cab396845c9efbdd72affe93e7f8b59596fa8d8e32e73c48412c7df3d54006fbc0dbc2160268e0c57695e91c9f45cde31", 0x93}, {&(0x7f0000004340)="7b4d4f3ae789a3a473e122f6e29c07b80991a15caf3b678623fc4add596d5098171c0257bf4f47db1290adcfa0c40cf6091171e9d66b47e07b7f5411a6c08114a2b4c759934fce4d141961b69fe3c94789e6a5f46171de31f6434cc658d087d437fdc6d3e6c5f3213fbc7df8bb1434d736b7fb63ebacaa93f67c01e586f2dc3ac4a6011c82e6f8c7f679df69ce52e4f9f9fc3957039c06d051f09e14f0de96425df93d3ea593ba8049fb7ba281fa82ae2b661dbb33530e864a3dd3cb09900da46bb7773ff4dfd6cdb9ad0b7d20845420f5ddcb2cbbae0b6fc949f4684267677c0bef29db926556b296927912f298c5e4c457d2f4275a7a8ef4d506a4bac672fd2b45818169864f84bb3eb0ef760f1b100c9054b50895fd2490f92caa283bb0c2dda891e1ee9a6a446d1faf315266333ccb0670966c1e850664a6fa5c04e4c8e26ae168cdde25a72df1f0f3606e4676f72322c88235f8883bec9e2b3e1967766593767798aa99f8205dc889a230476e2083bb07fde29dbc1097d887724d3cb520e0a8224994e8152a57876ee1dd4ea38a7836027be9aa24d2fd4fb5ec3e26d1689eb2fd31ef57368980cb5ccedd87af9f30d1c469816e12ccab0982d94d720a99ae6c2b9c1206ea401948a7faeff3bad89e35b8b063d3cb1be663dcf8ac08234c71d662538819b5cedd894166574fc50f250b567a78633f72a52bd7e1d73893f260954311d19ad71ae7f612d500f06d0e2c3be68d6a88ffafdc8e76f97229344bf3d47dc88dcedaa08b62ac795562a4f68bcc66680ced83a9c3f97908a0fb3e176a3ecbf5a933fdc35a6415b669c1052809f9edbf373db771214dce247ac8baf144725cb00cd89cfc1c0e940ddb20207ff734d326d411201b1ca7a6e30f81df853d8512c74c2979cfa144a727365a32bf315191f9b5dcc990aaab9914a25e62463e409095cbb8f738eb5f0654cf086e5a4f44a75ea3ce882ccccc91b59f645be79cf7a4ab2e7490518928023eed82494150a9cf1648bc9ae1bd1b2e2eedcc2e7e7e9f513a975c9bc195c6ccab4a91011d48ab2e65f08fea07be7c9d3245ee9223086249fb0e33e4886e21ce58db5ae6bf587d507665df202157bac11b650e95d9c5a37ada6c4155b95ce80b977dc74a2a29be003876993da3ea7fcbeee1bcdeb0e5b5a7f9ce639f193c3c0b6ba166df026e3d33f24f24c2f1986109cf7ddbbf316d8b7a574687c272b0d13c453496d9c782982865450afd2cf549b9bd498ce3fd288916c1c03f558e3bccd8f84942bdab8cae1c7fe5d230dd759ba1ac0bfe3c9da0991f701e5d09ba809e0b075a1855866807c2c58e54f4e4d7179e1fbc466d01d1c6dd51f89beee5f5566fc2fc0d85dea3a4f48749eb85272a119a4f7ba2a22aea72f363ff5de7f1682bfc0b7c9056a7b5811b3be5bc2e99114ab20bd21cf9a0cf46dd1dbc27196ad31b4b7bbaf48dc2f4cdcbcb6920008f92f0fa1bcae6f3952b8419637d4f1f3a6afeb5b10f20e6330e7d54deae385c3b1c38e0b89fa113bc018b1289522c671b2d2be6233da7cc04bb3b30205c0ed64545131ee3a435593f98b14e048d263eaae7aa360ac0ebeaeb6c5418e4df46369e0127caae6f69e2cd9b0061b6331faa888fa1c30e20ed826fca8ef222af44e81987b4d479f4885928e957a3c5b3e3958a428a520a6d112746968e5957add6ce2a084ec386d0150a9af9c4b7cf9d2b5011f36c257a40dcc56535a5edb16319b70a8d9b3a7e43dc93c364b56b41b1c2ba276cd238d0aaad61d47b7c6ca83fc8ee929d013f9e6aded695e74517e6f4dcf51af8189f8113407749e477a480eb4554e63765d1eae015db2ab99979fa3ffa7289b8942fe79d1328443ffb9a62201730dd55982ca27d3d9f65d35fb8ec3a49faf201e9e42b49a1be5c672ff103183c342697cb5e2b11d0cbd56605363e5440964466136bfa3e8af4b1bde828b4e3ad489665a9c69c1855c4af8fe62e852534ff93860dbe4876b9940b8c9c38d84f70c394075a90a5cbff9d67247287c1864f5e2456932076dfd48ac8651c4af3e975d42deb4b5ad0e62d4295d110e5a0df6f487efc4594f71da9dc8d32212afdd99b58465bac2a6e0623d9edee4b67aef0b9a7f2d6659484c101871f7180355113e19f2fc30e7e44cee64b026e9bec8a0548836f30a90999b7c0653e9caa0aa53ae8c1cc3197385bc78141490fc1c89d1cda2a493da627cef766f47fceae20c2e5c5229c9a52e4aa8a403d12b7325e99c00709990852a4a933d4e454e6ee9b9ea868e20c3cb4f4b64743428ea2444dad2a32a70842e24618209f92bfa332d0a05e54d02935602f36e132202036d74ca50303e7b3951a5bd6bcff88d32de3c4dcbe82701ccc67c55ee9897595979f258a04b78cf0c0ee7dcd4d970358a35593e4cf5c94a4387576c4ac3913237ab421d71a281fbcd822561abbd018afe9e9f5c5e6160d4aa5c7f40360ff77788d94d75d1d8324e7e71ebb4a5b8892a47ce06e16f3d9d22ea7abdf880c1b2971771545e834a96ce2ba1618ec56f62de191c8f39960d4738bbef32fd9aec25a5e035479a8d08bf9b439561179378de776eefd732bb6d4e4b7e577246baa8b2958353aefd284a0ec8f6ed3238e58ec15b98e23db9e62ae99a69a47f10e880a69cb9afdd58f7a9d82db24dc3608f6bf8c260547c538f27e18de6d13cc262efcc94c591c2c109ad1dd88be6448cd4a09765f1bd0396cbc34e03d4bc2bdd9135b655b32e6cffe26b8747205b0920fffa8e5dcc588a23ec639de23bb1f198f9d186d6a5dda93e72465a0fbb8a46fb54fa0ead10ba268048563be84669fbe6e708080a896186f1bfae5a99457640f6abcf50b0314c56c27cee0182b4277c48649f8f0b71603a2f46652fe29c1c822bf5c97e38e8ec2092f5e09d7572d14e6ff19998bfef27668d209a9536fd152ea4c838b4f8d802d52524429d1556063c4179d117bd5c255ab64e45dc7bb1d79fbe66c32ab9bd86a8c03b324cfaf9ded18f275a3ef84557bade1e7e51dca4fa3e1a8cfed480676a5a30f773b6ab4725d37fe8a37089ce8cde4e2121dc8558595e1c48c227b98fb7ade12d5bd6d78dba20ed20e881c2d927a6afa73c5cee88fef6bf0adf8b04b49cd84776708c4b127fa66da53423688211236cd706e41200601383a2e30d8109c2e7bf23bf58fd55ce7da50e770c121a958476c086c16539fc1757a6ef853ad8b5af52d9e6a0175fb1cc2f9f3318ba3a23a5071de5ad0d91ac045111e4c51930ca9e77150fbf6369e388dcef870ba5e43af9a23577ec6b07a832fa2cf324874e8e83c4fe917f4f936c5554b4cfa04dd21578b6fd66810b9731bad1fa030b6f7b6306c9f3db78b6e67fd02080195a322b01e53c0ac1f0689c6b5da6ee8a385804c49150eb65b71dcbd7f7d906b2a4c4c5bef4e749b71f0a30a83d76f8b71aef69a5cf99bb6f7d5adcdaa008c4fb57bb0d3eb4c63b2c14cdc4bfe3167bf2341ab506e18cca80d6c683080fef3fa17c44a491f5dff15eae3e98d5d025c99e0683bcfe74b294c894daf34d3276636f87d312dc93e15c3187e4d3aa4b5245a5dd0dcf2e2c404a1e9a961155c9b89d6d35486935b591be513dd07258e2cc2a1d90e502c457917aabd7e8cf8344fe3298bb7af6ea0900cd7f1f6aeda0eeddc6fe342c600abb41f8615032c4f38f29f224fdade8c5704363492ab4b5e4630124158fe18a38d7f5a50ee94533138a2f61790e8ccfa5825d2f13be04d378abc5f9fe42431e5b94db1ca8377349eacb5f40873a177f7b51c1eb80aca422ea8cc001cbbb7cc7aad41929d6cc65006ce9b74800747714c46c0def87eb3f327c8906dbefa50541a5fee00888271e9c964f1a70aaea79727fe5000b2c6d7bb0c1229cde0431e3aefed2ed43a7f9e5647c55b8d94a17c5109b4c95e4714ded2451597b488b19178ab0cdc13d2319159f04977c4ad778a63565cdd37ca6fd23951804d4882871f9e5418d2c124aac504018c687f04d76230636212ea85ecde848e306e7e14029c54dde992cc0166c65396b55fd44d3004eb1e4817bd8c9d69a2eec5b3619875fd66398184a3a13608a668203f2344ccfba153f3fe5eb421c8ae137649c0436824f576e2ca75cde8717f85c458b4d0594267e180dda4309033f2b8ce4f842750e78a6033040536ac04010884f318fb2098d824061d17faae253a6cb3ee3ba537fefea2b9e5fd9684ed1bf029d19102621d01726e209a208f0c389fa3d3d94a8f2ef4e03c0c075b23a2943b8aa295f72d907a293d62c9452a22a8da1e34e7fe439817ad024f80551dc1ab7d45a1e76904626563f8708fd51e0ddc5821b037e7ee22dbd16f29854536d802ba4f73486046d169bdb117e8ab0794536204e9015f6d123117af8c5f73d29c1257b697a5779133c91e1bff56e57fb301f84cdab9311d0df6cee24cb9a636fe7eff894f8a34ecfb91c1939c6ab67e3d2e3594e2e1fd521d27e6cd0aa91b33adc1e2625a734cfb18929507eea166c4964b8d72a0c8971220431395a063ba33bd832aab6e71d4012f74e164693997e1f7624838963471fa393b66a3606d068e0256720eaa06cc8d03f957c78bd73877d124a5ac897076279303bac01fea1e98c4abe55497db4dec0725a9fc5bc5112666d13d078bbe50aff90a276bd000360c7b28777693b5ef46b4e55d6de048fe0910e9740eb360e1ba587dc452f8f8114708d74149c9bae58f8649d92270db4c1cf2e515e31d548c7f81c156ec8552674a5edb18be917ea3dcbe1be6e10dd47352a5e127ef6a2057090aded6fa2fd18776d9a44a6f108731bc806353ab1a1628787050ff4dd2d111538b4fa99454c1157c833fba9d14e6b85f83808af9b79096f9b73bf251e16e75e84ff57b28d3f0072d172d00ec46eef89cfe3e8b2b522c0123d5ab49de0c07e7d16868f666080a70a0886082885cda016c9b8c2e6501158ece3d15cdbccfee6ef0827adc213c59bfa9f657e8fd6159cab98f75839764507b528a7c7934d8be2b35766759ae01917115286e55302086dee8881e114582caaf5b46e8dc65b138ecee44569e436a77ccec24001245a0b6c2ffbbcaebe5ed18881d7fa77b1423144a14846786e529ccffefb12d6fb77815821cf9aaffb8b6cb7d89f238fa4ca690c13cf36c1e356263f135fd2327c9d6d424b96e6ed6468904f0955331c5faa7f876549752d37c315f4c007b2aa073953318e2bd4800c07c00c9a9cfe643a49f7353cb919a3017582a62320c0ff189f659ffa9b26d117a7eb870bb2f0f79dba6780b0592cc20d7e4f963ffa2d535986841f4e3632afb93d1bf89aecebdf46569592d1e948159390f1d483c8f207827301be986831882cac2a31d55f583c989c6f86ceafda20a23fb1f8338d5b528e8a6adc05ba958ccc2e89972697871a6eb6591801824eec171b28959b1b2b9f7876b7ec5f0e79c1e291aa3ee9e10c9ad9bb616dd5602768a0bd00bb04cc864fb13ef7100513a2e63e1aa3af265c0ee34fa76c8c47b272116cfb9ecd7f8b26aae8d0295d7762151d9ff7e616c011e7231da29b88f7c3633c0cf3f9b4cf96c947b8fb48d87fc9e707544587d7c3fe5900ebdc822b299a413348613e34a2785e7a4d522435012e7777c3debcc7c081434fe5ba94d620a878868404c062e495f797f8a62d7348c69e5783bee0c71d350a122076a79957f330cbe5054e63b0ded895b084b22b9b09f1e7a11c9a517007781715995a504397cfe4055bbcad13583f6701077be956a2888", 0x1000}, {&(0x7f0000005340)="fc4259e41e9334ef8e7f29c39fbfd5cd583afbc9faf902c667fe1ca45254c386415796c67f75c675da6f476ac3bd5b83d94de4f5f8c70ef580d06ccb860eef5b4b5d3bc94299d4d21ba677b993dc65721c501bd3161f558155b14c9a3866999a68481ee18ecf60c5ee831ccdbf9444764a4d900231b8d190ecb74a472118ade7c7b8cca0e8a89fa3fea10b6324569e10576d2e6c544886d34a3eab44a3ccea52906d3d5f80344a66e37f447e05addfa0dc017a2b3bfdd530cd07d0785ecef2453d9b817b25732682d82bccdf3d62a99a3b0dedfdf014b83a825c52c734cd3b511d900bf3312dfa7c9fed4a4bcf35d33c2cd8230f690a68b3954eb6a68d79bf0e2b860044fdd1168983c427374960b256b4fce42e19b7fb75e371e253bf329094560608b8be5f625cf385929dfe8c63e1cf7de7b63ea6a5e25f61f260ed8e80fea02f4368be828d07616acd0fd6a9090ede8433f0f12a3f81927650cf8656c4ffe585c98c7a0185e7026dd53b5caa1922f319fad7d8153bde8bde6bb8ebe9aee764e85d5aafcd6ee86e42bac1d22ce9b8981381319bfb9810a350dc39c70501d903bc2d1da14fbe4f97c666cd5a7385d72a891fa70e0d36df3d59a09558052c400dde20ed5d15278dbf4a3d04382e0e8d91360816f7b428b30ad4c93d5481ba27177af37d0baedddaed1c0165fcbbe01f9b4415f298074f3dccb1d4c2c80d219fb835ec570ec5b4b6159fd63f0fcdc6c53d75e67ac48734eebe0f2d0db0844ebfc8874c86512121750a62a5ee12108777cf64f49113c6ab8416a425b0180f5509f8d0f56928983c4be698ddf1a0424a4bd0d5b23ad9cff94c9ccc9356dc33fa0651376852832519ce00bc3af08f5b231d39e04fc3867dc8ece51e12288f1f315a7a8adf794c9f8b154ed3492f75466c873549de5621bec2efb8da0d40cfc2883516d3ddc3ca6fcdb3b61a0970162d781e2a1f90433df7d01401e4609f33f30ded1f486aae776fd325635927602a9c7b768e9de151a62ea613450a925f1de616d5465ed6f0be965827551638ee93f292129545e648c6928127199860a432e70d718cd9a49527e26d32bc1130da70a7adfda133602855e3f582bde618350d01130361bff10bedeb0a0c198aee6b6f5a1f89ccb5853d9b095ba4d25c2ad9bc4d61a407da5826bf0e434a7261eff5e6eb4b82e21bdf5e74904bd2999d3f6b6a5f5176d1d11aaf0bef587c10c403dcb68e8aae02366681097f9bfd3fe30ffa2320ef7a2442990d23e9d9bc7277bc10b5f30100d861401e20f281ea76403336bef6450db778b29dd8afc5245986e9e4ead9d166b94ac4ceed0af1602a07ea8fc6db2c5da3f3ab0a45c2ded381d61fbc494d160006191f43c3b24e5e04ce11124ad884e05fecf8901f14161456ffcb648c069faee9c78f72eb57a6fc77315000f395cb95cfc363889784e218ed7f65bb17b7be37faac81b3a35dc2ab8c5da788df9ce7badcc121be3e7c71b4ccfbea51631c8fa43a69d5e8c0169f14e1e3c742ec2a813604bf3391eb78e349c0d3c227949d6cfde81dcdf02d0e4b10fc4723b702078b04b7cf72ee232846617270dc4f3830258fecb74bdf3740dd178c3dd1b361044d53364fbcb7a22bf4a660e4d1210aaa97bfde18d4add4cf6eef609828d4200afcf9d0cdca69b75d3d67835432d379084115e369b28b6cd2fb62f0eddcc618e4df229e219c47f407d8f21c65f77aeaeaf9663ef2b731880eafc3f193ffafdc7ba8908ae6936ce24fc89c314efd8758b760ba2a3d421a4fe203cfef64409a26143bb682c68a5ae2b90a5d1681a2662d6cb3f5b45f07a07f6899e3fd4d1f99af25a5dd7f655af82276832504b3bed1255b137558bd13c7ece490b6be5e5b07377af1d502115a947e183ef52a19d15b357f3c177ceb07fa236c9ffde22178f40ea10ef9d68cb3f4d39a4108212c22945fb0052c3818f7c8d219aa04478aa6ea504b34864eddc339cb414cf7a17aa9410d808e1a09bb8c58c0fed0c6038bbe8e6f6c4e9f99363f93bb171ae051508e0fc06c8dde15b02c3ef9032fd704c3bf9c33cb7e13e6fe19be7bc54e0b26f4ca9255a293e25904a3a6c68f0518d52c5f1a9ada641da8d700da1b61ca223aada4b4a2dd7b80f2cb04b35e633a3edd6485bb7e9dc51732dcda1bc915065d55d9482a73f77bf50047957864d31935640111684259c957e9606a4d635d3b1aba1c0a69e6c525abf79fa81eab70d587f416c609aaaa3c7361504e32dc02f8e7b68ac3e80ed629da9d928346eb2514b6267b19c5f000150cdadf739f7e06858d7d47dc90accdd909a5573f56529646ce66c8b50231466ae2ed5ad6354443ce770032db0a06e0f06ef024c5844f6a562dda69bd956b8fb792da3c4c77d0a447909ac5b2532bdb6936bc67276ca107281628750890174b7144ad889f50674064a147e012f8cd585a6aca31dd51fe58dcdb49f8a82b97a3f6c93116bb46d2bb3565fda74e6ac4329943367c85d4aa89512bbb6382c5bf9ed64c4ebc34bc119bcbbd930f38fec3bfb3a881e56847edc477f74b319f600f531a19179c8e89803dc6369b5b7cfb523bbf0849a092f25b12202daebacbe6e48e56b0928a0c8a6e730c06d5f19e1ec10a6467b0a46a9d15e0d2e84bf624e9fc4a8caca128ec65fff6822b37c8e14731a3f6c96ce17ff30818ea82db9ce18bb551aa596e005062cdd5bc4b81259ada5d45bcc39f0121c31ad5fa8cd55b3a1284f7b6f37f4bb8449d8d5f8147acc4f215a8a6337de23770e470a129b112bac0f43ad43fc3705cfa07e20f4a434a0967401c78a95da55072691145ac2beecb634402dea8488e6ba8934b507e55f239b5ac26bdff84be6e7c45a28238f0bb35bdd613d13379750f40ecc469a27b497242314c7decfd7afb2ebf4073cf8052ac6bca42d39aa69ba3e30d3cd09eeb15c4f503a335731dba43551986559b62fb7616d5b75968df4a7bad5425355bdb536f6ec3600f487b963e6beea8b2e96b5aad1cbc862297438bbf34b21b3d27c1942198e178408cf336a05139192ce62d483124a0fbfadf2ad62ee2504332a59ee505c851dafda2f9171efbd2a25db061482e44849ab81e14ed34fdc612bcb017994348e9add253860291a369b24486fc6171b5f32ab48113f181b57769fe1e718ea2678ece437a9703c7af904db17612f84c3baeabfe660434db94b6378945de07f8d1433470cfcd0d12ba13e9782be71288a44564b9983cd1363f218c321cfea30e619940498e9aa0e1b14714db911784d9898ea3912b749dca636f24e2bab0924b79c324ad9ecc84eca23ce9171a9693415b98f96c6789d90116104eeb5343b0653a496ffa8fe4641e2537a651f0bf5d182e0c9aac5d05b0524f07b96645f9a35971dc702ea47cacd3acf92a13895961f431ebfc91572981bf218e004fc742a944845f60034480bdc989f57171e15f6acbe956a78b8f2d05a18d6fdfa932a20066ef4ee90169c27e8b49dd98fa08a73c8240937659747b7268909aebde7b4fdea8a259eaf4206292e489cd61ebd9bb531a57b83927b17ae76594854c98339badd3aa806b6f9e6725f7f56ab0c45fa0191a7dc8f928f01a8552af48726ac0d7ac059e4dec97367a73421ab0c3b6e42eb964039b6d7f0dfff79be893575e1fdc21eb1aa6064cefd0c40f3475b4d00ea97bf58f7ba2cdf52c372c7fed8c686d0dcf49ee8256f8f126f85646b89287dc0e796e3c4d1ba27305f3a50aeaa049cc3a31a04a7b8a86aad800de12907a89e698a50f23b3b00ad53f1f8fe0e9c9c8fe9e1d56f8d4216faf126dd6d8a7e49bb1c2b8c01a3bb9c9e28efb6b1eaf2bcd4c34ced98f4966ef06266b708f3521c9148d4f1d2375718045f482b33d651bbfbc9cc8d464b8d3034caa1cc3a2a9c6d20bc48002499d54ca1ef7917535cd00078abec01b6b32b6db21c999e33837e2fdfa7bc9f158443889838b9b23bd082761dbf8e461032a9b4daef3ab29bb618cf966214861f8c8026dab28fb29b983aee22872d120f2d2dfe81711c6a62b3645cd7118285c9ac6a56c5023ab59c56bfef9af15a14d39c4ec719a033ff13f6865b7508096748f523bef34c35c3777d3ebaf867f1dda077c97a93d14e7a4ed1ec532615014ba2eed0db5fa47a2634d3672353979e31fd348614ed6592cc1ccca00e45e92e079eff24088f9881383e7279337df6228f209893548ad47ad20a2c4001b8ca94cfde529f6c3f4b779bb78499b66c2b8e7c0fa93a415fcf1c82fac6878a4181d91f9c060492a2cce6a3d51a2f98d14b5090462aaa0994adcc11b6aa73ee167f489aa7213bd4608266a265218e0159ffae1b9404146a14c0217e50155032f08177b24f479840fc8205b3cc3a73f084844caebbbe2bc23d03d0c85f71a6e879d7c8c28d82bfa9063ff775a2b78ce78691eebd3dd053543d3761ba0c375305d99c82f1e1633ecdc8c2055ee3ab2d3268efc8534c17cea7269a80577e73a89d0dd3fe3217675fca5a834afd038457de50f823c2e61343afc85be56d6d8010781deb56f739a7516567a0d85abae58336df12ddbcefb46eb32aff2df593c7aaf6f4658d06440242ca94da1edb70a420dcd56be4997796fb0a19bf5958c4cf2d282ff67c70a74aa9e79551b81a91875307743357389cde2d0f93c9aa543e5363eaa50cc7624401b7383274547dfca95748b1636cd18f4d802c46c7497fa5c73e1f4a2ee18052eb1c18c96cfb5554a1c15bfccd0ecdfddc87cd1bae4e4fee689794bcdc94613880facdb1b61b34252f722f9455d97f85b8f061b59fde3af910d75c8f3320f4318158fb8cf6f99d8a8ebd60b557d2a3349b1809bcf37491150c7ab3a48d75e7dfcc4e9959a64d1befb951cb243af5fd6f3555d14b66fdadb067d13eefd37dfd7f506a18b2031181b42d7fab7967f02a93787e7b05ab9ffc514141083ed88a26b21363520f50b29562c272936ef6333d7e1d8ce69bd3546dad6d0756e1e6a7978848625abe83e59e3c2bf270bfce3869a53d5f874d281dd4d73195a8e0924af637c0e6bca82501e8c6b786d1d0c78bf80ae756321cbd566e9a4cbb101356333b342ea9cf77be62f0f43f7daa3ac8aae1299d2c7be7d5efe61bdc3c12ddc5dcaad4027f2a9a0252236a0df178e01d7cea9be29a8f6b7dc7c1de991d989eecf08090b19d50f42c42435c7712314c3d5eb4f855eb7dbbae6549a2eab1bde23d1a21faccec148bc2b740cb890d0f47e12c21960b9e7bad3a7eac8d96a1dba2ecfd83801ede6c1132ec90553d4f57c0ce90abcdc2a4975013bd2db917234947fc69e18dc246373d0afa83a78e83a9f8073bc64e70ca89ec8fb831191b6f70775d384a94cde7d779d6d057f1f53b684f15b99daadbe83aa6a3f1532e5ca4005c16b562110692c7382582d8ca6a7874cbce5f2aaaa67411abc61a95a061be8b376dddd00a4da1eb1c2b88c702cb19908bf7ffaac23f38b38ed11727c02157ada0b749422bd374b1cdaff0717d7ed3b056350e7989010202f444568076cdf8a675a77fb0ec431add03cfebb47b35e1cf0bfbce7fce20edd4283d9bf80a9a966c602dbda1a0147ea4dab6f3b9fb7436be7af10b07fe7299f7b2da0295b9632dc9dbdc5264b280c94021c86e93cc0fb636217c6c833c8080840bef6dd5b13c99de8af1c2c23d26be218e59b147bc6687e8a12aaf9e8ca5cc033ab1ac0f90123228276def0b928d1a1430399b97ea7ae693bb6d5a978ffaaf234e677490bfc3341b3bcd1328ee04e49a4dc", 0x1000}], 0x6, &(0x7f00000063c0)=[@cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee01}}}], 0x20, 0x80}}, {{&(0x7f0000006400)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000007900)=[{&(0x7f0000006480)="990a4c01c2da74e9a6c1a6cd604f13f159d65d5056419a51d49b063f9e85dfde6c184ef4a55e4cc6731a8f3c105dbd12d5c1eb2407480d2736950f63f37a20f6b753eba388415153065d35cadfe403074895e132eee4bdda08f76a9c3d27b6931cbd55cc7da71f952854fe078517b97f7d9b9109ce0ffe09", 0x78}, {&(0x7f0000006500)="a06b340d1b1bc604b7925dea48798ac49a8fd2fc54d999c381950cb07f7d6e81d7d0eeefcbee16a065f111e03687c5520f4e74e069ed78ebc010fb31c67d56e67f9505466cfc03e0d0c5e1052874ec81c0f9b7c05de32d30cfa8715bd987b256bec59e638e2692b1f71af6a01a4da01093ae0cb1f487cacc4f0dfcae65ec372114e5eec30e5225cd044550544cc1fe367c6d4ac1fc9c6d331d4deb04d73a863fb672ece89446f1c5e7e1948505f623ad88a5cbc20e58838adc9acca9d7bb8122337f15b452ea5941e1964c4c3fe4198ee22f78c2f14bf3d4f11ea45d76169c5c261d7b9e00bc63623089570b9844576e3d05e2854b93da6883d84ab4fa2ec8d54f7c9cad73609acce06e082834027fe7add652ce2931e0e846d5ad5416ddcc76424b389dd2aa59c8dc0565399dc9e638183a0d67d5a7ffafd96922197f5c0e7ae1b119865ad7671e43fe09beb51067d5613f5550b85ffc50ff6b93e6e1a7bb59a2c1dcdd9c4262b20e34743f1a75757dc8443904784b82597254bd964e6cde1c4ddeefd72efac6c28e35def864c01d0d12f3230f940236612006b0321a293c208643b5410a7e7771e1d2c54afae526c294c4445c06326a09a39fb88fdb66d9fcfd1a20f57fabff4041be04ca9e367e91857bec7750df82057e5e66f3e456424c335432f15bfc97f597cc938e0337bc69bc75316df8d619b4b97124a8df344a960cc3e4c0f164e1cf0de224102715be05e5ac3aae82354bbed3297b2f5ec912e3db28f649867c1c83718826005b73139756264f3fe6678c8cb6c0c24041953382763ff1a08392fccc2c8089b742d35a1600f86a9904f66870ed52fc12f1f5b90741954d716b7f63b3e35d55db5f6c6b55e874be31b73505e3471be5f3bd88f74c4eb63404f28eece84b8435f31e13beed6bc2bf58c4a7e586d38a67a576baa69ca318398f0c97e30727020bd8a607aafc1a3e03122538af4719532d9ed1ac3be4873328d38299e361249a3fd3dd7b776daa09cb157d49774f4d3d6d6ab6d6589376733ffac975ad864bcfb2b3b4e38cbfb97e162e8dc92fb44283d06bc98a7673dca95ec5b3e99c01954df7299f02e1d33cd885f21975efbe69d98cfccba937f184fb618c98dc762d552075bb147758f41254eae89479510ea116e59c17f7cf39b5d1c54ffc49cabc150dcc5b849803718858c43c335efa7fa4763483e60ba8a9939da6def99ec249def67a311470dd66dd946376a34c68aa8890b26cb92ce3ada1211290db10fee2e8ac5482bda4c10f865b1e09acf13ec5516d9eb44f2d8ea09e84491a2f2cc62884dc481cb5dd9007dcf968f2ee08e7e272bbe8ec7f62604dbef61ba05487eabd7b8d934814462a0429c037a11f05f356b6e1318e3fbb98c54b99e1fa0f440b06852c36d12ad3e36b869a83d3fc00d7ce8dfb02ccdcf10413036fc020ad053a118707cdcb50f04c0b38200665f81ef33fa04eddd39078ffdfee5cce049e18b5307c6ea32b33493abd6d7072b16c1be1f58bbc145a1157c48c48454645e52c6ee8db34a5e9d3d3618cafe63fbd2615fd832d3ab6ab90cf2f0f346d19b6a1152d0862c9865b99a4108e56cd4b2813455c96725a26d9fb2753b98ed4dee9f2c0d0d932af910edfcdc439d4f061f8f8679cb4a3abed9517f026384f77fba19ca8be6b46db25e087aa438f99091e4d10979db47b80ef39116fe5d19d2a6bd47151bccb40c8964858ebb849c27c28d2d55a6a31d8608150d10b9278bb5dc52a300ced3edbebd54fa2fea803870bf2a6a1f4cadf90ecd1ac8bd1708784c0298f6adf07ffd098d4a9d387ed1e510781e5585c4c5464b4a69e778cd4d43e60f489fedc7f17d9e021ba4f07e5a9e9b9714b4d1668036ee9d85a018cdcb130ac6a3c2619a22cc43d6fc93e9e167e0a5f74d40f15e1b06bab9f6e734a473d350dacf3a7c21793d0256355ad71a36be8ba60322b90d397e2eec55205d195958ab75ae4048a73b2b7e8153a51541b660d79bd736243ce72ef184ba806ae06790b9a405d42c223b2c26f07722e1b62af9014b42f00962862e27a5f0ceb3054d62c3fe818ebd3e17caaf6b204a9efd5c26501a00e21d9fb455cf33450743b995a6ad40d03e3d43c52de2de33ef4921a19d4d966711c6ed23fac799f247c319e4788d02521d3671d4fbbd809878a1cd94d81edb3189852c279bab984540b33d5ac156da7d66c431ac8e2f7fdbae4a291fe957602138ff13235f80a6e6175e7cc3143cab4818f591ced30a4ba032ee5f8634bb480673cfee72f28c60a4cba42b119edcb7f21d9492668b83876f05ec055805e3c72a4119a02b7e1aef8f3ceac5191c7b6befe4be1c14f06851af89b408e0306fc69c6ffe67fff6ecb43279216e064ac54bea29fd12ca9a794011f7c673ab6604a576773404551ec791a97425d2a971caaca56452e9f8a2b68d6f8b0fcbeaaa47867fc5e486a53d6d0cfad2bfda8649743ea93cef1b32df4ed01131d38746d4bcbcc056b7f88e0fecbc2fb2f1bcb6a91d07b021adad9181ad15927cfc0c03850a71d35b9701b98d0aa111bf66fed7677edc22569551e03415101dd40c47d97293d407d30d674f3296348cec5771a4c5b5ba0321f7d2f0d21cd5a115f1eeb533c357155755a38acda015146d49f6d7c237544eb4643fdb8c7fd70ddd4a86002216fb0d87d269e8b1d5ed7ee7c339494e152c5708091b2db98026847fab4563a4b39010c15b3e52f39dee01b8578f1bfaf71398fe0a113443ddd9581c5a994ebc32028cdb78182975547037569c4a570c30c3dff2a8ad316764fcfc809e5c812517cdcf6e718c68179f1badbc946914862fe5c6c86d1fa213cec0b21014764b6dd55b64dd69d64a245409607c124b27f078c290fc78f324b4f053df11692ca5b7e46ff7c99423cbb9279de45572397ed7738bf09099aecc68f82afbb0acc32537718fbe3ff659a7af675a17618d91fe86c364123516d2952200da23d61e091ef26c2ccbca0f1bcdb87c1f2d642fe657424615744bd92aaf485e389db62aeecee6ad78ddd3e09bdb7714c517c11a9557570c8451c1b98ef80831a19bb900d32290ae3665461b3c7abe60763de2fc7e69a05ffba916631a115d3f3e10e007ddc3d7330d8758ec7d13b0a472a4fc376c382b643410217deea3ad0cc376135bf4404ac36b04e992c42b4c46f439e755a7790c90f7b28d8a868cc0793ea561c58499845cbc55b5205d168ed7f139d37e1ab100a9cf8561f02470b37475f136ffcda12c95ff39c4f0de11ec56a9c7f4e4129e706489c44ab39cd682469728e4116cd3b65df341e5173eec3965e2fdab8b230ca9762abcad5b2928f68f789bfae8d476ab3393caca3da4e281ee308d2592c4a23bfa2c5e7977a09d7ade1e8c1960336b61b429710b4d92b7a9f96fe7f4a502464db2cfcf7325c0cfe08732991408f417eb2b04576f3908b18e4cbe8df33ab5a8e84ab1585d191c1e8794018c3fc1a1b22d0765530ec9ede7b17243ab6ad6df900b2b81b1fb98b8098387b25fef82379e4f1661cbf29b5690799a8bbb395ca54255771a631df4c08dbb62cb63e52cc95c3caca1c90dbbdd8a1aae7ae6fcf0d18f90a7e5ba5fd65490e468200d947cbcd6f31326ab0d8a19a9d523abdc9d7146c9d1f065e1d272223586e9432a66517547a63e23f835d71ba465adf453cd4d35f2f3ac14bbbf04c952ab0a27f31211d690e1f4eb082683955d8cffeff4ae3600d4ff6dd48033cd5259e2d5cd41f642388c969bf1f5f29d4369c518b2640e52ef0adefdb4ba1aaa19fbafd4e51084c7da39c7dd710fb5ccdd60274aae9426b5f77adb5f4d397e41de3ab2586094ac0c59a81505168b6778e6c48a96b175bc2628e3f8427bb13c39693635fa0009c649a5c63d7409559287642f2e48af737a74d99e9ac378338581835ed22b3644bc0c7f45f08bdc608e4a6f7efe21a0be50539669ac3855c82c27dd2ac958cf5afecab4b8c29e3f4e96dcea1545670d2c3500a4e84b37335818ec2db6b69899cf3891c1918d3cf55f84a6ad6e7c7b232a700dda81d8adf0a172cc63dc1da443661941d6ca5fbff1d9ec1c1b582a9c440b613cd2e73aa17a4cf85af448c202f126e47be73d7afe97997101ef4eb08cadc22618e0a8e120604ca98f63c5b06c41a6857b269568d24f99d14986436192d280abb7c473afe28a2e5adef5d3efafd6998d0b3ca2be7f1a176713246018e2e5962760be859231639ff3080661a29922dae5ef818aa821c376fedee1032f95d1f4bec89ff3890d337c7b534754b650e1c37c7e9cc4c99f2ae77c9e3311ab80b9297f71a55790a1e429799a58e702ffbde08ff205b371118b54821a0f9ead0cf43a31de6acfe444502ad74ed66dacf7c6fcbe7c9ee40ebd0335e37f8c97513253d32bddd4270897a01344452458d147df0857496a072e8a76fe6d40468e8a989b3066fad17762356a41a030ad0fc05a99d8782e03d5fa614c87d701f4d97c7ddf8ea83b35c92de0d07a1b93476627f3a70413ab5096e715189613b523d963a4b00ee3cb7e2a318d8dfe9eaaac214d9995e4a27248b6689afab0a1e02c82c3df9484df8dd02c0082ff4dd4c97d0735e634ebe86ffd45118437562d4661db8a650d212dc4fa60785cdbde2118ca90d814b666bc890257c0fa0ce94958c37e9f54f5088973e9781406d899295ac8f57a1c155525c929ba6cb73d4b92606ed38dc4756a06a8e5fe05ecc2c78deb504588ad16ca73d794a833ed99f2b23b7df4b64c7b4171881b29b776e088b766de5ec287770fef9b23c4a2d4f83e5db3c4127a82063f5bedb3063f8442a3164a7b0fd759f38a0b2ba394059183f11e8a55f266dd86ded91d164f1d2db96563f4c8bb62163d081069ccb30623306510b901b3d15b2e2711fe2d3944d7c68b528791418db71ac950ca85cdda23576a0aa9f2d6cd8195071185598357e2e7e30ba638fe85b4f678f0b6cb5b1aa82c605fb4082d93e3a86af4c918de6571c0d0d5911c2237cc0b82ac546a5332b7f6bd27b76c3c23aa95621e80fcb24a5b1857c83f2cb31e75fe7203b61697e72e1e4fac5dbae1e85fa24cc5e45ecbfbc0473043a8c2117ca09f5b4998339e2fc963c1967efe7feb5aa1c59313db1da5676a155609803445981fe7b00e4cd4c8d25e6d26ea7879ef7bb0f68184f2c476870e71d8a6aac7122862863151a12427cad8c4bd25d9181cdb84255875eb87d855cc5b4fe4f7f2fa3dca355b780114fc1053acd4f09c29fdc5e968e5e480c55aa63684ba4501e30bb05385707bf55bb3bbd633dba9061bea18f771c42231f803447507aab7b9b321676a1d25256b53e4a46abef5007fb56c6543aea4268fe93bfb223803972a30214af7cc06556ffff38e17e496097a8fc6f7daef5e1a7e9be0c0b68557c3e8b9d5b99f47d46737654d350bca47e4b1f082b2fade0fd3d2271ad66be3091964f0c81d4b8aacb329604033167476b5845c7679d1cd204408bbfafba794da610fc68491b16e4e956bac17a250a36e19782f894a90996b997660db916b87b7efeff28604c0e6cfab4f5f7ba3c2a0d2477a289b9792793f175fb710fa8266b00d7b61b523a6080cfd1e1f6572487e86cdd9be3f657f3bcd6691b88beca90b13a35f5ea174c5c8f419b933aa42e484487e678c6501efaf73238234d0065efc7e56a8ed65dd92f29d67b6479987459391f0ba1080b836433a4b747871ae09dea3cdbbe1cba96d541870e9861c7c6042f1155afb8b5f7219ba338ea4d", 0x1000}, {&(0x7f0000007500)="a6c275af836e1c12bb9d3d858cf382ab13f5a9aa64b7ab0a0dd3dea9f483528176080e347a5892bfb94cb3efccf4746cba71146f8e460784053c5919ee21113fc0582d7bc1871fc7e118a973b73e1b8cf4c36ad337129f3dfef4b57b5f7300a18d29acc969587bca44f72215d8ef9e117e562c1c852284196cb1e674b2fc2e41649a458e6d92af2dc0162b57f8084bc28e5ae8d56bc66e", 0x97}, {&(0x7f00000075c0)="81ac386df879e8816f22f465f5d2f4a1a4b6a8aea49cff58c1d3eca626b575df6de2fcaa8c4bd3ae34589b4d79811012447ba0fb8cb1c12577ec5c8e9f2fb7888ef9e9d55d2858899e3b44e5aed91bb3b506098d55aeedf4fe7b9a7b50253489936343c53011a8802f856abeed4e285af9f4260adedf513723335012e53855debdbebad444fd578f90dcc0da587e5d63881df33b2e42d17d0364face927d02214b46af7f18f5a52c1416843dc2c8e95c3c5851d0a05aa058002aeb3f36", 0xbd}, {&(0x7f0000007680)="026780c2a77108324e072ed0071d83726a0a4fb02fa520d948b5f499055cc8821a94c23fadcea315a6399d59db2fe49e48eb77e46375acbde00df60baf813200a2ea07b28199689e57511f87dc537de87742265e772f5b7f0368b69e6e8b8b8b07b170bf67201bfd58ea3439b493b02ad5649d39d49a6f4b51bd1114740c73860db68a4922f8b299db", 0x89}, {&(0x7f0000007740)="ef1fff886985500e5d4695352482319cd86ab085cf3ec5d31d06e822e9e992e7d8e066d01dcd08710e74998ff8f6d689b9e480617e46f48cfaa6c4bcc5fa902a420f0ed04236562e2d7d1c623099e7443e8b2214aa1038d4d8197ead3b201f91a4d30411b9ad0e75834706fbdccdcb7b5e5d6b224789d3f267471d150783980751f2b2e60f8e5b026ebb66b2ddb2bedadb14f1611c5f", 0x96}, {&(0x7f0000007800)="a6ae2e0dc20d8759bfa97b72e78937ab683fb7853480a5dde16b3db7370d882324ceda38a19516d2909adb745928233e76", 0x31}, {&(0x7f0000007840)="8e64bec5da473363d7b627ff055808496cc03268b383aea1dfa00ed1e6774819dd0cb9522044969f8760cb7a3935b4dabf50c04235229ede05bed60bb0529a6e4759ad2a80dd8034a2f2498845c7045819e7276ecc0c543b78ed6a701d5faaf6d68fb99c4dc0d3b21a7895bce123cde793", 0x71}, {&(0x7f00000078c0)="c9f6c5dc1e42813ebccb3c", 0xb}], 0x9, &(0x7f0000007a00)=[@cred={{0x1c, 0x1, 0x2, {r3}}}, @cred={{0x1c, 0x1, 0x2, {r3, r6, 0xffffffffffffffff}}}, @rights={{0x20, 0x1, 0x1, [r2, r0, r2, r1]}}, @rights={{0x2c, 0x1, 0x1, [0xffffffffffffffff, r2, r0, 0xffffffffffffffff, 0xffffffffffffffff, r2, r2]}}, @cred={{0x1c, 0x1, 0x2, {r3, r5}}}, @rights={{0x14, 0x1, 0x1, [r0]}}, @cred={{0x1c, 0x1, 0x2, {r4, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xffffffffffffffff}}}], 0x108}}, {{&(0x7f0000007b40)=@abs={0x1, 0x0, 0x4e23}, 0x6e, &(0x7f0000007cc0)=[{&(0x7f0000007bc0)="a768cbbfbfb9799d2798e382b370d7a4f97aac04f4e2cb6cbe709ab9b39acd04e90893b203c60b3c5cf2f230e98e9ab5dd325d07c94cc27b98340a5747ec0cd00d0410b513cf1355fb42906f4d5bf3588cc4e882c988", 0x56}, {&(0x7f0000007c40)="541bc02f07b0515b4e82633cc9001a6993f8ef2e336a1268b477b9725dc2e8b25cfac37c52dcfd6b73b31903b6634b06bb71f0c89c287b876147c4d7d974dca4574de11d8e8126a744ed0870fb584b70af7b1d2cc0512e47b67a54bd49d062", 0x5f}], 0x2, &(0x7f0000007d00)=[@cred={{0x1c, 0x1, 0x2, {0x0, r6, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r4, r6, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5}}}], 0x80}}, {{0x0, 0x0, &(0x7f0000008100)=[{&(0x7f0000007d80)="fb76ea769ba395d6b15f2c8ea822b83a9e2029a24476a5e7ab2f2eac6a6be60d53828a71a8559b2b37ab983c666f868481123427d322f659e1cb78ac9901b76b8dd75f0d594742d206f1406e66b2231cd83a667a968dcb15147b25cf1118ac465916f417022d2b3b0ccff357c780994b497968ac6f427f404d3288a6ce5592e104d4ca24273a216c432ebf4fb69d18de5aed85167d3348cd1b2e43edfee4eeb252f0a749ba32b8ba42455af1734a2c97692a2eb5b5bc01c96378b10c103f7ddfeba03e61", 0xc4}, {&(0x7f0000007e80)="45ad8de36729a10cd6b48e0b0f0fb205a41e3a569cbdd1c571a333fd268886345dcd3354e4ceea3ebc1e5135dfb20f2648100d2d49227486680a2da2d9c3ba339f27764a010f9a21b225e8cae79a408379c446a07c0044e21d62fadc37b2b8dfdc8fcdab2c27c8cd7108d3951ded4e0512a53d9806856e6ba9cdf5b86b1f37fda6636c755b1d3589e1a6e022edaaefd532027f833f", 0x95}, {&(0x7f0000007f40)="81b19a1001504839980eee69451e2846e151ffeeffa50f5fa3feca563d6ec186489394c42e6a5a3a3fff6d7d2841f67d3d0b02ee3124f4055abf6a61881edb47c71e4fd98aa45fa51ddc306f396ca61ddc2cecdbf03aa6afd0614d76d35c24f28f77a233e35cd8a1efee2b7396951bb43ce871d4c51809b352ae2a669944932843fd858863d5a380a11c9ab566f5f38945fc628206162042", 0x98}, {&(0x7f0000008000)="29cb7f772025312ae8a20c0e3cc46d7e78e8368edc68d7c32c9136c3763d2430ee52bc53648281611ad141fad41d048b5adbcde1f95d5f3f65b06a64c24dc1f6db3bd328ed1cf5fd4747fdaa76d6d8b372ae47f2425fe9bfe60c034013bc415a819c4e87f984579156be903fc686912e0d2c6152c237a2da3d5477f62fe5c73c9ad3b004369a64dc719654413a974f9707bcacf545085ac8bb985061b31c0b781a1629f8620b156b0f487866c60b4cffd9c83b42f71f0184c970186f75ba990200805119ec77f22004dd1e65e2bb9f54a53e54f66f3c09e04fa326db", 0xdc}], 0x4, &(0x7f0000008140)=[@cred={{0x1c, 0x1, 0x2, {r4, r6, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r9, r5, 0xee01}}}, @rights={{0x10}}, @cred={{0x1c, 0x1, 0x2, {r3, r6, 0xffffffffffffffff}}}], 0x70, 0x80}}, {{&(0x7f00000081c0)=@abs={0x0, 0x0, 0x4e23}, 0x6e, &(0x7f0000008340)=[{&(0x7f0000008240)="470737cc8d755bbe577ef5e4ed9fa6c8c1d0cf5c4ea5bb62a38a50ed79b6b5af2a22626570f19df5972f85f5cb5f43d43ee1447a7df4bef6054708c084659ca72f4235b591e222b470a4d17553deda1857b4e77428d80df92c21e106de3f4e663523f85f224e514f9787989768c7a276b3cba1cc6be337d8963aac261e06d7cc2c4d317a786c23f5e5cffad913f0b06c40876c07b9ec77c982f97ba5c191ae1f4aa4da95ed1f6a4b5b804a1d2e0bf10c19b7d4cf33c0216c990394e571cdaab66b9a85a204b2586f2c0e33", 0xcb}], 0x1, &(0x7f0000008380)=[@rights={{0x1c, 0x1, 0x1, [r2, r10, r2]}}], 0x20, 0x4000}}, {{&(0x7f00000083c0)=@abs={0x1, 0x0, 0x4e21}, 0x6e, &(0x7f0000009700)=[{&(0x7f0000008440)="f008f5dd6a9fbf0deb81d4da26bdf1d9b25e1f6b8b4c689d1d5a8d1cc81426234a0b48429a1836569b3241be6dec6d75ab525e6fc3acac8ad4e496be4adf68a9ba40a3f167d86f2663f4f8dccb65a045fea840ae77f00ec733fa5dc2879b7657e8e728b6c0025c8cbb8015cf2fe8df70f4bcd418548a2ef5f7822fd0045c4a5155d5cc673ac0ca5bf7a6f85c00381d808ddaabed5194c42d9a4a932b42277621be62faa1957c2457a2debfa08a7e6c67c5d22776b53595b1c719cc16166170d1fab111cb5a48bf90cfd6019516d8a235375dad", 0xd3}, {&(0x7f0000008540)="653a34365a6c464bc6b533265a6713c660363b7a2a5657742f89d77f6379b35023376a4ccc0ba58ce4fbad08949ed8d9c29a88a14b0b3fe66d46cd74243b0236796b67e81655c97f8ba53e13a55c0a462a71c41a257238d8c588929d4fbda11d0920404f7f7d46f1eb511238bb5c19df040ad61a88e8549cf4629352dc9231d0134141d5fb2e9e6131e74c3574ce24d2af664629df4a622d643828b844775c7b6500160b6e8f86e498099357214509f27fc6247d78fb2f473088339a50495bf8b7cfa3ff677dc29ddf4cc3c3903dcf744935d26cbfe6091d40cdd6a5d62379a276517983e49de0d8e5a18abd314ac4", 0xef}, {&(0x7f0000008640)="3960b134ddb8d8249d9c7e7daf922155ca7cfc209beda5cf8369d7250547e54a0a2e106e5df7b7063c9f2e14548b90e230b61fece70594792eb11551991d9b7fb8b56afa169799fdb381d7c75ffa465b9e1d86bcd9ea9cd9e1521628ed5aa6a57a9abe1e3a9d45b4f13a25418749a47350af66e33cf8bd4fc7e14a5712f5f3bde595ddd0934628e8ab42d46adfff43e40434d19c9c0870e46a88304731f1defec315af22ec8f3be3466ca1001a8c6fcdc0d346a85f8850af4370fbde8f92d2003357cadc7bf35a435cf2c6e5f2b67d54aabb3cc2da98f1eac9d78cb722cb1f1bc49bdc32e262e2ddf50372072ddd3b7d302f2e4682807529b4fe5c9625d62e32d18f531386bf49d95d5fd685d5f3fc7caf6830b691270a2ef96903b83e79ed18fcc2d8fd59b86e32a92f5629b9ca7580959d595396d31976693741c39179e232c0ef57a96a546c594b5628ca862ecf639985cb805b876994aec44d31101e8a1531a0b29d947b20d3368eec3a185395dadcd5070569015278db4ca23808e87fdfed7b05bd586b8295ce2a5bfea8362d5d09b1a0835d7742d5f03fe1ac44a19acb864f3f11830a4e3d598ec4327daa5c45e7da6a78a277357af7ebc8ac4645fd34e8e6467f802411e34c922e9cfb99b21250b2d511cd66456c7c174edfc4f6c1546f33a4905fcefa843da569be88350d60f0c0ca7a94134a2f1369d196d9dd4382a70f7a9173e6b89ee1f975b8128ad6453d3f1ed9f81ec9dc307fa4e07855b920f9276460fdc2b03ec94ab35fb0f5489018633f9a9a390531f6396d259d03ff47ec54846370d6a1ef1b405abb6607ca5cb6a58d37824ff6f6177da0edd10b5b08c34abaa93934ab06302d42fcc3327454922dc04e8a3a4ba43c0db4f282fd9e9f08ec4365fc21088f180960f6c499cab0960d67edc2ece5b046e47cf41b43b5e51273acb71af774f557c70640a68afd38e80d0c19fc1fe330906fb7b18c2350d6e0a5f9a516e72b27b41c2d50f5ebb84ee34ec982773e74d8c597ea4092ef5697a72d836958f54ff0af51822ddf07dbc58840d04098cfb3650d6dec9baf6634d175a0e3b609cff7909bb46aec2249aa0455edb7d919c495fc0fddd7054be58b099d081b99cd526225cef7649796f5a964dc35ea3ef9efad79155501d62bcf110d2bf49622fd7d17fa3568419b03d9802f771b9a06e3e5ff56ee78d7809295399beb0ab913f7cc295455ee0376e0cd870a71f4004bd14462a57034e58075b14a70166a2b1a07f96df4d392366ebdf77d333d613fda39169771c65acdf16fa7cc8b9feeb7d2412e4631066eae0222eba74e6bdef4083f848c23baa4f9175f30a02fd99baaa1941cf319bf7daacf0422fe2cabe8ab62a6696006735655363875cba6cae80b21eefad3430320959277bf4c5f4948b1e1e09bcb3f2f90fdc7cf7ad36114c852534ae70c0cf4d19f6a450bac1babe019852d3f4aef7c6461b031207e92ea9c6c38fe2ba7e28f3d79d39efdf7dbbc721ae90012bfdf3d1dd7288e5749189781c6490df473ea1f7f13890b383001cdf56f2a3647ad2a6406ff1a78475b283ff1ac0eed9b84e14f6431e90a68da77db09e4b131c0a859d3facf7bcb99ee8032d555fe0d82822bef24949576be4163d45605c6530c2a632e40b8d471bd3f6a87c1c4b0b0db05b3b1211634f93a0382312c263a2238255daea75c800d7df46a9925882d3cea8abd50fda1bd49d385325d99e935dbdfc02f26736f2ccda2c20619ef4ea783aef87910d44d19cf032562b332d6ce0ed4aa1a277f14b1d62ca55fd3f500cbd95590220736d1a451cdb3bc6e21e14e722d189d7bab832c8e95e28cc1a4862f6d5966ff0ca9b4569cea6e61e1b31fe47672036c79797e0bc2f4959f093fdb72df4f026e30c04ab730691dd25967c15f19c05981580b1ac3ef3f09a5a2b1ac79fa2dd749dfcef32164234bb115f6cbc6d2839893be27697ccf96727780f7a3b10494e677503891667175fecbb61c144eedbafb9881f0fabf6588cc6d576a7802436a075fcc6fdbbbadf2ecb78522c8a796b031f42b88a1a8c933d6ee8dea401e064b8c7b4bb11e3c4e9588655f99715a719f8ffbf1367d166e93449c9d88b4f02e48d494c25de12c234ad6006858c987bd1b20bfd3dec5bdc09b8ec5506b0418fd5405b189d681379c7d7b103ec7fd3f5065794bc7014d30d5f2fc8c5819c2766a6547d71a257b884f8cc28d746c072d99cd60d95662e0eccb6790384aecd50dde4ab5068c0e0cf5ca0f54dee8ba21bc0daa8f32ff2e871a3104cb329aabba0c4626cd4e52f77660ce9b5bb586bb36d9b8451d53ad407f2441e3b5364d822fc9f1796527d037fa077ac6722eea9ae0a3c4c3732494286120bdba8e954c7c66214361a51a18957a06f8958a9a63c795ae49f1956c66cb8a5403081b2ca93130b5431a9181585c72ac10a4abb1f8719c72dfafa560751a2de9d63c15ba3bb4987edcf4ca7e2086ccf4d5b0f618f2a24e07fd59ab02d57b3153e962b55dc12c097a62c3ced2e0ea12bff7c41ec0f276a5275022b1b459e5568fb590e81435c7f53a2adb22ca4868be96045a245926424f9d562b956d4b8952482b1193db7dc2148bdb924770ada37cb0eff1c9446a20bae3e18965dd86a1bfe322706f27d4e986fa119ecd9668d0d2d9e7e5c03a537bfa2d395ec348fad62405f6b1281c5a0ceba3331e442afd62971c0fb04496e96c66a63cf5e6dea14f43bdfc0f667b6ec50cbff3c731a3e4dfdf9189db8de7fd0ddb7fe29de2fd2633586946013b2c4b0f408fdc213a753b59edee774eb9302374f80239dbad4fa8c1d84142821b2ba89b23a006c526f69807809e5fd407ccd750038639d78c4b6f1a45317da11e4e826aa5927105b97871b137de46d61a3a7f8ad49f2a93b7f95b8292dcf2cdb2acaa3295ab72dc9dfd657722a4253df647ada62266d9b9d61300c67f2d2a48f6c2e1da327de52abf58cb2db675920111fb34464ad2a6fc511b4e73613c4b7eb61aaf6f38b738e73aefb307d95c609d80aeefd228d65f8be99497b501a7cbe371195315f265938036dd6e755e7391b83e9d886139847193f89a8cdd4e6383a0e98e83bf7a45f1f2fc73edad450ecc50ee4f368f811a6a82f931d26cf91195540d2793b81a22a5afeb35f59c932705a1b7c1e781e9f9d5fa75e7b6bb7e2ed12d8407b3523ee97c9f027e87a3c85208b4b105c5883ee86379eaba1ef9283e59ed6a54483d7146e7dff2a4e1e9203dc6073b82161719ef8307721f43fb053ae76102c36cb04017b9bc150f56ce2ea3809d7faa1d46a4745a173c13189360f9e13bf832fc4625fe5aa438f79accbb55de2160e96530e130d6912d95d4fd809fd8475ed3273dc1ac9575ce3205613259006a3ea991549688daea9b93d50a978eec49b1b675e40c52b1815531fd564c0502f03d985ba2b2489029ee8167c0bd89852494bc609ced1516b803a63e088e3e544de454af3ea35cce56cf8214ac91687e3c78c338bc8a93b837196201a445e1f6568e60a38de6538169f8e50b19a289d82ebe11ced3251c181924724a4dd78edf16c46e741af87adf627ea4ceca60c5626f301fcbd24832b0951dcf9afd8bb11ae6d4676224bd53e3cadc18d83953673f0233879157a7f610be7cad7930ea7d88a192263764b279f3670bbdd12e0f71259cc8274541e7ab3a810dd51b9931c0cde907cf8200cc5fd6e7f0a711e0fd38f6fc6fe2530ac2533690099e460bdcc776f282de5dfcddeb404d6e29b5973d2527b316e9db2600096578e8010af7aebccce3915f5d974be60c889556f87c83764ef9b21e728e2098735082baaf1e364f8a9720c802aa6e941268d452199f45dabd0b6794c3e8f6db430945828aee09f105feea67956093e91fa2ccb118034c10c5686fe27695f379d4b746665a6a8d334217cb4773085797af10b3b05098b05a2ec8a5472eee5fe96cdbd56b5213d82c1551be67af4decb3b5aac6ef3b930e2006270d68207c476203cfffd1b56b605808f6c41b097b507cf085a63268f85fbcaa0da94578764e2e18f6e592609a811826d6a28eb0842a4b24243c2bce43d88aad383d500cf447b33ec88cd190321ec43f327f86fa6df9465aa8b04d9a4c45c32879b5ee8e9d1496a7565dc4f7e33abec2a65e342f809a5b20a28c00946bf090f7c4158ea58396a68a447f2f22b3d16297aee2174f54bef72aa0e556b730be0f97e96b4806c4570409e09bd9b34cf84644fcb4f8e85a71d0c6de9d1a8c217cf63ab7b5604ca2b3456b74d4c90864094f010200ca8529de83a3c339ae2717f31358d6d32ce4601e6e50b3b6f1b9ded705a8121c5486d14a2e403c5eea18e37b57688d14f62439e1a5ebce0cc894817504779588b1f606cfc8cbfbb50beb57f53887c70892ee2e98689514f8f07f3f873fd36f7b1d5f71f02e91969ab35dc05fe48856f66e60f9b7dae042f3d62959f5f270829f9f74ca482b7d1be1640ae3995dc772b911dda2630a04d755297e85c9cd65ba9fed53411707feb3132ef30754605b8bcea375172bd4dc2e77a89ac52bbe79ff94988e72e26d8e7aec911020ebc85f0791362211a8f587333f2c6193f734011ac8c944e3af9852780a5071f1721ff95be5901fe963d648aedfa8123bde49d9cb1712f79888c0ce01cbcc468cdb5748b3207359a5c151d4257ee9cc5e677eaa789fd6d753f530d72405d090feb895110fa6e03d7cad8e3786bee8d821d4815653f5aa57b7230047b78024cce1923a6de777fe47e980a48012620f75798b374f530e7b9b46c6477949c54989309ab50778b6a90cbf5a5d096602ecc736ed743546bda5bb4cad6c84a7322f6a387afb1a7eeac01bbb2c4bfb7ded97d06b3ca2548b8ec38c2d1f5490ab09be22de4274db0e001763a27030fde34674c875207a37fd197f289f6250f8195a46ec724fd9f6b74fa18cd62521361b4b7cc42ee52bfbde8e8ca31c58039d3781da78a6571fadb7095767587a57eeaf3d0233acaf34c5578f1510b9458479d5ae731303ccd60eef59bc968e3e12054d30ed8017ee73062acb6fc8179cebf2b97b35198a6dff6716c1a69a2f57fe1172fdb91d8944fd91c6088f3bae4c736cd4a04b0a55170f390bb3414c0bfd8e751b818889d530933c6eb2fd27256d2368b9c18a27698ce40e227ff2dcb9b416fa303b7e28603642c824cd3d7b57ed9a9968501acf438d56024a3de1de9cf1983a906fc7b0df3df291a230e1c6eb84f846b089d8eec5a3c3695d1e7cf12496b468889b89fda91d15e3572b608853d001c79897a3342759a9d12e6c7a79e07a4293ecf083a0e165c05b9656c46bd402a33c177f55605d8cac9eb4ef6b2e07265ddaf361005f4174bcda4ca45d987e60780b02de300a880c0bc094fb052c3c8e22b2f07d49393fc146d5a287e82e78fbcd8cf1df400e6be04d7a9760ed81700d241f31e6dcc910bf6e192fdf058a4ac7aa761cc26eaa8fa83d58775628799e8495214b7cb387e89ef1db501eabdad5cbcb8f40e7c714642d79475de10ba2e9774148e6a38ee88fe7375daef85645d1ea8b3a0001cb7ce1f647526f050ab928e524215850293ca4c4e3a9f7f49499019d3950bbb53ca9421aaad909a177ca4690040ba8bfff527afdf31778d8b9c2b86c0ffe2ed224b9985f258e9a01381fe3b1da68c9845ed9229c8fef8fd327c7264cfeeac61b5abc1bdd7f3f45567280981ffd2129af50c09c00e0382c777b984d5bb708e3b184dfa2b64fcbda6407", 0x1000}, {&(0x7f0000009640)="5e5c5176a6a62b034409da1324a58e8c873b0d3769bb8818b8fd244314bf549a32e8d86ea36ff497db2b08240409da74aa5ba10c5e8321306d262387465e04a1b659b31ff61a5ac46e95e88c67e2be23fe6bd0bf90b654bface0350db43918e7d018c02e67df02d733a312cdaa233fd56afbc41cbc19558e4f4d1e5492859dd0fcfe38b91e55c0de870d5bff40b9069c51f9c2a45a7e53b4feee5d39419f96901254d1433e11dad556188b06c89d38", 0xaf}], 0x4, &(0x7f00000097c0)=[@rights={{0x24, 0x1, 0x1, [r11, r0, r2, r2, r1]}}, @cred={{0x1c, 0x1, 0x2, {r3, r16, 0xffffffffffffffff}}}, @rights={{0x24, 0x1, 0x1, [r21, r1, r2, r1, 0xffffffffffffffff]}}, @rights={{0x14, 0x1, 0x1, [r1]}}, @rights={{0x18, 0x1, 0x1, [r0, r2]}}, @rights={{0x2c, 0x1, 0x1, [r0, r2, r2, r0, r22, r1, r2]}}, @rights={{0x30, 0x1, 0x1, [r1, r0, r1, r1, r23, r2, r2, r0]}}, @rights={{0x14, 0x1, 0x1, [r2]}}], 0x118, 0x4000000}}], 0x9, 0xc000) 00:28:47 executing program 1: bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000000080)={0xffffffffffffffff, 0xffffffffffffffff}, 0x4) ioctl$sock_ipv6_tunnel_SIOCCHGTUNNEL(0xffffffffffffffff, 0x89f3, &(0x7f0000000140)={'ip6gre0\x00', &(0x7f00000000c0)={'syztnl0\x00', 0x0, 0x4, 0x2, 0x4, 0x1, 0x20, @mcast1, @private2, 0x7, 0x8000, 0x2, 0x9}}) r2 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000180)='/proc/sys/net/ipv4/vs/sync_threshold\x00', 0x2, 0x0) r3 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000100)='/proc/sys/net/ipv4/vs/am_droprate\x00', 0x2, 0x0) ioctl$IOCTL_GET_NCIDEV_IDX(r3, 0x4030582a, 0x0) r4 = bpf$MAP_CREATE(0x0, &(0x7f0000000340)=@bloom_filter={0x1e, 0xadab, 0xe89, 0xcc47, 0x30a, r2, 0x6, '\x00', r1, r2, 0x3, 0x4, 0x1, 0x1}, 0x48) bpf$BPF_PROG_RAW_TRACEPOINT_LOAD(0x5, &(0x7f0000000440)={0x11, 0x11, &(0x7f00000001c0)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0x401, 0x0, 0x0, 0x0, 0x3cb}, {{0x18, 0x1, 0x1, 0x0, r2}}, {}, [@map_idx={0x18, 0x4, 0x5, 0x0, 0xb}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x1}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f0000000280)='GPL\x00', 0x0, 0x2c, &(0x7f00000002c0)=""/44, 0x41100, 0x60, '\x00', 0x0, 0x0, r0, 0x8, 0x0, 0x0, 0x10, &(0x7f0000000300)={0x2, 0x3, 0x4, 0x9}, 0x10, 0x0, 0x0, 0x2, &(0x7f00000003c0)=[r3, r2, 0x1, r0, r4, r0, r2, r2], &(0x7f0000000400)=[{0x1, 0x3, 0x7, 0x9}, {0x3, 0x1, 0x10, 0x3}], 0x10, 0x6}, 0x90) bpf$MAP_CREATE(0x0, &(0x7f0000000000)=@base={0x11, 0x6, 0x800, 0x401, 0x81, r0, 0x1, '\x00', r1, r2, 0x5, 0x2, 0x5}, 0x48) 00:28:47 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x8}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:47 executing program 4: bpf$MAP_CREATE(0xb, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) 00:28:47 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) (async) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f00000002c0)=0x0) (async) r3 = getuid() (async) r4 = socket$nl_generic(0x10, 0x3, 0x10) (async) r5 = syz_genetlink_get_family_id$ethtool(&(0x7f0000000180), 0xffffffffffffffff) sendmsg$ETHTOOL_MSG_COALESCE_GET(r4, &(0x7f0000000980)={0x0, 0x0, &(0x7f0000000940)={&(0x7f0000000680)={0x2c, r5, 0x1, 0x0, 0x0, {}, [@HEADER={0x18, 0x1, 0x0, 0x1, [@ETHTOOL_A_HEADER_DEV_NAME={0x14, 0x2, 'vxcan1\x00'}]}]}, 0x2c}}, 0x0) (async) r6 = syz_init_net_socket$x25(0x9, 0x5, 0x0) (async) r7 = socket$inet6_tcp(0xa, 0x1, 0x0) setsockopt$inet6_tcp_buf(r7, 0x6, 0x0, 0x0, 0x0) (async) r8 = getpid() socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) (async) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r12 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) r13 = getuid() (async) r14 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) (async) r16 = geteuid() r17 = getpid() sendmmsg$unix(r10, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r11, r9, 0xffffffffffffffff, r9, r12]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r13, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r10, r14]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r15, r16}}}, @cred={{0x1c, 0x1, 0x2, {r17, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) (async, rerun: 64) r18 = getpid() (async, rerun: 64) r19 = getuid() (async) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000b80)=0x0) (async) r21 = geteuid() (async, rerun: 32) socketpair(0x0, 0x1, 0x9, &(0x7f0000001e00)={0xffffffffffffffff, 0xffffffffffffffff}) (rerun: 32) sendmmsg$unix(r0, &(0x7f0000001f00)=[{{&(0x7f0000000040)=@abs={0x1, 0x0, 0x4e23}, 0x6e, &(0x7f0000000280)=[{&(0x7f00000000c0)="4f1268934c4da557c27dbe6e29d5d50ea096d07c8d4cd889f64bacee9f002f7625d3a13c2d050daf4ef51db0640bf5d269bed4a9c25b92deb1c808d7cf93c9c6aa3445229447bd84526441f07f71aa131b32d02c5991295aa435a8221ed5ebe30340aa7c0a22bcbbbd1882b95534f4e7899d5836a9c20cc26799045c8f67a5dab116e1ae697897c9d7eb02397dd38a9874ef4e447467d875e55fddab8d84af6a473b29316f53de7bb4cbb1e41cbe7caba0445272b9699a57d747b1e6a6fa0b7eef10f3fae80877733faf96270a24b5cde75dd484be41da7ab8406d0176ffda2b49519ac98b7174c7f8627f3cf1e4254f61a5560c2f87cff98f", 0xf9}, {&(0x7f00000001c0)="aa6ae15fc2430cdd0852feac4c44a0323eb5a322f20d77dff3fa45e15578e5ee99260ee5eeac6115518b84cfd771569cbaa2b64f274b57c7dbe75215e29d9c5c97d1b1355a020f47ce9f03966675997ddbda2e1bcb132f08c13fc0c3b2999c674fefe5e03735d37fa2eb356ebf0ef80e7711876483c6fa3c6acbb2a78e884d5db16997c2ce4acd0e3ef4ac5608a40974935d1e787c10909c597ac619713ddb8dfd564559e91f5fa7", 0xa8}], 0x2, &(0x7f0000000300)=[@rights={{0x18, 0x1, 0x1, [r0, r1]}}, @cred={{0x1c, 0x1, 0x2, {r2, r3, 0xee01}}}, @rights={{0x14, 0x1, 0x1, [r1]}}, @rights={{0x2c, 0x1, 0x1, [r0, r0, r4, 0xffffffffffffffff, r0, r1, r0]}}, @rights={{0x24, 0x1, 0x1, [r6, 0xffffffffffffffff, r0, r7, r1]}}], 0xa8, 0x20040000}}, {{0x0, 0x0, &(0x7f0000000ac0)=[{&(0x7f00000003c0)="fbef377aa482a296cacc9595527cbafb72d5733c8dc76a65aacb75c03898fc3f37a06fe1196c46fc6a17bc454bd6bc3769947c4b97e77a2275f381b66f40b8dbe0f80faa497c6d1674f1bb2c7a0898ab0edc39b51711217405b30b4c7bd7b077201c6eb8f61d6e6226aed0c91797b858217cc1650b3a43ba74d4099823c14a2917e74535fa9d512cd95ce7979da2eeb59a0be51459763aa419c0cd33de7cb997882182889bde144effe49fe1a37ca6", 0xaf}, {&(0x7f0000000480)="4c3a3ea4052484c161d6e5cf18180ea03443ec922887abd649e61bb0309dd92014edcdd18497e1f6b4a4d7cce80e2ea4541c9ac7c3674d3641770bc4b7cbcdd82cc3641dc418b1e1103deca5c51703494c3229df184f852bf7a322aebe4898f9e1baceb9dc", 0x65}, {&(0x7f0000000540)="396cd6dc1dd898c175c691ef48e5", 0xe}, {&(0x7f00000005c0)="e4ef3a93fa1a3ee770710478bfdbc78dfb301374b96e80659b49229bed6f57b299c0b69c406aa635cb0f30a156ac574a071027f80994ce130cb0d54f419e53800027803372dd5298f996109f463e8db58d1af6a99c9c93c8467a30cf1f830f9cf011b4ed26e13f380461ae75dd7893d47e3477785d1ed6bb0a446dc0945ff07c2388eb2e05b5a9da2a3d10aacf9cd092c83fe41b2b261bc7aae7767af7e259b667b4a2f4535353ede7d2fc9172ea9d796a7ac8504920360954712f", 0xbb}, {&(0x7f0000000680)="2445f1ba09ccb05d9cd51b21d387b7f24a7d17df0b3ba289df1bd7f1e33331549b31fc5830f6e6e858eab99289cbd4d1d4fb2b516baf8c8224023436ad9e2962f5995d90e3821242e64a8665fee5831b47dec99d43c706420b0e72f69daacc10c20454182551a3e29dd39b0d3955e1fb98d5e6c70d659830a220d94db1c66c2342ed9d83cb236bde080ba1d31a3bf90d1313b99924c06a134d33b266dc0c095dfad30163d62d123019de5f2df553441205c50c31c98fcd", 0xb7}, {&(0x7f0000000740)="8181cfc5eacf34bf4f87cd238639578cd31b65e1c6486a3df7fa39624f212462d047ab72d17e6c9aaa5ff83c3da7e780d2e42a31eeccebf1da29ffcc0b3a9889e88424286aced134160d687deb55054b63caabd480ad8557246d73a054d2e02471f93621869c912c20938563fe08d51a597cf608b7f3b67217cc704904871c9bd74bba1077aa1aba16db810989ac9711cf05ba44801f1699712979f2b0b69703e44e286633ba5cbeea8d540e94de6ece738fbbb65cae2100f0fade3c55ec2fd0fd7a2589da0cf33bf8958656fe92115cef922af9c07b5caec479a1a3bc2c54cf9c911c003cd6c8f703f3", 0xea}, {&(0x7f00000008c0)="a8fabea5b7de9ce8d1ebf60414d8909c9395b52882c9e2c29cc59f85c2cd0afa7353bdf9b123267142193cd0777fd6b8544d633ca86aa7801ac86b65782dab40a1ba1ef8e0e6f2b05f33f9b4a00894d9d0702dfb02b34ab5fe68dc028e1725ae4b729deb97fa01a1fea350698e85c59bc1111090d08b4034712aa8e00f9c7956feac0a6281a405be572e596b04aeb18919cf8001d0f6edeb10a91d26b9c0c49c5874046ed3dc2dfa14238534de2bcb29bfb6eb82ff6ca7fca46507537af52a504d8cb224b8d6e0587d37739df6c9f79874f946c057c2ed0af8df", 0xda}, {&(0x7f0000000840)="f0c424aed7e8638745fa5aaf081e8d634512d9229446939c31545bb515ba3fa8fb", 0x21}, {&(0x7f00000009c0)="76db349b779f1b83b5919546ba3260e6b76cb44362abadaad53f0e29667b72396631c262d8aade5b71ac669328663b9c942c0bab788f49b42fb6bf9b9aebb5c160ee3657e364b4932e8205893898be8dd56c84d5edf050ab32355d8bcd1bc5f2afd05bf056f85cbbc4740dcc1ee4449f35f41eba5ce5d0d6bdf3d736b10cfff34ee42bf6b252889fae21a31972ec9a7b607d510ef476832af1d8b96341eb59719aa2", 0xa2}, {&(0x7f0000000a80)="e4", 0x1}], 0xa, &(0x7f0000000bc0)=[@cred={{0x1c, 0x1, 0x2, {r8, r13}}}, @cred={{0x1c, 0x1, 0x2, {r18, r19, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r20, r21, 0xffffffffffffffff}}}], 0x60, 0x4000}}, {{0x0, 0x0, &(0x7f0000001c40)=[{&(0x7f0000000c40)="ad142f1e94dc19ec3a17757f6d986b91f83a75ca033688b035ee3ea2ef237c66bc8aeb6e7cc3cc78e28da047ee8e0503e8bb6303970f089a7f398605dd4c9d9d93903aec4394e1283d6862188fec396b39dffd5e172f5382a075e60717ad7e4e6aeedc30abb026c48a2bd7ebce850a4a0fbef88223621c1afa4152196c408dabd49ad6bcc0af0014d355e5329c472affef59d3c4e9e3b4b97cd392e7c72151aef1d8d4857dd6bf45e6c2a158ef297a0c902e2c82523e8ee03827c2a53a089aaab4177e74143065141e6bc3ceda8cfce5255faacace01f9d400362a6436e844393abe7ff745ff76e3653df2d5529a52572a55e62d773efb2257ec6c7d518e2dfdff3275fc9eabbf98881994527d18e94541adbe35aa3c4845090f48e470363ef93fca7df827e4fb3997d57dc1ebcdac8f864382921220112a214be818e740c1e59edccbb55446746238e42697bbe061e0780b632d2347aa14df615cfc26c19666d4db36a8fb828e26b7eb938a104899c4cd683a57e47970767021eb2e81f124f218f62554dbbda116d1fd9eb697edb764067435ed26e35de176bc0694758649310a91c3676bc33ebc82b7f14b2c9c4043f9aae68a956b06c407704dbff67e0f38ed98611f7b97609aaa76d8f2b1f7c84b9b6a0860b0e343275b35ec375dc18dab93e0b26167af76c5855aa64d96f52819a509891628a508d888471140b220a1fe6f866e103c48f83f41ef958f66128ce5587dcd0b8f5f10ca3730f4861c732e4ffb35da2653fc24e3cd68018acf1281ac01ea45b25e94ada62bd6780c3820850c59cdb2c6e799ae85061b3ffc16b60d7122db7e48da9cee6f4f1dbe035e06130ef5a829f80a3d33e42b55e1ca9491538fa049a639de8433c414f68d82ae8daf554fb9d6b61ab1f153041168d9bc87f50dad7bfabc64e358214f58027c03358b4af441607882070b9b49d8d59680aef2831007f0da97bf45df3168be08f8abeaaa31d6da923abb3091f0df5b1c4400c81ff6e51b858298b15bdbfc8f45323fa39de57227c810b1f7167d4068887e87c17c0566a59b441c68f39918416e6ee8a31af1798f07ede7eaef9a29a0006a2c2d322c9b09c04fca2831660495cf9701bc40fcd6f2e23bf4742f36faffcf12185953afa5ffa8284dd12519f8cc3873134cbb6a80e28bb1d47712e41dabcaa68a77b0613ffb2a04e3bb06717299908b44b30da497d81cf399357788e3703b1c484b55bbb3ded33452823df4743a5101d38b93591d42cd1dd9bdcb9992b779505b1c3599d24be9d72cdb139eed29501a7e6c94e979304e02d946910676efbd5d64160355586e84c02c1bebfda6d5fb581714c75f47afec5d25e5fdcbc5f3cf519270b8126c2dbf0baa32a0b2db0ddd0d69f3e0151c3af74323ae83587ab32732e9a3a9111b604dac6e926ed384d5dd0fc8b9cb0ec8ce115f83d8ebaa8a636d0aae2c9dc4d8032572f3acf25fe8e9967f7a651acb8ebb8558af65d3a83ef646c75f4c58aa0537a3c57e1ccf6422325d90a0c4ec34610767589e13577fc046632d9b2cf1c5f4960b30b9adff5af55e5eeb4889e636d0503773fd22fdf60602ffa2172a0b8a71a3677b03646ca87d698caf4b98478a96da31d0593650a5f6fe0964827f8c7d9fd0a061cb80e8a1119eef99d66775034c98338286b1e22edeea400628f08f6f67c10aebf14895afc4794ec470720cab7f5178c2f4114740de716a8daba7b51005818e7090e9faf35948a25c00e7bf71e9234870fdfe36b11ab5ca640376a15c0fd99fdb9b45fedb5af7b65dbd3498eb7294e255b80adc46a91b568b014ccad3ffa265d1a5620a5469935c19f5595e739f6a3de1a2849cd2bb042d9747995c00e6133d1d0a0a40d0388329866afc32cad2ce4855bd959a10ccaab1e5511cca53dafd4c83c091ac6fc02eb54bd1b61d6b55fb84f21ba94728fbf6c98a197ab6823f909f029c2f89018bd69a005e5827e27d54de5cef339a67a84c7ca5089e47ed893f037ea27e5049722e711cb13dea0d602b8275d8de65064ddc457a1186c9b5fe769d163e8c2255c76e7585b93594274784b4a42069d0781f0096970fc52ec0895285cb453205bb7a30723d183f00d353d39f682d5c838cabd299fd78358192bcd3fa744c109d24db6bcf49d0975f1a8c5edb78212d4b74cb18541b73c18af159f7550e70a307831ca4736ea0c9ab1182eb008bb6a32ed04738e4ee89ad6ccaa05e76e814433777e09812968a28af783ade5661cb8d62bb8360b61329140e4d2a1f7ab7e82b691fe6aab9799e1c2a9221eb5ccf5cac3b37d32e3964776e7af56e7a6d779f0e082f11917543f03345443dc4e84e676fd7c8f8d19dfa4db6032d1a20195d79bf97bd9905dc1c6af5d3a9f4f197a192f9fe108fd55c0d48a303dea46b5c3c22d8385b64abc88d04f8bc990536ea9a393c822958c7108dafc5c006caaf00b8088d114ee2acb0e12f574fc49bbb23de0295d7fcba8394f4bda5bca0a3f3356ae916fe942b9b6c6f65ee38cd2aef5681262e3f7b0e1c71f8396007c7db9749cdc0b7dc10ee19dcc48d7f6d0942d2b94a1fb8e4e5f586311e1556dab192115b9174b1968718ea94e4d6481f56ede8cf73e74af5522cabfc150d2c84bf51dab9f2a6ca69bdb23abd7447c2ff80e20c7433c1c472123a3d36bbc568ef293587bc02326700a5ca2f472ebed5e21a60ac5b10fc3be0c3343f29bddd4c342c587be2f259a83f104b215708f77f81d5fb6275d7f213d064674beb2e393448aea6407458333e68bf3f45212ec3660e58d9285e77e5690133512e36e108ba7bfc2aa43986c9c33e0dfff2c76fe6d6a83e5aa90cf0eba3f5ba9a5f51b9807744d9039e3ac55d52de4a2a2ba04632a7b21dcb25680e484e1c4b7170167df52ea33744c806940d488407ea83a12b2ece874f4c903b6a369130fc1184b0ec8abc3ff65cc89efe7c25745f1567c34a5d16292d84327989edf73e4259b1e70d1a57dbdc29e85d1b84846ed0bdfcf4cdad6d46111d2d52db1ff23e34d507563135d2b45978371ae5d7c9488eb7adabd684717a7215351273a5603747ddae1ed0dc2ba08379d6956f5434b2c15c46d6b4e58cc56f0aafdcae80bc63c3700234811c05cdabf5761a5d025c33b70e21c58a4b64c66a9a0310d1046f763bed78a9999c3e5c7dd347a0fd86ed9913c5c77126144dffb9c113f3a0f5a18b36df9cec1526e903c25e740685b083608fc0777bbbae542039ca9b3fccbe5b636e1f8739e9e93585568827618ae681a981f739fe1a75b2c7a86e2e6240d19072dccefb1150f0efd552fe4c4844f748df12a8c165df43da80dab27fa7a09454212ee2aa5b81ef1ab6cc20b99d629feb8c6addc6214dfc532406cb7f69834ac3007dec2ef52e2db43422a866fb1292fc022c4db2cce3b0a2dcf62b31813a9ebbae3d14e8ccc02d5f220d8370c787253ce850317293f419fc828e6c48497cba4986df6c585ce3e06a4b1649c4a18378ea7916fe2d767c441c86afb86aa9db22746f6df444dd8bf33b8dfcf2570643380414406b35a7676f7f59bc292f589d3866b1cd52ef854e73497a5734a3054c4935b1d75737a5e394a8beff25d44bf99980e8c8daaba6b4f0f702c77a3ff3730b271aa2b340b5fd4ff78e5f7fe69f46f4e7b4faaf2e1a06d5f0aef24da7b52c030f9e71ce966af727982d2ffb0b3c34e7ed027a71f97be55788b48479ec4504f2739e4c5a83f08e28a23882f289c8e6b0bc37fedbf443167b9f8b69ced21477cccfcaea018e5923fcb50381b2e4d132a4c2f813b0fe9e56a0f09939bfc930fd5f121ac4a5d7322ad7568ca63ffdf9b47f637e42021980ae4bf4d8ab76c963cd00291a44dc60260a51277b69449b9fc175b61e4f1d545fac4ff876fee6d25080a46b02261b9f1b62fc6dada23204caaba4704cf1ef890584930f02d292be9facd1eb270d1429a36adb38b8850610518609adf822d98cb9c1e86153b4753f324e55cc38f6ea9f35f118b6619ccb8b11f8e28163081b5dda523bab7561a71fe425348efa7ac221552f54d4a96f054047aace8e457d0ebbab91e12b5b8ca40191ecb8477cc2f2edb5293978c17e1fb9b8046055e1bed3946cd9c4aef9c0a95314ae675a3d2f37d90eb9c84348f9973013e5073eab98e512df4a12d10702cdfb28c87a83d5a0d516dee13121cb4bb96669c4518041fd6e015822458d23ad60dc6099a00ccd5c0823677c0170faa42be04cae9000a70773d18e53be864923d03fae31be77e36c01ea370f4e2cefa12f670dee3c53a9160fe76048c11e7b3113f98cf0e78cc07e64837d1f80159d6e1adc3b5af658cb9365710c84d5264e0149a893c97182c3f6505008fede6f100b1e1f1b54c2c53a2c3bc72a292ca6b354f3d2ce49fcf6c2ce3dc3c549d194679f90ffbd1c8ce55c4d69476fc878ce2bde130b8d91e080659db5b39e3f399f2b8060f0fa2a6da00f576e98a2a5203c3f3cbe132cc32b7af795ff8891bbd1520e78a7499a09961c55ed5d52dfe3aa773696a04c5d16f5a6b07a2e500a06e34001020d093010b67dfe479a50fc356ae5a54437fe6b5c37f0cc9b815f8f81c47d0f153729c21155d70aa8504c39e31c8cd9c9764057197d5beaa895cdf48b08c107fc08a17252f4b68b946b9f40d800e865ae6b60656c0157174eee4111379f4dc5ba2f0cad0b5366cfc398e431c91458c114377ea93b83109ca7f48b7b5eac5090939de869098090d1e14aaffc58f7546df7de6e9fc0c0c5eb55cc572a2796da49e559be7d73ba48e8b15743475858d367fa5d30ab8e08b4f270760db24b00315332ef6491973ed8a23f296f7587a18d54a0c0923e50f7a596f9f6a489bb5c82d80bc7cf0b0e458e8313f3d9dacc6746b4b5988b6ece979196170eb28134b754e63bd7e8f72ba53a24bf36dae2eb323ecca6049ef6e8573240f034507a36ad2b64c5fdf002cbb119ae9f2fed395330bd3db9694bd82a6b930641825f050227658571f98a265f5829f3d0ba7417075d10c4d4f8316aedd6c95761ab498172f9cc331e15d524b6ed42558650476356c5c2fb26978fa54d1721c3d7e6ff6133390ffcedf4842d3a3c2a170f7882df3ae4f57460d34b1a288064fcd3cb82bf4b84e6f69d5feb5ddc1469dcfa3c2ca0eb46110b5ddee5ae65ff7ac74c35d562c76df64282d6998aa1e7d34758e2839187724f329dda822e021901cff38f326d54f6b29972bfdbe7011531d1d2c121cdb6620fd6cbdd95190705075cc0d6b822ab61cfe545b455a8fab3f3df005f7e7afa02e332fa466479d7ab0bd7d972d4191c8b5dbe2d96fec963886789af89d6f69b72ed6723d82ab42e46a9aead28c1a80aa63e35e40203556e47c08064936142af795dfb26be56a11b0508a42220c57bfa485e12d1de2841206ad410f2b2a77eeb579f106b1d80315ca7baddd6e5dfc927d3d5d7324f5aa0fb2522c9142320e27dbf13d8732d4c9eaa9627d3addd6d0ee8fbebf27c01459e0fde0cf867720c5ac033e99562fdc4585e332bb8cb3c4e643f390bcc3b6a8cf102cccf3d5f31679e77202f0a7a0180d40eb2837a3ce45ee13dd802e2cf86448caf93ebe319a28b8b64a964a978e15f9ca7c601869e2aec486490abf5308a5b09d350edb9b4031d117a7fc9c9437560e3b0aea664550d12c97b2b5615215fd8fcb7b04d4a0149e85cbba9dffe7ebfd54fc0081f6180541292531e4af82ff3e438f9bdc9bda186028e10515bf3cd7071d2ec00ddd88f171596fa036098", 0x1000}], 0x1, 0x0, 0x0, 0x4000}}, {{0x0, 0x0, &(0x7f0000001dc0)=[{&(0x7f0000001c80)="3814051670d4109952abc445432cdd2b17bc640e5b204ce375555a0dc31e33f9e5118e407d27850b8e2b634b73292209e24584d2fd54abf6d48af78e22f4746bf8bfbfeaad6b3274d8366da3b0f47e548c6708ab38227b46a089b20812a0a42901353264ca0fda09238fa6245e929ad8861d1cfeff2eba1139ca93f2bfc398c5334c62b51ba06c0c138fbeae9c62d70f4b9f0208643e3bb010a3fcc2cbcad88f769b09d31538c0084e173aee8a", 0xad}, {&(0x7f0000001d40)="70c14ba0ddff01ca84f6fe47d99f9a3bca5ef3f0a48ddb15b97f40da34b63c536d4352bd5aa7d665f1daf123a221c8f463f516d6c50cec3e001989935aa8aeefdf37a0a722186c86e0dfde370d910f98dd365ade37888d2d522e033aaf161e3c0bf7429b1539379801dc5b4d1c39c750187fdf3d380f4f928d240282ce6c29", 0x7f}], 0x2, &(0x7f0000001e40)=[@rights={{0x34, 0x1, 0x1, [r1, r0, r1, r1, r1, r1, r1, r1, r0]}}, @rights={{0x24, 0x1, 0x1, [r0, r0, r22, r0, r1]}}, @rights={{0x2c, 0x1, 0x1, [r1, r0, r0, r1, 0xffffffffffffffff, r0, r0]}}, @rights={{0x18, 0x1, 0x1, [r1, r0]}}], 0xa8}}], 0x4, 0x80c5) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:47 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xf003) getpid() getpid() getuid() getuid() 00:28:47 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x9}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:47 executing program 3: r0 = socket$unix(0x1, 0x5, 0x0) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000000600), 0x2, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000b80)=0x0) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r6 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r7 = getuid() r8 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r10 = geteuid() r11 = getpid() sendmmsg$unix(r4, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=ANY=[@ANYBLOB="28000000000000000100000001020000", @ANYRES32=r6, @ANYRES32=r5, @ANYRES32=r3, @ANYRES32, @ANYRES32=r3, @ANYRES32=r6, @ANYBLOB="1c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32=r7, @ANYRES32=0xee00, @ANYBLOB="0000000018000000000000000100000001000000", @ANYRES32=r4, @ANYRES32=r8, @ANYBLOB="1c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32, @ANYRES32, @ANYBLOB="000000001c000000000000000100000002000000", @ANYRES32=r9, @ANYRES32=r10, @ANYRES32=0x0, @ANYBLOB="000000001c000000000000000100000002000000", @ANYRES32=r11, @ANYRES32=0xee01, @ANYRES32=0xee01, @ANYBLOB='\x00\x00\x00\x00'], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001880)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7ccb3494e59c5e8ce8431ea3af6e0829fb4e8f523b80d22325b6151c5b4964fe11e8bd5217a714d310377c0ba5b19fd7181f4417ccdc371e0ae73e42570631082b8cabb246031894a55c875641ca17c39192378dd51f40f5699222c2a3a7df70344754107a0c5f2d7aa04f10f09cb5c11aa5e53a7542bac86581894cdad0e224dc67976013a7597503b05b8845a7db5314b17f6a9ba3893db1d5f9d2de5f0352c393cc072fdf967ae68ca17f69844bcfd753", 0x144}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) r12 = socket$inet6_mptcp(0xa, 0x1, 0x106) getsockopt$inet6_mptcp_buf(r12, 0x11c, 0x4, &(0x7f00000000c0)=""/101, &(0x7f0000000040)=0x65) socketpair$nbd(0x1, 0x1, 0x0, &(0x7f0000000bc0)={0xffffffffffffffff}) r14 = socket$nl_generic(0x10, 0x3, 0x10) r15 = openat$ttyS3(0xffffffffffffff9c, &(0x7f0000000c00), 0x55d7c0, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c40)=0x0) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r20 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r21 = getuid() r22 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r24 = geteuid() r25 = getpid() sendmmsg$unix(r18, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r19, r17, 0xffffffffffffffff, r17, r20]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r21, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r18, r22]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r23, r24}}}, @cred={{0x1c, 0x1, 0x2, {r25, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) sendmmsg$unix(r0, &(0x7f0000000d40)=[{{&(0x7f0000000040)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000580)=[{&(0x7f00000000c0)="83cf82f9060b679a7f08be646de6b08f77678861332db9ca2cbca6", 0x1b}, {&(0x7f0000000100)="dfd4db580959da40b397e0c63e8b3d87c86cf1aea7d80827be861f6822431dc856c50df9a4fbc82cc040c0f5f03136f410c8acc146402f4aab40218d9c73e56c1ae8241726118267b459599fa4bdb7973a7540cf9accb13f1877059cdce7efc1cd706e8e50629d7edfb2a87f6d0639435828d5c4ca7219a89ce2fb9d2a8b9513cb84aad43d2f329ee1db7cbfe4bc779f66e3df27de4e2b1f51085f3f6861e5d319c07eb08e187122127e44b418467f2084cb28a1742be5b35d61e74d26fae23e86c59748d42c", 0xc6}, {&(0x7f0000000200)="5258b90232dd5dc3391d28ada7927fce7e855695ed8e02aec490081ea011305d2efe1cb20c4d55c5daff2bec67a98bba1480ac24368c9be46e8185911adf35a759306b67aeedc93a8644c4a6d9647ae44356cd731148d1e6b5bb18f566b2559d4c88e1ed0eb248e0fbf8ecef1eb3fb3d8e9adcc4ce76baa91973918834fe864b500e72bdbc9c4e06e171cff227880be452fc92dafa24b1b85e7a470925f1374ead14d5f255b2543b7fab6a47a9e82592305952f259398f1b3b297b27c175218258fd8f05376b5c811ab83a7a0bf6a8393793b159a13d935bef", 0xd9}, {&(0x7f0000000300)="920e4581afed1ee67dd8dc0a367370c293d97e3feb028eb72a2815b2255a6e5b1c5d533ff3a21a85cd6be9482b4cb6d8f2ecc638bcc06a0f334fd66bf75ba5969ccb9d4951f597318baeaa56cfbfb6c488f93bae6129892d27b46abf85ba5b85eddb06b250e0408d720da2b0a87ddfda4850c921e26bb18616dd57d9486cca4db6dcbeb4b247f22f20fb", 0x8a}, {&(0x7f00000003c0)="fa0cce0f4b622ba48d4826e60cf36464ca7be5f5291b96bd97594132b527492181fcf3bb61b750780cb5603a837c22e13a7a29cd683ca2c5183b1e1781ed1daa0d2120114cafbe6d1636180f1e091a487850660fd5a86653264a84294e4a57c8fc6fbb560c2161a4e59a02eaa48545a632df4cfc73aa0292d6376f113878089cbe7fc43bd127bd864bca06d63fa54c78c4c3d26012b831e76a6f94f16e813d88ea25e174c5fa363a91cda51ae1", 0xad}, {&(0x7f0000000480)="ff89e60b90cda2b57ba52787d3b6b94b1f39f4c1bb1ca5a8905ef8165fc76bd47f4dce35aa0cb9f31642d0ad473836e8262f866aa03968999e073b6fb0946747f0510c25e0291b7a9c6de36474a090f7701971df35aa3d3909dddfbe23aa103db519ade03fa8902ac14054733cd1ce648fb7e38f81e42f166abc8fcdcd9995cba61316d3f1091181e373644e03be0267a6e0548e484c7ff9928c4c67e968a2d9c20388d266eafc464e0344ca58ff051235686e1a0b13370130eb18fbe3f763ca5a85a3b9dd58717017a2569f39aa67b6762507d50857fdff25b03193c2f1d682e3c8a4b57320d4d15b3e1fb82a7a24f8e30b", 0xf2}], 0x6, &(0x7f0000000640)=[@rights={{0x14, 0x1, 0x1, [r1]}}], 0x18, 0x44044}}, {{0x0, 0x0, &(0x7f0000000b00)=[{&(0x7f0000000680)="43a9cc03073ccb92c777dcd1f269fdde4542e1a8b3e9f7bb81cd6df8c2fa5e9056814ce4bedf2c0bb323d4a659f9c0b4fec943f8ec19e6dce4aba0333adf98ec9c07b63c7c7efc33d03b698e687979d0f479a1a61bed0abab3ee6708ddfb5789bf15a5bf9633350f57f72e9c6f7d4112ab815b679c12ec7c326a8a42a6dd076f31f396304dfcffc7ebc2ad48b296159de59012d334918e8a8e74e950bebc", 0x9e}, {&(0x7f0000000740)="1b5d9835a1d0cf5a711dcd222d18e5c74e4005712dd07d2678ddda1c5c8ca1319ff36f11ab1c153a44c8b380e09be4a404d9f165318eea07744d3ec144c39d484a1b4264edf710d4bf861818d752b9d6bf68f378f8ec37cecc81a996bc09830779770887b65ddb6ed6e67edc22cb641e07e59e2ac4c3175cce4d1b763fe1c1a218a3f360e4d1622b16bedb7a0082", 0x8e}, {&(0x7f0000000800)="f611def51abed9593db5dcc8ebc03ee4d4d4c68ece61810b3373e9f42bb82691c8715922278c91c951c855d34ef91ec3af4cb2d4e120fd83", 0x38}, {&(0x7f0000000840)="13e37db8468046f8feccd3169267df31d7c53163b7a63b91b88c88e916cefee236d47889a2d3cbdb8a3609a4cb80d1f3c3d4cab69371517cc5bc4fcae2e825228c", 0x41}, {&(0x7f00000008c0)="72aade7d64b73a8be93eeb66b8f8bff8dd21310e83afb3f18ac6e1991fc33381852e3bbc91a4aea287abb11c23b95de72cd640db6277ddee4c3fa63dc757f905b76427559bee28fbedba690f7ca2ac3e7a79454be96904da256bba54fd737fa80f4e7f5f84b5828ba06aa86fb3a5af2cc3bd59bffe951e557c8677969b37c5cd0fd329984a6f0b", 0x87}, {&(0x7f0000000980)="3f653667865a9c8f8d59302ae454de6ab4182def183ad3c1889fcbc53636c7704da28e936e61f294d0ae25b4ef046d9a443153b2933cff6b6336aa9063ee6a7c5ea99fa439ff052a3f26250a80f6eb39c1d8f18bc3fba7b09255779571348a4909de7b23717f3fdf4f4ad7bd06440e5ae94b6d94c844583a8c3f87806f3e31ebc6207915e0082d314011b2740f32b986ae69a7caa4405444be44faffe11e7bcc6cbea08ef52717b67b38bb5890714fcf5e4831c09feefa000bd26aec0652d89dfd63a197e614634f99c8266a499919badd97286169fb7f7fea731b66907cb8008716e09ea4", 0xe5}, {&(0x7f0000000a80)="5f2a7b9af0c75453eb890b9849b306b4238a4de99d93a3bb48bc59397f0fe95237178f5a44ca450dfe16b7f03d45c38f07854668f8b0d587007022b1f792b1db483ffca6330c1d464d706ef3b2754791be94713c11bd51891183131554abdff3cc48070cbc7aad46f65f2ce85971f616083440e3e3d934a11689c3b2c9a8cf", 0x7f}], 0x7, &(0x7f0000000c80)=[@cred={{0x1c, 0x1, 0x2, {r2, r10, 0xee01}}}, @rights={{0x20, 0x1, 0x1, [r12, 0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff]}}, @rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff, r13, r14, r15]}}, @rights={{0x14, 0x1, 0x1, [0xffffffffffffffff]}}, @cred={{0x1c, 0x1, 0x2, {r16, r21, 0xffffffffffffffff}}}], 0xa0, 0x800}}], 0x2, 0x11) r26 = fsopen(&(0x7f0000000000)='bdev\x00', 0x1) fsconfig$FSCONFIG_CMD_CREATE(r26, 0x6, 0x0, 0x0, 0x0) 00:28:47 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xb00000000000000) 00:28:47 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) r2 = openat$smackfs_ipv6host(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) write$smackfs_ipv6host(r2, &(0x7f0000000100)=@l1={{0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x7f}, 0x20, '@Mm+[%/'}, 0xa0) bpf$MAP_UPDATE_ELEM_TAIL_CALL(0x2, &(0x7f0000000640)={{r0, 0xffffffffffffffff}, &(0x7f00000004c0), &(0x7f0000000540)}, 0x20) r4 = getpid() r5 = getuid() sendmmsg$unix(r0, &(0x7f0000001980)=[{{&(0x7f0000000040)=@abs={0x1, 0x0, 0x4e24}, 0x6e, &(0x7f00000005c0)=[{&(0x7f00000000c0)="b717bb341df74e9d6777781b0feb8659a05825a132cee03c2cb9ce8db16b232db8a87b160706c91b2c4435ec49d70c3aeef358b9b6333f3e7b9d2f4c220bd2c81c0975017164fa40b7625daf60ae8de69a11aed22171090c5b4ac689855c7c7a352be8660316bf97aaa49ea70c43afdcfbffb07af9abe5c6dbc752a91a044b27d1cf2101722387de490d2e6b4c285af79358ab2e47784df3785a8c23b51bd5ad898d19e5aa4376992b8f3816f3674bf2a06f318e9ef694addb309a9965d438a454701075ce2d2e888186b0a8263a4bae72028c3a2248023cf98ed081d9b5590b47cda6a183", 0xe5}, {&(0x7f00000001c0)="1c7eb0d6fdef006522ca793a7adbc3467ee2f1f2a63d8010bea062491aa13d51fec04b35453dce6785ba201f7259a75f41633a43e6e77afcf4a92b3bdaf2d72f5466a758f496ea94ef0448581609d14d122a8c0e8be313d798e2e4fc4791", 0x5e}, {&(0x7f0000000240)="bb7764035e5e3f0b6fc7d0a948caeb1aa98600d029a3f0dbb1e873df0db00926e5de2032334c5cb2d6d20b7e6f9ca07688dbac381cdc8d1356dde70387d838f4de0101c03941561e136b373ba9bb88bba45b4ba5147873767295e43e960adb177add5a90dff50cc8801f5db8411042a8bb87e6cccc02b2d00cb35b2d0fa893258748ed37549103afe60e23645acb7672aaa6fbd706113e4b4428b8a4a06b03bb5925b85da12e58e4ae50650c8e393c4bc048f5170d4aebc4b315c4d854", 0xbd}, {&(0x7f0000000300)="f65351bf52d2313687035d03caed565df5f2062b6748c84887baeddc0fc88327af9c4b73f0f9153fcb1c8b421580aebccbd2fd81d29c990f2b50690a24be204b450d31c868565183c65c192d84004a95569fde", 0x53}, {&(0x7f0000000380)="1657c6a3e4c7f7e5316393c595ec5782434800531a41c04adcc3f4738f8c566105042decda11488bc8e5ab7425844557cb806b0f941b43c553d5242b230d92015d5f94", 0x43}, {&(0x7f0000000400)="35ea73eb1953bbf932fa13842445b75647a26526dd0462b70d5a2b37b0e6227378d8bae64762932a7ca8feb18087ac0613d5835a783162f2827c63d2d1e88046ec381483a016f22432fda496f7737030f294e35e477cdbd4324fb8e6fd93faa9822764cb83d73c5c62ffea5bf641b4bb6c01072bed8fb925f8f627a147fdd28b31536edd0e9e18417615bacad5c1dd36df4cc212f8317601ccfb40e3", 0x9c}], 0x6, &(0x7f0000001a00)=ANY=[@ANYBLOB="1c000000000000000100000001000000", @ANYRES32=r1, @ANYRES32=r2, @ANYRES32=r1, @ANYBLOB="0000000020000000000000000100000001000000", @ANYRES32=r0, @ANYRES32=r1, @ANYRES32=r1, @ANYRES32=r3, @ANYBLOB="1c000000000000000100c7bd011c6e92e6e9b19c31b828e729fa2ea9edadbfc7154bc4c2794caa137c72e26567a79792d8c6d4b0fda86ee15dad83a07864864d872c8a283ca088648f", @ANYRES32=r4, @ANYRES32=r5, @ANYRES32=0xee01, @ANYBLOB='\x00\x00\x00\x00'], 0x60, 0x14000004}}, {{&(0x7f0000000700)=@abs={0x1, 0x0, 0x4e20}, 0x6e, &(0x7f0000001940)=[{&(0x7f0000000780)="a10f5a30325487d250c4bc57e0648929ffa2eb02e0f9a4f78de17ffb92216c6526944d71755f215f3981", 0x2a}, {&(0x7f00000007c0)="1651d4eec4090a2a3c4382d70c12a652ece34aab27ab47ee0511e6b28c68d3095835dd0e1ed1c6afab230885cab82a8b7df985692fdedc2885ba20fd5b01f77719e8833e339d2d7b2926096b71092281c7f89d084d87d33231693535076f277e542edf43ad1a993edb58923cb21a7dc3c30cde45e1d12b13a1e7d03dd04b649ef409f0aa7a76911817c0466adc08", 0x8e}, {&(0x7f00000008c0)="81dca8cecc10d73455ab3c5a3d931ead7dbd174dc93df93cbd216ea454bae74d6a51f383679c7fcc2545ad25d5b2586af492498fc7bd0ffb95dc2b00b20b544359e17d1478548a68d1fc", 0x4a}, {&(0x7f0000000940)="301ddef419afce3f365a8fd662bc5a0108b5b31d43ae7c68fe33a021b6367802edea31d9059b79f295b80be9bda4b195de80c4d10d928fbef8a82c4855bf9932be1501bfcf83daf86654fa2379e534bacf22c915fa8f40b9b6cc2fbd576aa92bd6215b1e43c1b45fab01f1d096904124f469e55a4ec3b42444e040107730a55ec71846b728b9ef16d2f9c5f1e5421f0b27f6d8db3be02b0c1d794122297fa147b00769516172fdde79bf5a70f41b98c48fd73494828a5d404cc252696ab70807080aaf9daa5030ebba782aeab4cbb192a6cc464de71f8f6da53de65315aef43eaede70e4e3bd86bdb36e014596e946b67e479929e8877c19138dfe3241f9a73c3c7d7dc4a439d8ea6b7a31bed1c2bc6c310842fa72b7440268e444a2a01879e7e3b353d285975b186b3b375dd1c01ff8920b831c4edf8a4e74e00164477f5e36504f3292de590cb499aa25861c8d37b10ec8942c60dfdd0a2bbbdbcd733ea18b9b33abf3159d7a439ab070264b39eaa4bdf78ec144d80e26d50a225109172458b20bbab94912ccbe4f8123c3e883424cb1b4fb91624f653705d70aaba371abf3157f2cb5e95ab2f2bcc99c4aa7f337cbd32e00eebb9150d906fe7a7a6d501d8f3ca9c0b8d3b6288abc76e85bd5dbc12b89cf024ca6f7d8041dece2601fbb3ad911fbfa2690953928ab80967990c7c8710271eb4f2ac58fb2d15455d2826634aa5473508ad362fe0d322b4d459c1e01c60487ff596d7205e8c1a0cfdcfa2cff4f8adf850011bf8182fca4a26610446d33ce311da4c608fff89737617c0bf696d150fc001665dda6485d7075de904831698b94a8140ff3d2613f027c221998f068cfedbdaddb2f1f1365ba913e8ace3fa9414f50efa2910678763ad1275dc43cd78fefd1935cbecba4dfe1f3e0f5bbad02d977dcc84bd8d1c4677bc9798381b60161b373cb143c48cc8c00fa8491d82c7a15de9527452a1368beeb86f4b5501f0ded2f4c44af6ff0fbdaf8ce100aaf7511f821040e7c88da6c6bfc86608c7b398844cf862a06950eaafde3ddc7a33fc762d0e9d63606bca157a2eda3c0f8310f79786bc34cc227a308722bc54aee1b2ac4828908d31430a34bd08495b8fda5260ad0b424a9aa9d5ebf61ec8a83a7acbea61bd02d6d1fa12d26c0e3f8b71404821c123f1501624cc6b366997cf76651ac1418ea26e89ee2f6bad65ca3b7d737846ee708584d4de7dd3ddb71c04ee4d505227112e72f64203287f8a95c315703b5225d9421d687fdaa71b4ae98a2e77cbccee3120b1f72e70c3f84e0453b6054f2fe510b09bb22588d6a6f8d02e606bb33c4c9f5dfd3abf147738768ce6ec23363c7aec1d2aac5bf6dae182f12d12d9ff5dd047bf0911dbe95c606fd49d8282f07281db3db4d2148472aee8fc06fe3cec7ea3c4061e97a6ccd1381078e81384e32dce1fd9fc090b8ae5d203f374ab430fde2a779597edd1300a907eca6339a0baa666d4342db893cbb5fc1b9e06a93bed0afd9d301cba082a27cac434ca93fe285cfa115f7be0259c9d8a7575c887edd76848efd8f3fc6f1fe9708bb6bff5d9afb01876974f94a9df029e624a23fd5d9fd88f6e661c5bf53d09fddaa4ebb7be3248f01f712f9efe5a32e2c9a106d76c53684537a39e7df87f541a9152aa3987d4e0dcbe32971fd192f6ab0fabfd58392be7d92fa1f19aef4bacd6decaf1b7ff72303b187a3bb03ca05ac2d19a6583c02dca516ffb0743bcb38e4e8ea566f96697784ccd60e4d4fe86275a7f9dd3f94d631e09ffa7b2626f20dafa5b80943f099d0299eaf0b7526c2385a137c8b205964c848bb7624e4eaa4216c48a733ec9bd33deb12f4be09eece0edba1d6fdb1e2101da2887251be6464a9275c48196ce6cd46147f60fe0861ae2c9f5f8a0af0e9064556b9774bfd7aa4b06cc54a1755ba8e57eff0514a516669fa70c3ce065a1a885f7e8b903c892aeee8c0ad7bdc9fe4fbe3ce3d1b900b631e12318c672ef667398a3f602f64329a1c0c631124512572f8e149c4697c9b34a53b5cdb260d91f4459267f7f1cbc753f4d306c653b1960c0800a685eb476d008c32963e2d3b94f08419941599b708dc3552fd568a4bfe5bf2243a3a38d53872e53ded92818c3eec52385a97d7f581355c28e755ed353d9486d483eca9b59d7fb35d7d3ca79e10ff245e9a1f7e83b7676d35c0c2e2eb4e53bd2b76098d20eaa45752399c5a79608b90b042660385dbc8839515d8ebae25e0ea383a5663bf7207ab8e98b61cb7fc3badb88255e04978742ec2aea7ee0b433e1ce226535937dcb62901d0abfd94cc85b871fe694acf27646712015d6901a065c5eecdc1cd6cc5c6e64da97a8ee66f5eb5abbda13ee6308d059abf74d9116dd8ed092b5557ee678b64f5b3bb6eadd002e87d6f4753a8428c323d431847266a2cf0461ad06ceb1664441db92a6992876f0a91f68de6b107c0d12c10408de9b67715131b8c2792a6ee7f637e9866f8c1c3927678b5731fb4bf6827e603bd2425bc3b5f043dc048f39a65c6d8e115c2b9cf5203677e0cc263923981ca6aecc31d0723270f21c23986816c1a63d9693a275030634276ba015cb6dd3c3b23f2ca0fca02157d187550d4c42bf83d0fb62a42fc12cbd834d189e665155a3350bf60a8960cd6563260566f3aac88d52a320e65a707649e9d783d2180d7954a78dee92fa0c2a383f32754bdf67342b4d75a1ae4b18368763f138cd0296f0da9f193facc3da4ca2b1da04647dfbaa5f9497fc9fb741d43c3a2a155680b3cd0bbfdd3c3685e5ad76ec091a50fab87a300121ea140f28b6fd2aa0df616044c05cb350609778353d8ce111df21525d302b8786f8350f00255744bf05a2dc3fae375a3ea8b6abaa5db3ae2310f545afac4030c104e196b4fa257f509879ce76e6b8d2b92abcfc1d0f88852c67ac3f0c4cfbff8a580b4a121b83ca9d8ca16bbd46a70ed2567951a756a9e7a4bf37441f3cacebcb031b761035dac540ef78530ec124e2a44f4bfc8d095b882e96f099ac29396dbeecf60ccf4a6f5d4b517bddd6f6c4bb816816123d37e6e2bbe62e279d31131a5f13f75c050050a8678031c418b881b9f8680fbcfe9bdbed9ff86947bb808c4c6a990367af5271f0c6718ba76a1684ce755c208175744971cda7d0b910d0d7bb8919f4820cf5940b890b90976be89006a14bf0706e15830a016d3bcfc52f67f591c76672d2e107a0cf74d801bbf92779189dd5b3ef92daacb3958e591b9ad9fe451ba8e533814238e8a739c86973edc4a327b7b09b26973c3e11fece0042895f6be1a1ad953c7856f03cd3a941cd4104ce15cf8c9b53c44643b7fb9961f5303d6e6e72a2790ee1f129affb3dc69d9cd8ce1d6eef2c0079a6d9d0192ebf7c4f364ecfdc411cb3713c5444d01431b8cfd839f0bd783f304c7b80e3c992af7d246b4f23dd1918715f4292d1a18c0b60e352470000978d6cd1e80d64dc7473ac897d1cac97c0d9bc1f28f63e5eba9d6ec588427512742514b5a8062d75366fe77937ff5ba6a0f7bf5e20f5afcfee327588559a2c13f0bc4534c680c4eb07db972605a84f8708a6e5898442ce7fd2b0afce81353e9b30bf75a2e1b533e2c3ac0b8dbf75a85517982bccf4b20a263216174aa72a7b919e82ba6d5b5339a1f23299750fbec736e0678841480b9f72439e5d05e9bff369d6db221d8a26119f30d97d54c80d0324cfa1f298734be6843b37a95f7ed6150a2d003cbd05021ee59ada06ce6dedafa413d9417fca3f539e9005684f0860fbcbf9409548ada3f083ec513db85a30353adf83b30ada08324f76d9567c06b1b875746fc2590401ca15c0c8a84cbe65627610cd582222d785c7db45a89f3dbc689914da25443dcfe6d7fad3d6dc74653871d1be8461f0f9820b4c3d34a2fa90056de3d8dea21a0723c75d1f559fd769918829011cfd9c46c29903b05d411e7db7fc859482c3298cac162b533cec391a38a63a833176695340db5cea30da030b9629ca9be19d9a9c92c33712edc2d3c11f0dae9416b5e666f3d10ddac2a0164baf3794f5fa9242b5375f4772f65297a126663cc7e2d2ed0f0e234b3a00e66d65d2b704627c9f4c18da4f9fc60197948a7f0a4dae427595bc4bd75f1ec7db718595504f3a708dbdec3e11aee29eb7aaea12a96125c8fc38ea8799d5a9a242866fffb1d62347f94e9203cba17c1983da65246afe8341b372b40cf7ec5e55810adbc4519221470f0bf2e0a4775794eede0d98a9a5d84b619882a00b09b3a3bf067522b0a1120b0ff2d4b324bab9619fe21e7d1e8655382134fbb8bfc2be02b7e6d76a391585e2f50b0113471cb049da6dbb0578d641b3bbc957db56b38a18c52eae3caccb4f38d98c96c8774b345368c95eb2372189ea02933a5c7105bc9b336d0fd51fb9f3c8dd46bb51fbac9500d85865d2a7ee8a8f8ee511250a1fe7318787d63c5725b26ac49820f1d1cf56c7b06c15bafa5f203f7151563f1c65b710a8b2102c1d8cf80badc8df5e3e09f883c78a38cc85c7d77bc8895d6e81affbe87877ae43e984934343124a79b873430e6a4f8984cf53d8bbcb4535163ef34acb8859ff7f416b3b3046c58199a7d1e32cb37eba8fe82ae3a1013e5f9180b9ef969d5fd9c981fa4b10e90589168c5c376bfdef3bbff234542cd51825243a0e012568419a93f18e48be4c230f1189787b0ab11d595d8b067db805cc80c8fe8cb54f7796c6542a42fb85f17bb02fd63a8a569d2a67201e6b5fb139d30ea204b47fe74df8cecca610003c0419c6f1b9805b5c72ad0e980e10d6f15524094455bc09e62f560062684c714da90b8127506d5cc37aa1b07c76da6738d427d8fc4ebbf020cacbf24f859640cdca0df429662f4389f479c11ddaa193bdf214bca06c18dc98aa133970618bc269ce8dc0bca5f4a64e70299e0480a167762052206db4f274396bd022409921d06807d41f74b60504d26012d909ccffbda0d5f30158f8b6a9be9e49141acf22ab716efd7c7130634897317762732ab7b33749c39d22ec35fe79e26deefd86158c5d7b671be0c5b617950fdd76059274c00c20f02006fde7de0c158ce1aaffac4d7cb408bd35404611a42d753f6ff9f29e6d2e842bbb3989de396d025d17c120666fa41e96235cce160fc95c75ed430b5821610fd276c4779303335a0737ad8403cbfc5637405105f5e443ea5b7938aa002d1700e6e7daea7611f218de6d47b61340b57ae05047830f68ea298e39610f805360f46a652f22ca86f6f44ee0d7de7a7104bd66a85d0129a4e11ff5e16eec0b414edfa2dacc8d363dd64665e53a331c3c0e4153f23d1bba3495819b2c9fc95141c96e032bda7d25ad84703ef686b50b6b69b5514f62e14a7148d71599d296ae740abc5bfe212c4bc11eefcb5bf7cdfb0659944d55a63f3468b96be2be415ed15792a08d5ae410819fff0f5cc770cdb516c1b299c7d8d104918ac00cd9c5e35d5ebb139a3def2fecaef3f74eb16a45137772a79aa81760f5dc1898b24f5c255d4b77d932d0316df38e5163f0676b9bb9945470115bff4f8244208ca7823e1b3c09988d2f35dd16557d7b0a3e85cc343ff198c730bd125541034c6938fcd897675d13b41cf504b76a5d61b8e62449713b846d27f623f58a7f3d23f59c5cf1aecfcf950df1cc5466e748fae1725e2dd039a44cdff703382950a0ef86c2dccfa16be053e74d92be3d3c850097c110b2e0871b5fbd87795b2df71a27558e19981a730d702655bd683f89f48472577715a5006cf1c", 0x1000}], 0x4, 0x0, 0x0, 0x40000}}], 0x2, 0x4000) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:47 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) (async) sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r1 = socket$unix(0x1, 0x5, 0x0) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = getpid() r4 = getpid() r5 = getuid() r6 = getuid() socketpair$unix(0x1, 0x2, 0x0, &(0x7f0000000000)) (async) socketpair$unix(0x1, 0x2, 0x0, &(0x7f0000000000)={0xffffffffffffffff}) sendmsg$unix(r7, &(0x7f0000007540)={0x0, 0x0, 0x0, 0x0, &(0x7f0000007500)=[@cred={{0x1c}}], 0x20}, 0x0) (async) sendmsg$unix(r7, &(0x7f0000007540)={0x0, 0x0, 0x0, 0x0, &(0x7f0000007500)=[@cred={{0x1c}}], 0x20}, 0x0) socketpair$unix(0x1, 0x5, 0x0, &(0x7f0000000000)) (async) socketpair$unix(0x1, 0x5, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) r9 = getpid() sendmmsg$unix(r8, &(0x7f0000002ac0)=[{{&(0x7f0000000000)=@abs, 0x6e, 0x0}}, {{0x0, 0x0, 0x0, 0x0, &(0x7f00000005c0)=[@cred={{0x1c, 0x1, 0x2, {r9, 0x0, 0xee01}}}, @rights={{0x14, 0x1, 0x1, [0xffffffffffffffff]}}], 0x38}}], 0x2, 0x0) r10 = syz_open_dev$tty1(0xc, 0x4, 0x1) syz_open_dev$video4linux(&(0x7f0000000000), 0x0, 0x0) (async) r11 = syz_open_dev$video4linux(&(0x7f0000000000), 0x0, 0x0) ioctl$VIDIOC_EXPBUF(r11, 0xc0405610, &(0x7f0000000180)) (async) ioctl$VIDIOC_EXPBUF(r11, 0xc0405610, &(0x7f0000000180)) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) r15 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) getuid() (async) r16 = getuid() r17 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r19 = geteuid() r20 = getpid() sendmmsg$unix(r13, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r14, r12, 0xffffffffffffffff, r12, r15]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r16, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r13, r17]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r18, r19}}}, @cred={{0x1c, 0x1, 0x2, {r20, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) (async) sendmmsg$unix(r13, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r14, r12, 0xffffffffffffffff, r12, r15]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r16, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r13, r17]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r18, r19}}}, @cred={{0x1c, 0x1, 0x2, {r20, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) fspick(0xffffffffffffff9c, &(0x7f0000009740)='./file0\x00', 0x1) (async) r21 = fspick(0xffffffffffffff9c, &(0x7f0000009740)='./file0\x00', 0x1) r22 = openat$dlm_plock(0xffffffffffffff9c, &(0x7f0000009780), 0x80800, 0x0) r23 = openat$smackfs_ipv6host(0xffffffffffffff9c, &(0x7f0000000000), 0x2, 0x0) write$smackfs_ipv6host(r23, 0x0, 0x9b) sendmmsg$unix(r1, &(0x7f0000009900)=[{{&(0x7f0000000000)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000005c0)=[{&(0x7f00000008c0)="fa37d126110b0896ca8e4cbaba406067c131b2e9e0ad5ac2bd8378a4523865d9919d2ac94f8deb2be1b52a56390465941b590cbb1a660d1b72cb56d5f1325779c64f435e8a6563595e59d518ef0ce94628446a8a823ad6db0614779a870a70108fb1599c1e47a9825efd5ad369c7e9ec5af43dbbe5ce40926e4072310392636ae16a8fe85e855bf04f4d5f9d555cd15a0168bbf4950337bcd29104410c46f2e02848da30753457c0794810bace066af24f791080321d332770e5d46a94c4854955ec5e331c714d67073c78120f18d1fc97a974a95c2cbc1d043abf8a549fff259d72628b2102d91c76e0005e970cab93f59fc8431160e38d8fa1fe373475fce4cb77bf907be2922e7403b81b4ebff75e76ae35d5a5e284b76e8e1f591504fce19e49398f7c6540b1c57f3df9ad6000085038e5ad8b0112de53ce1d970f35e043a7e1e0fd6defb9b2bbec77a9020dbf31ea9da2e1d6c6f431372e934c8c725eb6cea4121aa6dab87700983da7bcaf60bd0f5b0021d86552bd07360e00e09592dd373fbff45c1af62e2c0e7edfc84aff2d44fe12d0690964c57db055558e584345bdaf2b716e4bed0d329e572d1dfe07507123f3cf9121d4370fce9b57c2b36ee9ab468f6cf534f9ba06ad90b56d5b7550e0779e03bb86cd4899a03e34fe780c7b7cca7a12b3c2a32d4dbc8d87c2c1cd00e51f477674a31da27787f78771ebcf3dfbf87a0f1c28cd01e687302b876bfbf3ca8d4f228f0706c7be37395554ca5e9851d698627c98e52caa6d3168025036b6008fafd4d40d8fafe632c1d33e3baaeb34792ebcbcaadde90730305bee505fc3ddd7ce8d349ff7baae47081283bf72b968bcc3b3ecdf5dbfc084e94d8cecd920e8818df3c7019d75fb783d8fdf2288f15a392450afc4903472318c4805452aae3643fada5b693568f983e46f1120c3c67c885d9d36cc63c31f52c94462779e6b8a2839aeab4427f322aa38de378e0f9702cac011173aa433356aa2041eaba0c8dee901743a35c668fa79de9447d0c556bb6f60fd8916771c96f83e9390cfd30de438be1015eb7e76be51fe9aceb4a2f649bf50bcabcc6397eb8841c68ddcdd1eb70c79ad2c4f4f3ee6baf2c4ef0db550df29c16c94dd6f0d043749496e25c7f641c580f48f38b6e96669f886080176d329631ad0d7ad2a068fe79158dfe9ce9f8bfdbc3f1d9f376d8b056765d0ea352a3b9d684a6fc6ce6cbd666356d27b1ee8689e9168967196cb4ddfddff5289ce786921100a5808b262f53972ebf0e02ad9293fe6b051c2a2e40390c89c7de689d190174168c9141cf155702de67dd151ae631f6a34642bfe9af1b29f56b7c02e6d109d42e08964bc288b5f51f58d984b1ef303ec98776af8ab69446a839ab189d8079edd5e8e3063f3767958a7de47fe0a55ce40c906eedcf5cdeb4ecdecd41590bc10a92c5c03f8b114157f1b98d60513193c8fe1de8004736586df037afbc6896546786ee1260955eb7770601f645e95c87e2c7cf2612608ae426b47c27aa143c0bcd833b7326603a5f4f1ba48d5025bf370d98f74f550a18d876e0343af27ca98060cd867eaa11df5bb30efd1ccdeffb88abff851768e8da4f7bcb721311370bf00000e017b9e41397f47df508840cf0718f6e518e8b79895b70cb1269acc5f5258fca35a2facaa1533ebb7ecdda967a95ea57608c1cfca395d83b0a6ef8aba6a27d25713c1eabe1fdb002c4391e116ee9e900ae126dab7a1c9ef3117168aa55818539b5d089e6bdcf68c70c9c460c01c027838c7c4995c765ae969c7a20d7b92b36b779669b1fb479730b0458b40a0b9f6bf557183f61b5b4df4b7a9a76e07ddd94ea61c82b33add088bfdaf1bc3a47266bbb3a4a88e04cec90bb731e5bb1905287bb5936a7ebc00edd6fb179587867c7fccee1f35a92455a38ec06feacb4c1b432d855cd78c3b29ce4dfcdf21cddfd712b43928b72987d8e88020fd861cbe2879f653d6c08492c4ed412adde82755363ff58f46b76405ab32064fe1c75c693222e9ce9342ee4680270fec9bf53a585d0f971943919d37b07d98bca22208e40411a8153dd62e8c319cb3552bc5d7f258fc9068c8aa8d389efcda53b512094a38ffa8e9129389cff014b4df1242efb4e53fc8a55d00eda1fba5920a565296f0d75d6a414de3b5efcda341663b1950709afde7cfa29504a3aee7ede77173f8503dfbd648b92dab827aad42814ff5c6610ac972c510142d1d7e488b25af38f3cb4250c449e4089c057d2671eaff30ef11e746f282a46f5f4ca74fbac3cf2d26449626ac29f118e7bd3ad4bf279674fde588ab250c06604af2c109d2c8f7a5aa33254f80965f310c15fe48499f845b1ce53c55f5e9b5fcc833d5303503a47e2ce8110e72acc2acbb0af99dec4e25e4234fda0bb25dc3e3e47ccdee2a539bc2c0d25e69e8ff04e4af93f3e16c6a888022c06fc30afa8bcb64d284ed54d4873f3596f3590a91e467ecc6b7e5dc2bdcf5e111632fe4c1d1c92c1e86fdc5da79dea1420ce1e322d50aaa02d5e75c7fe7ffb93f4f7cd372031caeee4d5054e755c5d4bb986d54fa918c54247eaf143078528ae6eecd05b86e94e54c4b69de5a98da9a1f795e64d0cdaf800a130ab854ea0ff80fcdba888c0d7e65125f6690388c0cab82c453bd88ec556ba6e4d700ff4e10cddf3b557b1dace833420b2e8b69ce047cdabfd752428eba9b28fa4d0781e3bb4a1c99e003253eee35dbdf0a007a854074210db70b885334b9c920ac2c4c982c05a967defd92e26bd110ae885d5c38f1303e9308910e354f1e0e4b1bea0a7469a197e413a310b72e6c877b740169f0a53e4b396516e0ba105e28c0c51731bdaf946bd5e8e95148486c97f9be35bdf2385f13334dfd7804c010ac4696e960d0b436fe51a4fb27b472415b7f6ea4fea1f0e4bd51f891114931a9e08b00a680ad2db82db728d61da51ec5f2205903c942829dc3e697b7174bf863d6600049f19c888e805e4103c2d0ba517cafa6140ea78000cf6c38fcd670c36d32b6a1cccb52dd7d467fab225f1df1f81a8ed1bf717346600adb129e354b586baeef4eb0db5e956687a4329c6357bcd719708672b6b9d6868c6c5e2ae72ddaf3b30250f4ffab74515c68ff3aaff8b2806119da6ffb79be709bcf1407eeaac9290ce56daff4d136d6b8eab914976f8041e0171de4e2b01e9cc7e765a90341c3436ad0c41486c40c8a7aa02f4ecbdd83433358f5c8bcb0ea97a67a4895acfe04536535e60d5aa21588391fe6abc15be83dde9b70ea85229086956839a4ff0b6405353cecd82d7ec4b153815cf21ddcad042dcddfd52dbfe2c3ac86c2b9ba8491e57dd8fd84dfbf249300f2cb5889e17816ead325dac3cf3dca9a614264d8cc7bbd26630c25fc0e61bf4f4c40aa00f31fb1ba830a892291269815b3d3a6b3ec721577caa20237c3fd2d4ed8e0ca0dc930a4b718cea849b5f9b4dfe509026a12d7c7eb2429da5dc5b2e65347f5eb386593ed53199389800b56f716af258827046459653919d9bd0b469eabf14a31b3d7dd0b343b71d77ef84d1096903cfba2510aa83c2d1e9fc79dfa3b36a8ea8be8f1779f2275d03f99c6392e9832abb8a100974c63df854160e8b593ff8477da7718cf1401ee9242454bd2541b41b0f4e0107af788f9867b5feec4276ba80b495b2d78ec81a93e55c8d3bd0febb9973dac2d3ad6c4f2e51093dd544d999570130203f029de3b0dbb77171f37c74ef0435d4af2245d81bd86ab77063c1b3e5d69194b0ed63c95080ac7ecf31e1147de51f3329cff3dbb2a8b364f4a53503d89f04bd2b16436d0cd6c45b961c113e68ff816052d0705b761bf45d1447a1320e9cbd22564537691383882ddc681538c43cb5bfbd3c5a3f4e65fdb5e899d76c1d285a518fd7fe737f88a4d046c5179be907f7d0c428bfb83ec4c36135ed5b5c92d9f525f292419c2a0d5599e2486a795d2d2b8c3e52aa17269cfea0e189596976d5bb15380f8653e4f80bb1f595162138c65844041dcfa6f9d86137c228b3ed455486461525a0ea6dd682b7e8b8c0fd1510e8b1e71e3bc45bdc41ad55d37af1a07631a30716e2fb9bba2c17d74fac8e22a1c110656565ec448685d8c9f5856adf9b19e4687d0c0cd6b42a83bfe9ca455e25c757d504413286d35898b9d47a06d3afb2ea136cacc1c91584b1813f6da12f64374dcb72e3d3355ac2c6a4d62b3de3f19da8e1c17da2b92248f839bdfdb5585c2984de38ab782b73447e39d15ce933644be7aad8e49ae7d08f002f79e504a688672182d74498a39304e4a31bb295a889d49aeb0f9a3c92b35022fc73ebbbecb97b129261448429dfd9b0b8478ce9d800b374cb0384434fbeadb41052e45bc1f946015ace6317e180c8c02bd0542ac45403c624c76176b461b7f618cb91b7acff8996e098e882d45a50625b9c134f95c3a0003be93588c85578b5403f66a5bd3a963b27cd44eb4fb1c8a7610747f3fbda234e11b5744f1068eb42fe0bb9a07a1d9c2183d0fca4648cb316d636944b09f3d8aed69d1ee517e3608469a880038aae7296977fabdb22d24359f66ad8b4a25f9476505956a4beecdc44ba114b0a3219c63ccb3799997b13d9b96bfb71e38bf9e0da7242020bf8d7e213a02ae5dace52a765b79ea497339b3583a5647dcda3ab358247d634516f67feaaf0c44a1b9529ea376b5b6b00589700647a9db5fa5899ebd157c9da5807a1fdb618dd7041268dc060b11542d2a93d74864df7cf01741c5ac5d94941d2a096510f95336c3d22cc24c7fee47f903b87c2fde37428248192257c293235129029e442aac615707cb15fde52d5f05c15b7065a3e9998ed7b784fce4aae2129d34b63be4c773870ba542e276bd15c6e2771a0ad7ba1f90ef4f053b14734bb6a7ae7449bcfa723b96bf2a5cfbd02517ab16990b667e4aa3089bc5179ad7a6b47854eea2a99a2e8fb66737ff9dbb645ab2b34c5963a3f8c1a955c0e2ea9c734705da5a799a32640c0095341b5eeb37488df41609bf82a21c97a813cbac254607dacb85cc6b31305ad31eaf900f53ba979bc323e93711bf23e0d32d2dbf8e8fe1432919f3a2268531e46722eca0a033205db5c10378765b6b3a562757b3995b03ab3384d93f0230c6589db4ded9d2de1e575a57b042b47d904b707570fc7891112538bbacf61cf2004ee0cb4b7ed7db65ee269ea9c8a27afed7120361ad1e729401ab37b26ad22f56ad6c179aa9cbdc612baac1b8f8e90a1fcc4be1a42fecd5800cfef77c9790d69960e8e97444a3fc79fb4db599027e2e6c97d0bf2d1a8f28b29d03a1ed0e5ce9a676b1ff77946f255ab094df65b31e196e4c589b181bf4ff556041687b3a1f3e00638317d91ffa66ba90d00bc8dc047714ab3e2022852fda88c7c3fd5ca44782dac67fd17d43a7e3f934ff00243318bd7b80cc651496a5bd058568876e8b4bc163447c36fae13927279380c5afb35944bef08275ca10c0c0f3addbd7fc0527d4085065cc83d09ee2ce3d5f8cbf3087df0a010bf6d85b3d17f515d64406bc875537e3ea13eb26b329979b953f0256d27b978aec9de0c81b7d5ce00e02ff36a0a68c67b3180f73024c9c28d9f39b91e7c247977318742f3ee8076a90babb0c2facdd4c8f97609276e6d939892dd5b3d9131eb28240ac0868f5257c9991df4663d53638accea55b7a95ea0e0e6845d7a8eac15dfbaee4227682c28f66d0c32c26a524138f3b11d630bbe2cd7b4b40365aaf2e5069be2c4d6bb00e680419c8c24e3f0e", 0x1000}, {&(0x7f00000018c0)="1cf3ab47869f60ce64dbeeeb8b768c966c54e95c5c4d23a2aabf62a874e2b7cb02d0faff70b9d1c2eb12840d6bb9055e219654802c02edf95fe117664554259ed3b647094ee8c3e5967ea7446cb0643179a917a5bd99c22bf7f061e2b228b19c062b15d4b3b838c07c187caa350a1118d37ab5946cc0b489e128dc0360ffacd83e81fc3bda12a2117b76920be1e38f0acca8ce05f7d0acb3339ec1ad1be54b51f2460d5d3f62c140fd6f89c0f74dc10094d924880b12f43524f2a6d961fe7dcebe83e3cee44c1a435a81c062ad483c0d7bc4cd0e969545c0d31706fe982d749557afffd8ad6c511fc9ad37467270f5336afeacb1c7e0c4f7e3d10cf30f1639de03d049e4ac5c1e9e1b46f1bbd6df9204b6ab1c5de489404dddc1c494623c76032adc7a430cf02d8855c8d8d0f5cc7ed0e7f7cf4f60b1214dfd0a508470eed0de7e1cfd0e11fc0db5d563592b12847447230390d0d111e52089de01810beb82e174208c10dabd9ae75f10aedc627e8ec532abe91a5a1c40a74cc7dadfefb1400bf90da1dd28ec86633c3b6d9af3e27477e9ff42efe8b5e99f17a34b9a5c47af12db6a8b9c42565c6be3bb7dccae26db24ac305e037216a6c8fc84c4c0a4e261ad264702bb23a151f68cf66d3e634e9c5e96da0040aef24f197afb284dc960614b807e106a366f89e152bada4d960c45e9c4969806423d270c2bf8591492010755c6e1e585ea510d9aabbce42aa69257e5e66bc56f6046eefde47f1b27f383ad385f3040066cf2f1862485ab85f25dea5b26a15e3dd273c1f1b57b6efde3840890a717b2b21ddcfd618457b97ed4daacfc2beacd0c771516d57824a367fee820913ace4a167031e04afc78c98cf89782ef365d7ce85efd5f45a6c1b922bf1a4a91952cab59d32cde2bb9d0d54598f41062a24e8ecedf16e6e64553a1b3ef1881482d300990d8ac9b7f1ebee7b74bc17828bcf31048360482545c449d2ff55fc2e14fc20412bbabd723e1782efacab5fa3527dcdc6e7cd82b331f68f9ce3d1026dbe09315da525104ca900c060a30596b2ef34f83073be322bd7f1468f532e25e3830d2f3d768047e247bb5b9516254029b2fea69f197df86318c5196462924601669ca39a9bc7b89487fa5e2c8702f9128fc2d3fd79cb1b3c59d2b8444b266f2420a25790caa23928a9eecc2e06dd8b8728f53c97a92e1d4e132904bc9950eb8afcd75ce796995b3321c69e0794691ecd63c572c2f9fb07fbe0bc00d2bc96e9badcbb0a189d98c86b40d7e91aeef3eeb3188657c0c0bf57adc350a54453469d28f7e9f5d9fcbedccbc4a9f7eb3376460a44bbe01daa2bdcd245d1f40f2a3cea2ec494904ea06d1ece1650b2100ae8325f58d39dc5f2efb5223eabb581eab0bc0f342c9ba045b033b676e8d02ecb0b2880fda2283fe903bf01b6ddba76351e06c47997eecc65d974f0678fce0171b0e060b688fd565ee0fcd7aa940755d33c40228e232a58d92e9d51227fa8ffa94c582f7f0690ae89c0b940444a8f2ac29d27a12a60c9875472693058077781a51f84fb7d8a32cbc4ad130be83f80154e2742e7e5474d9b4cb9f7b7f9d17bbf15cecf49e1b92e52fbf579398b13790d3fdf3ca4c077d5d98229137a343b745ff89b3c9c14f2aed73fc9a7e3b307b7b7433f7cf93493b1899e80956fe33ef88291fe066fe794bacd5b5e1c344e0ed067d9c57f59b7396d4710b8207c99d27ac388a26bdd94ac63162e7651692a21e2ee01af814bb213992b08b0c260cf3c001f112bc6d6242d097991a41e981bb1040f20532fab540e3b25bab2bbb0eda5592f224c909283ba9c722cbdfebab4a162cd676def056cd6da19111c532aef3d73a3f132ef92a05bfa02265bd1cf57b20c8cd0bbba1628a06445c39a0d624104ab2230595006e612a14d3596ae3121ae86cabfcaf31808acb7b95c0c8f6ac733c4963a9e4105d69bd65709352d107d1345de415bc025d68543ba1aa49a0133d940000038e1cc67a4654fa4ad557d4cec7fbf8657cf6d4e7c93697978f251d28f25be94deb695ee612b5aa7abdea898e7eb91d38d890337348c4156a6913dfadc1674d0f882e32f2a0c574228a401289aa5375a625d38989eb4f0bef6d1a231033cd1d1f062a9d219d459ea505e95c164805355ae54beb029f706387e6fd34213ac4785665d6467442bd54197850182e1b4836c1c7b652713f94cdf3b7c051b76b16da599cd37d8fa9cccf8d64e91d6a4cf5e64ad3223cf3bfc3fc25a1258b827d0d798093ca7b850e04d4982abc44ffb8ae030471dd99ec4df0726aa8705c0ffbff88ba03ba990f5f3169a48d439fece89a7d27ab0df28f0f4bf85a6965f902612ef1aa98b4f2c7723cd8ad3f4de4bb30402ad37da1c3d312bc9b20dddbd74c47e24df1b533cf7713858069c84da81529473e5d7b6abdd3bfc7814af63c88a04a70150a2c77abc55977846dfe01c5173e946541d9d535af5038e706592e84cacfa9dc2970e723b3775f4ca66da94a9f8ca80014ee9974d839362ea7fc031c4c9e722516c0e4c536af7425444a2ade2227b760cf4b014adc48b3c9d72042ed40da53d25c7ea00018a3a27c320a49636fe54c30d36e58f5b6225957dcb370dcdd6b3d08a3b836640c8a59de3e0e23e09cc37012ab7a490b5e3769b0005339979992abd2cc0d9a1843b370bb525654ba723bd17667aa598806b0f0432fd73c36772728adac21d512b23031b1b6765ee019ca2f3aac4bfb3ff80bcf767983635e2f6d5345c0d3eb02eee8e5150503f0f5c0b8f3defd2235bd9bf2a845496074c1e55f8cf1226dd9d37b39305ad41cfe0dbcfe2f85d20c8adf0001fa22cd0e113e59183f97f45ea0507b034a83bd5d2efd4cc5c80f582044ceece90ef7dbcb8ad13dfb76f6ec7e4d80878de4b1eb9185929b05bcaafea6f3580dfbc5df8e582ad89f8975c6e9fd7df4eaa2bdd6541833cc979350faf830072aeb67ae91ea4c2d5a963d026b37fe0b7a93fcc0ff5dc5108725e3a197eeefa338037deee0c43c43441a603bc67881c1bd4d1c8b8229506c67b2e1c86c9a825f0b57d781647313a70b2688162b133b27e931d37f7da718f4a7836d0d7381ce5dc9d1752d5652db02e15e2743afcd77c7804dfde067a135c76316dd7057b79d6dd850df5a50bd7660fff1bf41ade2b7bb441dcd5c4dd38cff60aa7fddb49c2f6069e2c7328ac43915006febefda1e86ff51523dd9a8611af2cc6e9942121595217362389f6a6199347a452e13b9c48cefbbe2716a78bf2973a0fb8439d372573c20404e0f2ea8281a26dc0c1bc6d67552b9b1857f32dd8f1e421634efb6a0bf332d44238ac330839dfd9c4070796ae053bba14ab354ff9397cde73f3e175ad072e4b3ca2a68fb60e25c269709b758d2b4dae878a2322bef1a656cd2c10b127acc9d0601e663a36b3dac4205b424eb3a8377affa53b36ebc7617f172b9ea9a6b30de0b8b2a5b3118f5d213c076b12bb7336a542a546b6576bc9fa5ce3aee2a168226b7b6aee0bbdacc61f5c1b03d6551b5b08cc09baca71130d23bc336a9573b3fc2f50e108c2410c37788d32d89d57242129a1378e801231a2fca9e259c2619c6f85d22b3532cc5782239c2df02e50198105b484da56e518ff533380d10dbfa4a8aae8bb7b3d833eb590088755c414b2ab85970a9e34aa636d7240b91edd3605552bbaba74d41efce775a55cce314400fae308ea647205e80e9c3644fc999afaf3fa5b0418f1f8a65d4c85d352a95a750f972977b16fa81f37e61e241075d4c2bc2d5ad57318b55c8baefe198aa088f6eb63fc3d73a7bd737d43f9200ad55f3656fdf5d8c3d009c170d54f3def0e9a6e5c3dacc8b8720429f1bb5b45864a0cff037e7420b18a7a8bad6ffe095b270ebd3d0424339d73027417e42be2e4ac0c45d0d880d240ebf4f1d575b52ee49e47f247195748cb8a518b746e49647ae58323c1388c76acbc4040578560d4df314ed9913f63d895488f3288f00aa1a96a053d9e8efb5990b1cce5180b475e6efc8973f68580bd88b40d89495d8dd3c128ba2b50c7f8bfb6d1c5fa6a5c214e6af74d2b2a78f9454096734bb3d119d68e17836b371e7d0f908e055d9c518d71e1455896b91b10f2801b9932b60580231a781d1c8904881c6233e14a8cf8a7ca9187cb005dc4755e035c1a998deefde9839693cbcb2d9c98707bd18d5d315af0b6b8da8ae06b993b28bd19769ff4c920a39c05b8a998e053486e8029ec6b2be6bb03755a9d8af8a743608c20e26a8c0eba72d3c88d9d128a6dc426dd38484f7e31985433c0c08dcc193cb35660eca143c88b5155e30fdc0d0bed09aebcb065a4772658b9c903edc33cf4180a3aa4790fc7709971350caa41d452e55551e50b0e87fe8550365a98a62401d4da18c43eb1e3e2b085a34766897fc9b034f1c6806671185ceb55020f774fd90378f7143e8b10cc93aca0fca42e880cee056f5fff19f4de211470fba3476675a704b0cfe26e7fa0ea0141a1c68c30034a4ec2be8ce5e72f3ac892101f16411ca74ebcddb1c6a7c79d102c088efa221259021f234bbc1ad7978ebed418011a7663e0299386578a51a3ce853b6fb23d116366c325c82bda0e11055fc31091c2e08db8000a9f4c6c9fd62177f8507bb008cd07dfd08416bf4369bf2c89da96321b47a42131685a4380ceab21b699419c22fe243bb02a904252f9bd938fc8847f10070b6a07f5fbfe74255c383d249f5da78fb57a38adc436c63dd675694037b4e6c70d497dc7efba84345638b73ce9821903ba4fcde323403bd5079efa902213c15534934b2afb4c044c1e13e188ac3380195afe57f91dbc5f9624ec04d54977073f77056a01a8740a1bcf936c553409236acf0ff81e415af914e086c25f4f303d9c839c50924e1ea15a9b909adfc5e8171711aaee2c1ccde4f2c799155dcf80a55e0cd328f9c9d3ae42dead2be3a89f2fd3f031a70567ace2e85ce9963dffddc654197c4348e2a3f72a9a19d7df72d5a42b1ab0b7d520d427a023e5ed4653d7da924af6cdba07ad530f34ecc9224b1203a894e58c322f032e1baf16caf1c20eb5999e2904e0de2c581b934673e300f30f39f36685feb0ff73ceb2e720e477ffe9d86220063768370a9ebb7397f3ae0bebc3ff4588fbf9881ec80516b6834616357f7496e111bbbe9181b8077af79c0f1d7001e508ffa221a4ce1511f1a4b5a04b9a540f7767b09e55c0db318646b99c6c6c915c00788997095a421a3c028945ddfd83834eaa02241b8eb827fd4bdf18f0728e4a2cb703666b8caf9d90121f9efb800e35d9693ce9519193b44d43d5cc6ebabd6cd242fea7f4b2b3da50635a2e462484c01405851886f439fea0227e03180526d7f9392001eebfccaa28fe209f8cf2e7e8727e7cf87147cc3c64d7687620dca17e4289c4ef44e1be18f4da3cc185e3c2286be3bc70a6ec90ea868d2ee3622c6cc3e120badc3d6d7b666715a5e180c749fdf5d25b50c59546ab37ac9470bcf1a832a42cf6bba0c9cb454e3e25c9438ff520bdd34f8a46c9e9bbb0082f691469ab24c171541ea85f2547cfff76ac1628da19eaf4ee1646f0f6424681026d3ff1148cbd3b9e86cc2c464471ce0b905379a60f23f0231ddb5f93095715849249883d517661fa3df9131db2b82f5fa06e3505c97527f317b289a081164c45afb8568e662ef7259c7f767d989986caed0bd5b0277e7ac6f8b3b446102f847c7d9905348cee65608e2d778afc5dff62bb9f9ba1776a4757c16c7a40578", 0x1000}, {&(0x7f0000000080)="7a69e9dc6049f3c61eb38e0e19a34580dc43dafda681c5d6322138ed9fc9d696ca57625793dd380af6903c63728765ae5378b98a00cc3d4a1f6c7774345e9fb5a1c9b77f1c326b682f0f98b23f19b747f0307e875fcbbda49bae5a6846789503971e8657", 0x64}, {&(0x7f0000000100)="66d95f5d0b4ad7206d502b282a0159f9dac3a4b95992ad0f285e6099f6008abf461e509d82736be80ec45a2212ac41572a915879fde382fd9e75452cf3f0b1d7841d5d076598c9d3bca11ce2f2e6332109bbfb11a98ba543e7d20702a1eb9174997d5c7ab9235dbfc814a1c5605cdef858b6d5ab803ef431327599d21c09ff811c05305de8212a8da4a7b1c987336e0897c0ce59f78482d08632e1de5ce2e7dbdb35e1c07e079998dd1d7a4c4773f41788df37efd2e257ff0c6b8819a721d66c45215d5fddf4d46cbaf913a46497", 0xce}, {&(0x7f0000000200)="7a3a3027669a68962ea5e5391e93cd5b361c37190316679784cf473fa60402f27dacd35f00b5a11954bb4763affb19b1f2b78fe9b6ea44ad62d3410d1551f5d8d5a26e05731cc5568f16f7673d0c07bdb09a5a1fcbed0c2734d3ff7954263fcca072bb51fc3cb2f4dddede4c547bac12362d0ccde1c6de92d7749837d124aa7cf0bbf9c2651ac065033ccf2e0afc740c8ead508ea9830f044a5fa9a0513a73d9098cc19e22338393c826e9903316c7651a7a39f29c1ead376a0ffac55c0c305fd1", 0xc1}, {&(0x7f0000000300)="597a74b23d2b86e0ee5312e828f1aa8579797dd769833e76ae62185704cfead2e3cbf494628560af36e54edf804711786b4f05d0c9c78b908fce46f88166bf61d63b9c454a4c5b724aeef31198f2e9f5065cc52190e8bd5bde9310e422db21afb8f17be1016eaca8889acb20a50cbaec0246367d5c2fb8eee475d14bd32aa42fb38c800f05c52ad4eb4d5cec691106e08dd9cb19ce49a1f8347b572629280f8e7e6aa2aef0ec521cacbf8f6fdb3825fa6bb384a3245c663d17111f48c19196524676ca76b05eccae64a08a83db4dc7449a868811912c67d08804d638b04807e23bb8ab6e64e456a92a1080922aa70e34090b", 0xf2}, {&(0x7f0000000400)="ea3895d28b4058bd94781c7499bd237c9603ea7733914ebdd964fabc2d3716ce201d6d14105ccde43a9b362bbc614fdf6b242a0723fc95b37e890d2fa294083528e5bd8d2f3d04b42bc7506c9c827a699ce3959010ea4c28de4ee08c63e043f613d48fe15385b3e0f2e96d78d7e2f7036f459f42ba6f45c5e359284851cf985c4eb51f9ca022ab5894df9d430252a20676d7dac450baf3a9320899041f477922a929c2bf190ed62699396c7bf670f9bfb483c08b02d68d43604113a5e502277d382bf921aeb10d1bae9825b00ccdd85e6cbf163831c311ff517a4c3c8b242d49f16ce2005a795f2649533b6060", 0xed}, {&(0x7f0000000500)="8894a5dc8fc187d20a68d2d047dfe041645dabd06adc2004e6f90a71bcff1f26cb378c117e9e7786", 0x28}, {&(0x7f0000000540)="e3532b949814cdd50f21b0810ab69cca147d923b529d4e2769881cea7b34c844b93bc0acaf880c75647bfd0622741cab41b3355d2e1fb919d35495587fed1d636f1ad4ec8a63613fa2c170bc4b8ad1a85b01161e6f7df4c2e2f3", 0x5a}], 0x9, &(0x7f00000006c0)=[@cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {0x0, r5, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [0xffffffffffffffff, r0]}}], 0x78, 0x4}}, {{&(0x7f0000000740)=@abs={0x1, 0x0, 0x4e21}, 0x6e, &(0x7f0000003a00)=[{&(0x7f00000007c0)="352785b775b70002deb8fa0b00015e2272bad043866bda538535b59fec375bec4e4ff475aea666872ae945f99e3fbb7700923cdd07d5c2394421fce46504a01ce938e40dc65a24e530dbbfcdaa820ea80de6fc9dc287a88ce3c93e8ea76a38b171964caca6a8eade3da9100ba36445388859", 0x72}, {&(0x7f00000028c0)="7dc3ffd76b3d30468ce9dd4347e3730944956066ba6dce166224360001f2ef3cd5f3ff11b00a9dc200a9e5aab72cc5ed95d5470c70ab880b155707ae7ee9480b289115ec4350a6834ae1696081", 0x4d}, {&(0x7f0000002940)="e05a03afeb48212e01b58efa94f8faae255dc7dbfd3d2921e1a9fb40cc23f8b6f9db2167ec5fb3d94976f200db93c9a49648ea5956407c414d0f596d676dd147a7a2c72787d03c3eb9af592304381f71795b8755fcb49fdfb99f94b0ebf6c59a27422ade90db42b9974b5fe939c86714717d9f156385597b2afb9474c93ebe", 0x7f}, {&(0x7f00000029c0)="637d1ecf2ea1ecce9c77bc", 0xb}, {&(0x7f0000002a00)="c7b5a7d2ae7a3930655e7107eae6c3e0ee94ae88c430a7d7b0f7ac66bbfc6e46d4675ad186a616b83d65fdad901ae83a3afbddcdaaebe408c0de1a7ee7461758f257877d4bdecc27372aceb1b0d6156c1feb4603ef01e6d1d8ead56408a8a4008130088ad09a2a5c19177f29017191871e3ef6262e559997cebf83c3c4e2421375e8093d5cb61e8c7c86c41d5d5d67dc1376aa8fec686003776ccdb3a891bf3a063a0d3839b602f00d684ff21985822c18da2df541080946551b1c485b58dec28aaf1ffefc7ce2d80a7d5522a7d2acb830fbe41991deb9700b7df8d07c94f788db83cbda1188cabda70cc997f337578662a0a4698c3c03dbd8373850015f23c49447ca4434197e7c4a5f04c4fd73e3e60eca9460ccb97b3a8f88d12cd82d6c2eba411ae33e641cdf329c46fb5a31aa15ad5b92441780445674b46590ea6e22befd743ffeca8f50e8a9dcc85263f40d717cd788623d19894e36e037b4ad4e999c14accf0f417da8500a895c6bb1b44e4deb5f6eb29bdc3d8ea252a0d8291cbe0e4a1d1661645e9b2580ab02cba738eb87a8d54ec3aa2bcd81b0b386c60f4d993cd012885a9a8a2a61ac5c1785e06d7e550f4d477ee3c2c218471517cddb6291abbd2ec77146e78b5327be7b9de611d7bf1c8e5e0a461daae48e9877ea1aa280d116f7d6e257bbff9a1907bec4f4744bf064fdebb839c5621433990ca52abcdcedcf4cf9faa9087d51eb32a59fbed2a6cb665b6e52157058382ca48209fc1ce45c41961f90146c9935da2439bc229dc79371e520aabf801eb42a928680c4c1451c924c71f5c971fb6e5108292fcea8e3aeca60f25ecdf9303cb7465384c09203d42724aa37dd7749eeaca8e7219a702e8730c334933fe9545848b6ef18681ffb22b44531f458b1947bf4a0ff30f31257f90928ebc14d48675b049a2aea12be90ddde30539aa2775204ab7e64379c24bab9281616775b805c024b81c11f2d846eff8618d817180e1fdc30c63e183f6de9afe9b455f5c6e07e5a12ab823f7f28254da826ccbac164a9276a53b5d547817f9348c087c34ab8be7707e7ba6655b3af522b7135246f3cb5707c828b252aade9f4c8fc423a03b24fd36c077a00f89d94317a85596bcef84962b581c72adba8d5155726b0b16ecedda5b2d2f4b5511926a3163cbf3233893ec5c0f42979b71b48a8bc36b5fd31cd4cd3a68a582a8ef26c78f594a426774c767b9cbc0a51f7a378119ebb8477f159b2d8767ed4bf4f079ee9c94ba772f85bc887f58eb24a491334a87e468a48749c08082f5a30e99f15d0546e4926b6509d311037cded8dde1f300db5679391b42bc689e5e332128536c1ce66f82557f404d0668ba8ed72d47cd1595ff4ca73f3569dc12fd09f6735fe7b34cd0016b9546206ff0615762a786bafce23f7e977e58ee659dbf119c523e1be4fa6d01b214d74196bd31e6c9bd4d409bf9f72f019e73155fc8593506788bb58a6238f6e89228f067692e7a56112d05db1c0d990a08879dcd8fa2d6ce2cc5c7aec0c2e08076d6e68f6ad7197e92a9d0a8aa101f5f9f7969e715ecc995436930917f4b6315e06af1e38064b2a17fd35bed2d364e7b997cf85a2201ec5076ddc00baa7be4b99e38b8e8cfcaf3af75d179b0fcee9f567f3160ff3197b0e770c8f2fc7f5ef3c53b96f0b59115989ff9282c49bb460a7f4af81dae9578b51ae1fe2f457325f9f7108fd1a7231cd05cd82b4bdf2929a8b6541495d181d7bc05592cf9f869f90e8222588d191612e01937a8df8855cb8c16302bbbd0b9930ff89e909498efe76928bee389391d73f29b6eadd9be765bc4776a1a97665ec953e9a1f42d48fd778890927060e85c3c4ac90d4716a415a7884356aa0e9192115272589371d05d78c94520532a0472bb002b3644bf0c2ac94f252c02be56fde68d7ed9cdaef35fcd7266435ad071af4e9f2c4011f88d11436f1351155828b40814b22b4916638582dcfb6b1e0323a24e849a1b043f158e05770ba67c030415ff1658c1183bbb14c8e1c29f131dbcfe843f4c6925437ecfeaaed6e200643b1e46a8418f2fefaba3059871f4125fc76a562dc495298f39312b6d790cabad691f1068d54d4209d863d390d83c00ad052a7d0c50969c7781dcc29d95dbf2d9101ec9218a0a84be266cf826f110033ccf5d887bddea81916fff6910846b7b4c279750a51afecabab01df8b143c7b7210020eda081216f19c3383a2ec7d60c4c83d55632ce28e50a537427131d99770a0ef2f983fa091390540fdc1047ae0eb8c7c3945762595229b9693c7b53fc7da97b98e4b03acc1a6998fa7ecf00e958aa98cae45517a8059c6ab6083c014c418cc1473e5e25df311432711e9cd582f8eb92c6890ca16fd2a2ffe87e95fde28d0b7ddb2f7da90291c0e5cb3418713bf8fa33a9c82972b5a9b6c2ba33009f31d3f3f49fc0e81814db0d1ce1e17121eefc6c81d956a694a2552529b525e50ba1b5947d3412bfeeb6b70306c7033a0009ad0fee9ceb547d299998f6d0d087d92cdca87a25b7591aa236de697c677771f93db1c23802a35dcc09d71485f95021044d227f4477ad8d463a99d3e9b6bf05685d88b2d7dce309ae2b5773cc46cc4110632562a6c56f6a67086d501f003d43494557c391f2e2129a9dac704b78f4df2b3747d64bb62fc6002300a0153db3e57e93e3f844ecce523ee9635710831f61815a40f6d505bd525ebeb9b2024e5b82ab4ad31ed2a9b0a6b9510171ab3a63bd38bd0d5e5dd6a442e62ededdc92793c20090d4096458ae4776a19510a34daa6bdc0c6a997238e46994ae36d816295651b48a0cb7fca48313613b4e82afbeecdc4ff2101ec05de0b1a8d17aa33f3c318b77b4c9aa7c5a6b7106f27217dfa8c5c913275a0e62dc37cd927d0858d18fa769d99107b6b0ddf8d84a91e2716f04e96b8458f51c1855d738775b43083ff3626cdcb1a6318fad4ceadf50130db78a8efb161552b33fd43d899615bd344669dc9833f983342bc84f79ff152eda4846015d567604a54c1dc0dca43bfd9be95bfd604a775edf26e0d11e80cae32ea8d6817b9caf9404d5e57dffc148bc9b913f4af57dddde165d07d3e63759122dd9256be4f9b549d6186693b9ba59a5115dbb7201bbe174c0f2a5490da51fd8481a348b803faa1c7cee14936a61e03a9f9f725e93b476a5358dc7408c6f2108ecd0422a9293314b43e64aca5244756dbfa13c946fd53b84e697dd0f40844bd15321c53a14088332a4c9ea037e12c7d84d5bf829f1c20d5af3c68c4eed67593fc29be21ed94ddd85a787a8e60da347bec452d036c867e0dade62befa899dac74f82e14ff1412613eea444f9c03a1d256da00e19a989815f3ff9e844157132f0c9e04de330974dc73b8eb84e2f1a0cd8eab31bfb7fa8b08484cb7ef7b9ce67b408395398d01ad7cf69346c47e27592aaedcd557dbe9c925626be1c8ef1c1b18cc88e26d705b58933074dbcc723aeea4a9313e1668a52a5178ebc8fa793e5766c9e09a529aad38f3c7b087226b4930a5cfed85ce2a0a1423925169bc76c253c5474b12bed3dfee057d8a4b3b19af4b18e31b2b3bc4d5ed76502deda6f183a345f718fa04e9dd1ce289d2cd4c1fe57f63299a7ab5973ab4ec8c970d68217a1fa7e87b8b0653f60e3ca6cd3caaf5d39b9e2155fbf233c3634a85724c0abc4738f1da5e833fbc86ba87ff9b2c5370935b085fb4246fa076c8ec0bab34bfc092f613fa3682c6aaadbff22466cbc651e3d42b87ee1dbc481caa7c932486e45d2d93833eeda01b2b3c43801450a28f27f2e829838076f79659c353ae764093bf4d776c3856576442f20f2be892402f428531c85d02e2bf05a454dbc76c3d2b3661501d7ef1773f63bf1b3ef6c50999d72d6a4502b34c97b2d52487d2eea16830fc1413c468c41739c6a9b8387836d7ac380b0060d87a5da451adbc1bda921f1150f85d46cc308710ce8b374c68848160bbdd586744d47c16a5be91a38c880649c50547bb33cef645c6d4c1697abb90c2936a5a4777edd99a187cb37224948d276504948b01982684d704be9e7bcd166cf0cca6a18b8b27af81b23f7e97782ba404ef2c5b71896e73a8f9fb80d585383dd35ade61562eacc87cbb1af5eae31aa3c94d13fe13c03c9585fa20241acd6ce20065425791cde6e3932d1e5e871f9d08b59f6d21b1a5760bb8caf23df2c5dc97511fde463937e20ceb17f32b7726818665a44e6661e125c87a80c884b2e1202ce72eeeb8b0b13f29feb7537baeb47536cda8bad897ecd99950421b2b19a31f8f9a26b2e0a9b0900102ec0c142a4125f3b750d7994335faa8fce4ee30f797045a0ff5a19e84d2a92c4b10042315630dcfb822a748f1976f194fa976423aaef35f7a6ac5354101f8f069fa7712c32722532668a1955a26c1b9781dea3dbd6fbe3c728de7bfb63d6cfc724d54ec8c22ceaa1046c3bf9b8074f4cc744c123bd39811136818c6d6271099289586928e3db6b24d8b3349ec5c72a2d6a961d83295e8fe3f677e69ea1603f008c08fa2602589b42af23f90ead18a6d9e2b1cb98ada803063cdef11561145fcc494c69d62b2866f9d750a04c90fe29fbced3c59f74d78755a912c3c741e294262f723ed0b973e4db06a6dcc68dbec6b7d7ecbb41e3dec682ee72c3b5f9f83336cb3f0916bad625cf7065f19481e0392003ec49e544993ab258fadea7c023baf0dd8bc4a7e12cd65ba2d749315c44658478a16c92cbf3a2b2a0d355a85995a443dc77c64f3a2efff639e3f3df9584bbfaa1222ff599ac35a6e36c97a1fa7db13573439c4cfdc811f24dfd4b3f04d5e3b87a999bd610d838a8768be189407971ab35e27c8aa421d5d265b235c95dc7cd77de9fef14e2f004a1c97c2bb37c5e68b1f8a1b967ccfcd4e02d3e27002009aeaf4b75efb7f86927a70c1a3bb2c02711c2fdd99b421a67e951f9019609ab484fef6c0259c12aae069ca6bd6602375d85f5ad7d3274d11e131b26e8f4ffca0b3af3c931b4eba78894ea2f2c1ff9a5fbf8f888bdd42d3bb4692078c18ddd0c59b3bfe5ee1e3573844b37cc54c98412a1afea9f36a4bb4e5dff95f73e60fcba1dab62fc65d6f2b9e4fa49498d6255444fe350faaf5c58da1e5f7d3d4d5cd476f32ba9658f2961803227a16268ee8fe99594eabb6364d307f8e8abb35c01cb00a50c7f83575215bc4268114c66e59869979995b8bccbf107ecd3bad2eb1afaff320e6b439a052679d9197c7fb90745e8edcd6d1e50cc218607226a2be3d2ce2097f6ea909c7d8f62afc3fcbfa93a8bd6efd0d3849a3db2e4e30a2c144813f4b91074d806bffd587b9c7920d3856ff97bb5dad847d52b3d9d95d2df50da5c2bca618af08bca015bd2600209930d8695600a0bc6fbc458b64342cb440e293acd114cd5daaa545582d63ebba5ff6ad6eef26c26fd0c4fc54a5eda7d23886ca159833b9d7dbe65d51893d5b2a9a7ba06a2de4e43dad43b3690669dbf01ef2654ef140b06bb3db6f14036a1a04eef010255cca929f6e86ea5e337a602980653fc6f7022b598d70448a86c42ab41a59104c1efbb95e90b47fe3442939e0ae8969741dccf1c1136254f96e5aa9f738e28947390a376ebe309e93f5578d0535f72989cc9a4e76e8f4e4566cebcad06d7c384cc4bd1cd6a2d2ce87a5773a110f90ae83e60baa74b6c20704e8a1c8b81b7f5d1546e2bbbaaba8249a17a506fdfd2084fccc0492cb97837e90d2ea1379c5c7b8ed98a321383924432cd57a7403048bb0da69f", 0x1000}], 0x5, 0x0, 0x0, 0x80}}, {{&(0x7f0000003a80)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003fc0)=[{&(0x7f0000003b00)="93ff6611eff8657623e1ee6fd9b708304afbf0d9a637a6f1d9919cbad0ad18410d0b8827dd19fb88ef2e0f20f9eb0c038e7812e32df4a0c119107cdf0faf80680ee407a3196883ddceecd0d80f111f0d55ce7bc097066cd7a28006", 0x5b}, {&(0x7f0000003b80)="9a1d914fe13ab7183ddce7984aecdc90c01f16c1937d29b54527ddc0eafc0a67bb4cdb7a0c6b69727c1a33bb4e0182325e8ea3d1f3214c5fb6676062d98e8572585151c17f42a5bdc084ee77dfa252d6fa4f60073712783b5dd75e6cd2f0d0877d97ff44c0a9c37b94ccee98d94d37b98db64a0945587014510c627f2e4aae71801ca31f0fd4947c1ba4ffd98346e3966fe13ff17a9d60332a97fe20723f950b69fb1491b09e120c63df633b168eef37121aa2682ba3ae84cb44c6fe3e19393a42f9bc15fc4cab073ab085e617321adbb8df3dd1633f6a20e43f961869dc31", 0xdf}, {&(0x7f0000003c80)="aee2cd94b96bb045c083e66bd17bd57d745ea95102f3e5a7653b8e295b4aaf9dcbbaa494546db8157f5d86a86c0d2274f2cd8572a155d73fcc0eed3a00e7967b5dccb3d3b4ca559416dec340c9b390b1354a1ff0a21074152186b8bd72a4e7f766a81a9c8e8e0af80cf2669b7e4007415c99bfb384385ccb6d850f217f21f20b858556cef723dbd1658600468074fd687a082efffe3b4be72c61317617dba1fc45994eab262cc257e42504fb96803eea4ad0088a2c081e2b9e09c2c1a69d3ebb4e5f62c9287cbfab4ef3f1c7aec0c713b597da5219637b42b26c8ae24f3863cb93d95238d14f", 0xe6}, {&(0x7f0000003d80)="428c17a36c4ec58b0147c9a453486af940f4bed12995101c294db45fd60abc9de1f464d15454929af4793d7e79fc65ed56f769a312d6687dda6f994a676a098854ab67e1647bf998bc934bcbfa5e92a61b175756844f30807bf4638c9cddc76ddb2ee7250ceb25f370b0e0c8548c6fa493bed2f0491b9f963dee395825ed4301f15fc7cf91b23cc5e268b1a2fa65ad781380257a185e99913ed377ccc1154364b7", 0xa1}, {&(0x7f0000003e40)="3f4108cd190380833d4ab89250830aad3bc1d9e2867398717b567ea37a63b3a8766a44ce38ba9b0add9423b9bdb4896deed28fabc912ae47ec8d0da27dac1498a0ed9ec28671da0486541d1720e3c24e2b57c65875409ecdad306ada8cb9c19cddeeaed5d832c34d235c1ee6cf8d3b882489f2e59ad5dc46bff747447c86d855410bd40197c6aee8896d09dd102089703b5068244a938c4ae1585fa003e4e7a0b52d79d1a87e1d1e15ab3108c66ee4a2235d", 0xb2}, {&(0x7f0000003f00)="083f5206abe479be4e32320f648c02e3bb5f8371db74dfd6b8602de5f39908c105694c789a4cf28bff8f4dc5c790fd408ff02d3ecd537e901bba127e2f7df76a76bc1e5b0fe40afebbec329c8d516616b174cca300e15251f271d249bb9fd581ecb74bd9d1619755914dbe3dedf723027716a8e69077055b38d558906dd0ece9d5482915d5a015205833ca75302e296b7d96e6a3b8", 0x95}], 0x6, &(0x7f0000004040)=[@rights={{0x2c, 0x1, 0x1, [r0, r2, 0xffffffffffffffff, r0, r2, r2, r1]}}, @cred={{0x1c, 0x1, 0x2, {r4, r5}}}], 0x50, 0x2000000}}, {{&(0x7f00000040c0)=@abs={0x0, 0x0, 0x4e24}, 0x6e, &(0x7f0000006340)=[{&(0x7f0000004140)="b2ac54f987983f553cec73a48162d568c71497b2a34a0964410ae45af25d2ee5262974bb9298de01b5981506", 0x2c}, {&(0x7f0000004180)="7481ec864b88325813ba249f004091eb2fdce6fcb3e6740cc6eb33ef5494155e36bf889369ae354303d674d8565cf44e3ce51074740b5c9fb6750f7f14bbdc65396c9f8a11dd62c8939c4bed90f2b53886412bd146556aef9d99af830a360fb7f71d7599caf45580a1ec09716a4e30fd6880c390f1dff8c7b0ae3e725b79cd273dd936593c0cd4587f83a5b801f898549982af796c30c996d008ed73cb2c43c97e122ac0875daaf010160040c5", 0xad}, {&(0x7f0000004240)}, {&(0x7f0000004280)="fe692724ea2acaf9c9003924deec4519a92fec2fe89ff751e084885c5e07c26ec0f872ee0d5da8a8ff66509dd0d91bf3954468c546fa2d7bd1eb3edc727c9f59023dce5c36a79c2e0fa10be91e795f57d3384983b2252fe3c7ec74d8c544a266d2b24cab396845c9efbdd72affe93e7f8b59596fa8d8e32e73c48412c7df3d54006fbc0dbc2160268e0c57695e91c9f45cde31", 0x93}, {&(0x7f0000004340)="7b4d4f3ae789a3a473e122f6e29c07b80991a15caf3b678623fc4add596d5098171c0257bf4f47db1290adcfa0c40cf6091171e9d66b47e07b7f5411a6c08114a2b4c759934fce4d141961b69fe3c94789e6a5f46171de31f6434cc658d087d437fdc6d3e6c5f3213fbc7df8bb1434d736b7fb63ebacaa93f67c01e586f2dc3ac4a6011c82e6f8c7f679df69ce52e4f9f9fc3957039c06d051f09e14f0de96425df93d3ea593ba8049fb7ba281fa82ae2b661dbb33530e864a3dd3cb09900da46bb7773ff4dfd6cdb9ad0b7d20845420f5ddcb2cbbae0b6fc949f4684267677c0bef29db926556b296927912f298c5e4c457d2f4275a7a8ef4d506a4bac672fd2b45818169864f84bb3eb0ef760f1b100c9054b50895fd2490f92caa283bb0c2dda891e1ee9a6a446d1faf315266333ccb0670966c1e850664a6fa5c04e4c8e26ae168cdde25a72df1f0f3606e4676f72322c88235f8883bec9e2b3e1967766593767798aa99f8205dc889a230476e2083bb07fde29dbc1097d887724d3cb520e0a8224994e8152a57876ee1dd4ea38a7836027be9aa24d2fd4fb5ec3e26d1689eb2fd31ef57368980cb5ccedd87af9f30d1c469816e12ccab0982d94d720a99ae6c2b9c1206ea401948a7faeff3bad89e35b8b063d3cb1be663dcf8ac08234c71d662538819b5cedd894166574fc50f250b567a78633f72a52bd7e1d73893f260954311d19ad71ae7f612d500f06d0e2c3be68d6a88ffafdc8e76f97229344bf3d47dc88dcedaa08b62ac795562a4f68bcc66680ced83a9c3f97908a0fb3e176a3ecbf5a933fdc35a6415b669c1052809f9edbf373db771214dce247ac8baf144725cb00cd89cfc1c0e940ddb20207ff734d326d411201b1ca7a6e30f81df853d8512c74c2979cfa144a727365a32bf315191f9b5dcc990aaab9914a25e62463e409095cbb8f738eb5f0654cf086e5a4f44a75ea3ce882ccccc91b59f645be79cf7a4ab2e7490518928023eed82494150a9cf1648bc9ae1bd1b2e2eedcc2e7e7e9f513a975c9bc195c6ccab4a91011d48ab2e65f08fea07be7c9d3245ee9223086249fb0e33e4886e21ce58db5ae6bf587d507665df202157bac11b650e95d9c5a37ada6c4155b95ce80b977dc74a2a29be003876993da3ea7fcbeee1bcdeb0e5b5a7f9ce639f193c3c0b6ba166df026e3d33f24f24c2f1986109cf7ddbbf316d8b7a574687c272b0d13c453496d9c782982865450afd2cf549b9bd498ce3fd288916c1c03f558e3bccd8f84942bdab8cae1c7fe5d230dd759ba1ac0bfe3c9da0991f701e5d09ba809e0b075a1855866807c2c58e54f4e4d7179e1fbc466d01d1c6dd51f89beee5f5566fc2fc0d85dea3a4f48749eb85272a119a4f7ba2a22aea72f363ff5de7f1682bfc0b7c9056a7b5811b3be5bc2e99114ab20bd21cf9a0cf46dd1dbc27196ad31b4b7bbaf48dc2f4cdcbcb6920008f92f0fa1bcae6f3952b8419637d4f1f3a6afeb5b10f20e6330e7d54deae385c3b1c38e0b89fa113bc018b1289522c671b2d2be6233da7cc04bb3b30205c0ed64545131ee3a435593f98b14e048d263eaae7aa360ac0ebeaeb6c5418e4df46369e0127caae6f69e2cd9b0061b6331faa888fa1c30e20ed826fca8ef222af44e81987b4d479f4885928e957a3c5b3e3958a428a520a6d112746968e5957add6ce2a084ec386d0150a9af9c4b7cf9d2b5011f36c257a40dcc56535a5edb16319b70a8d9b3a7e43dc93c364b56b41b1c2ba276cd238d0aaad61d47b7c6ca83fc8ee929d013f9e6aded695e74517e6f4dcf51af8189f8113407749e477a480eb4554e63765d1eae015db2ab99979fa3ffa7289b8942fe79d1328443ffb9a62201730dd55982ca27d3d9f65d35fb8ec3a49faf201e9e42b49a1be5c672ff103183c342697cb5e2b11d0cbd56605363e5440964466136bfa3e8af4b1bde828b4e3ad489665a9c69c1855c4af8fe62e852534ff93860dbe4876b9940b8c9c38d84f70c394075a90a5cbff9d67247287c1864f5e2456932076dfd48ac8651c4af3e975d42deb4b5ad0e62d4295d110e5a0df6f487efc4594f71da9dc8d32212afdd99b58465bac2a6e0623d9edee4b67aef0b9a7f2d6659484c101871f7180355113e19f2fc30e7e44cee64b026e9bec8a0548836f30a90999b7c0653e9caa0aa53ae8c1cc3197385bc78141490fc1c89d1cda2a493da627cef766f47fceae20c2e5c5229c9a52e4aa8a403d12b7325e99c00709990852a4a933d4e454e6ee9b9ea868e20c3cb4f4b64743428ea2444dad2a32a70842e24618209f92bfa332d0a05e54d02935602f36e132202036d74ca50303e7b3951a5bd6bcff88d32de3c4dcbe82701ccc67c55ee9897595979f258a04b78cf0c0ee7dcd4d970358a35593e4cf5c94a4387576c4ac3913237ab421d71a281fbcd822561abbd018afe9e9f5c5e6160d4aa5c7f40360ff77788d94d75d1d8324e7e71ebb4a5b8892a47ce06e16f3d9d22ea7abdf880c1b2971771545e834a96ce2ba1618ec56f62de191c8f39960d4738bbef32fd9aec25a5e035479a8d08bf9b439561179378de776eefd732bb6d4e4b7e577246baa8b2958353aefd284a0ec8f6ed3238e58ec15b98e23db9e62ae99a69a47f10e880a69cb9afdd58f7a9d82db24dc3608f6bf8c260547c538f27e18de6d13cc262efcc94c591c2c109ad1dd88be6448cd4a09765f1bd0396cbc34e03d4bc2bdd9135b655b32e6cffe26b8747205b0920fffa8e5dcc588a23ec639de23bb1f198f9d186d6a5dda93e72465a0fbb8a46fb54fa0ead10ba268048563be84669fbe6e708080a896186f1bfae5a99457640f6abcf50b0314c56c27cee0182b4277c48649f8f0b71603a2f46652fe29c1c822bf5c97e38e8ec2092f5e09d7572d14e6ff19998bfef27668d209a9536fd152ea4c838b4f8d802d52524429d1556063c4179d117bd5c255ab64e45dc7bb1d79fbe66c32ab9bd86a8c03b324cfaf9ded18f275a3ef84557bade1e7e51dca4fa3e1a8cfed480676a5a30f773b6ab4725d37fe8a37089ce8cde4e2121dc8558595e1c48c227b98fb7ade12d5bd6d78dba20ed20e881c2d927a6afa73c5cee88fef6bf0adf8b04b49cd84776708c4b127fa66da53423688211236cd706e41200601383a2e30d8109c2e7bf23bf58fd55ce7da50e770c121a958476c086c16539fc1757a6ef853ad8b5af52d9e6a0175fb1cc2f9f3318ba3a23a5071de5ad0d91ac045111e4c51930ca9e77150fbf6369e388dcef870ba5e43af9a23577ec6b07a832fa2cf324874e8e83c4fe917f4f936c5554b4cfa04dd21578b6fd66810b9731bad1fa030b6f7b6306c9f3db78b6e67fd02080195a322b01e53c0ac1f0689c6b5da6ee8a385804c49150eb65b71dcbd7f7d906b2a4c4c5bef4e749b71f0a30a83d76f8b71aef69a5cf99bb6f7d5adcdaa008c4fb57bb0d3eb4c63b2c14cdc4bfe3167bf2341ab506e18cca80d6c683080fef3fa17c44a491f5dff15eae3e98d5d025c99e0683bcfe74b294c894daf34d3276636f87d312dc93e15c3187e4d3aa4b5245a5dd0dcf2e2c404a1e9a961155c9b89d6d35486935b591be513dd07258e2cc2a1d90e502c457917aabd7e8cf8344fe3298bb7af6ea0900cd7f1f6aeda0eeddc6fe342c600abb41f8615032c4f38f29f224fdade8c5704363492ab4b5e4630124158fe18a38d7f5a50ee94533138a2f61790e8ccfa5825d2f13be04d378abc5f9fe42431e5b94db1ca8377349eacb5f40873a177f7b51c1eb80aca422ea8cc001cbbb7cc7aad41929d6cc65006ce9b74800747714c46c0def87eb3f327c8906dbefa50541a5fee00888271e9c964f1a70aaea79727fe5000b2c6d7bb0c1229cde0431e3aefed2ed43a7f9e5647c55b8d94a17c5109b4c95e4714ded2451597b488b19178ab0cdc13d2319159f04977c4ad778a63565cdd37ca6fd23951804d4882871f9e5418d2c124aac504018c687f04d76230636212ea85ecde848e306e7e14029c54dde992cc0166c65396b55fd44d3004eb1e4817bd8c9d69a2eec5b3619875fd66398184a3a13608a668203f2344ccfba153f3fe5eb421c8ae137649c0436824f576e2ca75cde8717f85c458b4d0594267e180dda4309033f2b8ce4f842750e78a6033040536ac04010884f318fb2098d824061d17faae253a6cb3ee3ba537fefea2b9e5fd9684ed1bf029d19102621d01726e209a208f0c389fa3d3d94a8f2ef4e03c0c075b23a2943b8aa295f72d907a293d62c9452a22a8da1e34e7fe439817ad024f80551dc1ab7d45a1e76904626563f8708fd51e0ddc5821b037e7ee22dbd16f29854536d802ba4f73486046d169bdb117e8ab0794536204e9015f6d123117af8c5f73d29c1257b697a5779133c91e1bff56e57fb301f84cdab9311d0df6cee24cb9a636fe7eff894f8a34ecfb91c1939c6ab67e3d2e3594e2e1fd521d27e6cd0aa91b33adc1e2625a734cfb18929507eea166c4964b8d72a0c8971220431395a063ba33bd832aab6e71d4012f74e164693997e1f7624838963471fa393b66a3606d068e0256720eaa06cc8d03f957c78bd73877d124a5ac897076279303bac01fea1e98c4abe55497db4dec0725a9fc5bc5112666d13d078bbe50aff90a276bd000360c7b28777693b5ef46b4e55d6de048fe0910e9740eb360e1ba587dc452f8f8114708d74149c9bae58f8649d92270db4c1cf2e515e31d548c7f81c156ec8552674a5edb18be917ea3dcbe1be6e10dd47352a5e127ef6a2057090aded6fa2fd18776d9a44a6f108731bc806353ab1a1628787050ff4dd2d111538b4fa99454c1157c833fba9d14e6b85f83808af9b79096f9b73bf251e16e75e84ff57b28d3f0072d172d00ec46eef89cfe3e8b2b522c0123d5ab49de0c07e7d16868f666080a70a0886082885cda016c9b8c2e6501158ece3d15cdbccfee6ef0827adc213c59bfa9f657e8fd6159cab98f75839764507b528a7c7934d8be2b35766759ae01917115286e55302086dee8881e114582caaf5b46e8dc65b138ecee44569e436a77ccec24001245a0b6c2ffbbcaebe5ed18881d7fa77b1423144a14846786e529ccffefb12d6fb77815821cf9aaffb8b6cb7d89f238fa4ca690c13cf36c1e356263f135fd2327c9d6d424b96e6ed6468904f0955331c5faa7f876549752d37c315f4c007b2aa073953318e2bd4800c07c00c9a9cfe643a49f7353cb919a3017582a62320c0ff189f659ffa9b26d117a7eb870bb2f0f79dba6780b0592cc20d7e4f963ffa2d535986841f4e3632afb93d1bf89aecebdf46569592d1e948159390f1d483c8f207827301be986831882cac2a31d55f583c989c6f86ceafda20a23fb1f8338d5b528e8a6adc05ba958ccc2e89972697871a6eb6591801824eec171b28959b1b2b9f7876b7ec5f0e79c1e291aa3ee9e10c9ad9bb616dd5602768a0bd00bb04cc864fb13ef7100513a2e63e1aa3af265c0ee34fa76c8c47b272116cfb9ecd7f8b26aae8d0295d7762151d9ff7e616c011e7231da29b88f7c3633c0cf3f9b4cf96c947b8fb48d87fc9e707544587d7c3fe5900ebdc822b299a413348613e34a2785e7a4d522435012e7777c3debcc7c081434fe5ba94d620a878868404c062e495f797f8a62d7348c69e5783bee0c71d350a122076a79957f330cbe5054e63b0ded895b084b22b9b09f1e7a11c9a517007781715995a504397cfe4055bbcad13583f6701077be956a2888", 0x1000}, {&(0x7f0000005340)="fc4259e41e9334ef8e7f29c39fbfd5cd583afbc9faf902c667fe1ca45254c386415796c67f75c675da6f476ac3bd5b83d94de4f5f8c70ef580d06ccb860eef5b4b5d3bc94299d4d21ba677b993dc65721c501bd3161f558155b14c9a3866999a68481ee18ecf60c5ee831ccdbf9444764a4d900231b8d190ecb74a472118ade7c7b8cca0e8a89fa3fea10b6324569e10576d2e6c544886d34a3eab44a3ccea52906d3d5f80344a66e37f447e05addfa0dc017a2b3bfdd530cd07d0785ecef2453d9b817b25732682d82bccdf3d62a99a3b0dedfdf014b83a825c52c734cd3b511d900bf3312dfa7c9fed4a4bcf35d33c2cd8230f690a68b3954eb6a68d79bf0e2b860044fdd1168983c427374960b256b4fce42e19b7fb75e371e253bf329094560608b8be5f625cf385929dfe8c63e1cf7de7b63ea6a5e25f61f260ed8e80fea02f4368be828d07616acd0fd6a9090ede8433f0f12a3f81927650cf8656c4ffe585c98c7a0185e7026dd53b5caa1922f319fad7d8153bde8bde6bb8ebe9aee764e85d5aafcd6ee86e42bac1d22ce9b8981381319bfb9810a350dc39c70501d903bc2d1da14fbe4f97c666cd5a7385d72a891fa70e0d36df3d59a09558052c400dde20ed5d15278dbf4a3d04382e0e8d91360816f7b428b30ad4c93d5481ba27177af37d0baedddaed1c0165fcbbe01f9b4415f298074f3dccb1d4c2c80d219fb835ec570ec5b4b6159fd63f0fcdc6c53d75e67ac48734eebe0f2d0db0844ebfc8874c86512121750a62a5ee12108777cf64f49113c6ab8416a425b0180f5509f8d0f56928983c4be698ddf1a0424a4bd0d5b23ad9cff94c9ccc9356dc33fa0651376852832519ce00bc3af08f5b231d39e04fc3867dc8ece51e12288f1f315a7a8adf794c9f8b154ed3492f75466c873549de5621bec2efb8da0d40cfc2883516d3ddc3ca6fcdb3b61a0970162d781e2a1f90433df7d01401e4609f33f30ded1f486aae776fd325635927602a9c7b768e9de151a62ea613450a925f1de616d5465ed6f0be965827551638ee93f292129545e648c6928127199860a432e70d718cd9a49527e26d32bc1130da70a7adfda133602855e3f582bde618350d01130361bff10bedeb0a0c198aee6b6f5a1f89ccb5853d9b095ba4d25c2ad9bc4d61a407da5826bf0e434a7261eff5e6eb4b82e21bdf5e74904bd2999d3f6b6a5f5176d1d11aaf0bef587c10c403dcb68e8aae02366681097f9bfd3fe30ffa2320ef7a2442990d23e9d9bc7277bc10b5f30100d861401e20f281ea76403336bef6450db778b29dd8afc5245986e9e4ead9d166b94ac4ceed0af1602a07ea8fc6db2c5da3f3ab0a45c2ded381d61fbc494d160006191f43c3b24e5e04ce11124ad884e05fecf8901f14161456ffcb648c069faee9c78f72eb57a6fc77315000f395cb95cfc363889784e218ed7f65bb17b7be37faac81b3a35dc2ab8c5da788df9ce7badcc121be3e7c71b4ccfbea51631c8fa43a69d5e8c0169f14e1e3c742ec2a813604bf3391eb78e349c0d3c227949d6cfde81dcdf02d0e4b10fc4723b702078b04b7cf72ee232846617270dc4f3830258fecb74bdf3740dd178c3dd1b361044d53364fbcb7a22bf4a660e4d1210aaa97bfde18d4add4cf6eef609828d4200afcf9d0cdca69b75d3d67835432d379084115e369b28b6cd2fb62f0eddcc618e4df229e219c47f407d8f21c65f77aeaeaf9663ef2b731880eafc3f193ffafdc7ba8908ae6936ce24fc89c314efd8758b760ba2a3d421a4fe203cfef64409a26143bb682c68a5ae2b90a5d1681a2662d6cb3f5b45f07a07f6899e3fd4d1f99af25a5dd7f655af82276832504b3bed1255b137558bd13c7ece490b6be5e5b07377af1d502115a947e183ef52a19d15b357f3c177ceb07fa236c9ffde22178f40ea10ef9d68cb3f4d39a4108212c22945fb0052c3818f7c8d219aa04478aa6ea504b34864eddc339cb414cf7a17aa9410d808e1a09bb8c58c0fed0c6038bbe8e6f6c4e9f99363f93bb171ae051508e0fc06c8dde15b02c3ef9032fd704c3bf9c33cb7e13e6fe19be7bc54e0b26f4ca9255a293e25904a3a6c68f0518d52c5f1a9ada641da8d700da1b61ca223aada4b4a2dd7b80f2cb04b35e633a3edd6485bb7e9dc51732dcda1bc915065d55d9482a73f77bf50047957864d31935640111684259c957e9606a4d635d3b1aba1c0a69e6c525abf79fa81eab70d587f416c609aaaa3c7361504e32dc02f8e7b68ac3e80ed629da9d928346eb2514b6267b19c5f000150cdadf739f7e06858d7d47dc90accdd909a5573f56529646ce66c8b50231466ae2ed5ad6354443ce770032db0a06e0f06ef024c5844f6a562dda69bd956b8fb792da3c4c77d0a447909ac5b2532bdb6936bc67276ca107281628750890174b7144ad889f50674064a147e012f8cd585a6aca31dd51fe58dcdb49f8a82b97a3f6c93116bb46d2bb3565fda74e6ac4329943367c85d4aa89512bbb6382c5bf9ed64c4ebc34bc119bcbbd930f38fec3bfb3a881e56847edc477f74b319f600f531a19179c8e89803dc6369b5b7cfb523bbf0849a092f25b12202daebacbe6e48e56b0928a0c8a6e730c06d5f19e1ec10a6467b0a46a9d15e0d2e84bf624e9fc4a8caca128ec65fff6822b37c8e14731a3f6c96ce17ff30818ea82db9ce18bb551aa596e005062cdd5bc4b81259ada5d45bcc39f0121c31ad5fa8cd55b3a1284f7b6f37f4bb8449d8d5f8147acc4f215a8a6337de23770e470a129b112bac0f43ad43fc3705cfa07e20f4a434a0967401c78a95da55072691145ac2beecb634402dea8488e6ba8934b507e55f239b5ac26bdff84be6e7c45a28238f0bb35bdd613d13379750f40ecc469a27b497242314c7decfd7afb2ebf4073cf8052ac6bca42d39aa69ba3e30d3cd09eeb15c4f503a335731dba43551986559b62fb7616d5b75968df4a7bad5425355bdb536f6ec3600f487b963e6beea8b2e96b5aad1cbc862297438bbf34b21b3d27c1942198e178408cf336a05139192ce62d483124a0fbfadf2ad62ee2504332a59ee505c851dafda2f9171efbd2a25db061482e44849ab81e14ed34fdc612bcb017994348e9add253860291a369b24486fc6171b5f32ab48113f181b57769fe1e718ea2678ece437a9703c7af904db17612f84c3baeabfe660434db94b6378945de07f8d1433470cfcd0d12ba13e9782be71288a44564b9983cd1363f218c321cfea30e619940498e9aa0e1b14714db911784d9898ea3912b749dca636f24e2bab0924b79c324ad9ecc84eca23ce9171a9693415b98f96c6789d90116104eeb5343b0653a496ffa8fe4641e2537a651f0bf5d182e0c9aac5d05b0524f07b96645f9a35971dc702ea47cacd3acf92a13895961f431ebfc91572981bf218e004fc742a944845f60034480bdc989f57171e15f6acbe956a78b8f2d05a18d6fdfa932a20066ef4ee90169c27e8b49dd98fa08a73c8240937659747b7268909aebde7b4fdea8a259eaf4206292e489cd61ebd9bb531a57b83927b17ae76594854c98339badd3aa806b6f9e6725f7f56ab0c45fa0191a7dc8f928f01a8552af48726ac0d7ac059e4dec97367a73421ab0c3b6e42eb964039b6d7f0dfff79be893575e1fdc21eb1aa6064cefd0c40f3475b4d00ea97bf58f7ba2cdf52c372c7fed8c686d0dcf49ee8256f8f126f85646b89287dc0e796e3c4d1ba27305f3a50aeaa049cc3a31a04a7b8a86aad800de12907a89e698a50f23b3b00ad53f1f8fe0e9c9c8fe9e1d56f8d4216faf126dd6d8a7e49bb1c2b8c01a3bb9c9e28efb6b1eaf2bcd4c34ced98f4966ef06266b708f3521c9148d4f1d2375718045f482b33d651bbfbc9cc8d464b8d3034caa1cc3a2a9c6d20bc48002499d54ca1ef7917535cd00078abec01b6b32b6db21c999e33837e2fdfa7bc9f158443889838b9b23bd082761dbf8e461032a9b4daef3ab29bb618cf966214861f8c8026dab28fb29b983aee22872d120f2d2dfe81711c6a62b3645cd7118285c9ac6a56c5023ab59c56bfef9af15a14d39c4ec719a033ff13f6865b7508096748f523bef34c35c3777d3ebaf867f1dda077c97a93d14e7a4ed1ec532615014ba2eed0db5fa47a2634d3672353979e31fd348614ed6592cc1ccca00e45e92e079eff24088f9881383e7279337df6228f209893548ad47ad20a2c4001b8ca94cfde529f6c3f4b779bb78499b66c2b8e7c0fa93a415fcf1c82fac6878a4181d91f9c060492a2cce6a3d51a2f98d14b5090462aaa0994adcc11b6aa73ee167f489aa7213bd4608266a265218e0159ffae1b9404146a14c0217e50155032f08177b24f479840fc8205b3cc3a73f084844caebbbe2bc23d03d0c85f71a6e879d7c8c28d82bfa9063ff775a2b78ce78691eebd3dd053543d3761ba0c375305d99c82f1e1633ecdc8c2055ee3ab2d3268efc8534c17cea7269a80577e73a89d0dd3fe3217675fca5a834afd038457de50f823c2e61343afc85be56d6d8010781deb56f739a7516567a0d85abae58336df12ddbcefb46eb32aff2df593c7aaf6f4658d06440242ca94da1edb70a420dcd56be4997796fb0a19bf5958c4cf2d282ff67c70a74aa9e79551b81a91875307743357389cde2d0f93c9aa543e5363eaa50cc7624401b7383274547dfca95748b1636cd18f4d802c46c7497fa5c73e1f4a2ee18052eb1c18c96cfb5554a1c15bfccd0ecdfddc87cd1bae4e4fee689794bcdc94613880facdb1b61b34252f722f9455d97f85b8f061b59fde3af910d75c8f3320f4318158fb8cf6f99d8a8ebd60b557d2a3349b1809bcf37491150c7ab3a48d75e7dfcc4e9959a64d1befb951cb243af5fd6f3555d14b66fdadb067d13eefd37dfd7f506a18b2031181b42d7fab7967f02a93787e7b05ab9ffc514141083ed88a26b21363520f50b29562c272936ef6333d7e1d8ce69bd3546dad6d0756e1e6a7978848625abe83e59e3c2bf270bfce3869a53d5f874d281dd4d73195a8e0924af637c0e6bca82501e8c6b786d1d0c78bf80ae756321cbd566e9a4cbb101356333b342ea9cf77be62f0f43f7daa3ac8aae1299d2c7be7d5efe61bdc3c12ddc5dcaad4027f2a9a0252236a0df178e01d7cea9be29a8f6b7dc7c1de991d989eecf08090b19d50f42c42435c7712314c3d5eb4f855eb7dbbae6549a2eab1bde23d1a21faccec148bc2b740cb890d0f47e12c21960b9e7bad3a7eac8d96a1dba2ecfd83801ede6c1132ec90553d4f57c0ce90abcdc2a4975013bd2db917234947fc69e18dc246373d0afa83a78e83a9f8073bc64e70ca89ec8fb831191b6f70775d384a94cde7d779d6d057f1f53b684f15b99daadbe83aa6a3f1532e5ca4005c16b562110692c7382582d8ca6a7874cbce5f2aaaa67411abc61a95a061be8b376dddd00a4da1eb1c2b88c702cb19908bf7ffaac23f38b38ed11727c02157ada0b749422bd374b1cdaff0717d7ed3b056350e7989010202f444568076cdf8a675a77fb0ec431add03cfebb47b35e1cf0bfbce7fce20edd4283d9bf80a9a966c602dbda1a0147ea4dab6f3b9fb7436be7af10b07fe7299f7b2da0295b9632dc9dbdc5264b280c94021c86e93cc0fb636217c6c833c8080840bef6dd5b13c99de8af1c2c23d26be218e59b147bc6687e8a12aaf9e8ca5cc033ab1ac0f90123228276def0b928d1a1430399b97ea7ae693bb6d5a978ffaaf234e677490bfc3341b3bcd1328ee04e49a4dc", 0x1000}], 0x6, &(0x7f00000063c0)=[@cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee01}}}], 0x20, 0x80}}, {{&(0x7f0000006400)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000007900)=[{&(0x7f0000006480)="990a4c01c2da74e9a6c1a6cd604f13f159d65d5056419a51d49b063f9e85dfde6c184ef4a55e4cc6731a8f3c105dbd12d5c1eb2407480d2736950f63f37a20f6b753eba388415153065d35cadfe403074895e132eee4bdda08f76a9c3d27b6931cbd55cc7da71f952854fe078517b97f7d9b9109ce0ffe09", 0x78}, {&(0x7f0000006500)="a06b340d1b1bc604b7925dea48798ac49a8fd2fc54d999c381950cb07f7d6e81d7d0eeefcbee16a065f111e03687c5520f4e74e069ed78ebc010fb31c67d56e67f9505466cfc03e0d0c5e1052874ec81c0f9b7c05de32d30cfa8715bd987b256bec59e638e2692b1f71af6a01a4da01093ae0cb1f487cacc4f0dfcae65ec372114e5eec30e5225cd044550544cc1fe367c6d4ac1fc9c6d331d4deb04d73a863fb672ece89446f1c5e7e1948505f623ad88a5cbc20e58838adc9acca9d7bb8122337f15b452ea5941e1964c4c3fe4198ee22f78c2f14bf3d4f11ea45d76169c5c261d7b9e00bc63623089570b9844576e3d05e2854b93da6883d84ab4fa2ec8d54f7c9cad73609acce06e082834027fe7add652ce2931e0e846d5ad5416ddcc76424b389dd2aa59c8dc0565399dc9e638183a0d67d5a7ffafd96922197f5c0e7ae1b119865ad7671e43fe09beb51067d5613f5550b85ffc50ff6b93e6e1a7bb59a2c1dcdd9c4262b20e34743f1a75757dc8443904784b82597254bd964e6cde1c4ddeefd72efac6c28e35def864c01d0d12f3230f940236612006b0321a293c208643b5410a7e7771e1d2c54afae526c294c4445c06326a09a39fb88fdb66d9fcfd1a20f57fabff4041be04ca9e367e91857bec7750df82057e5e66f3e456424c335432f15bfc97f597cc938e0337bc69bc75316df8d619b4b97124a8df344a960cc3e4c0f164e1cf0de224102715be05e5ac3aae82354bbed3297b2f5ec912e3db28f649867c1c83718826005b73139756264f3fe6678c8cb6c0c24041953382763ff1a08392fccc2c8089b742d35a1600f86a9904f66870ed52fc12f1f5b90741954d716b7f63b3e35d55db5f6c6b55e874be31b73505e3471be5f3bd88f74c4eb63404f28eece84b8435f31e13beed6bc2bf58c4a7e586d38a67a576baa69ca318398f0c97e30727020bd8a607aafc1a3e03122538af4719532d9ed1ac3be4873328d38299e361249a3fd3dd7b776daa09cb157d49774f4d3d6d6ab6d6589376733ffac975ad864bcfb2b3b4e38cbfb97e162e8dc92fb44283d06bc98a7673dca95ec5b3e99c01954df7299f02e1d33cd885f21975efbe69d98cfccba937f184fb618c98dc762d552075bb147758f41254eae89479510ea116e59c17f7cf39b5d1c54ffc49cabc150dcc5b849803718858c43c335efa7fa4763483e60ba8a9939da6def99ec249def67a311470dd66dd946376a34c68aa8890b26cb92ce3ada1211290db10fee2e8ac5482bda4c10f865b1e09acf13ec5516d9eb44f2d8ea09e84491a2f2cc62884dc481cb5dd9007dcf968f2ee08e7e272bbe8ec7f62604dbef61ba05487eabd7b8d934814462a0429c037a11f05f356b6e1318e3fbb98c54b99e1fa0f440b06852c36d12ad3e36b869a83d3fc00d7ce8dfb02ccdcf10413036fc020ad053a118707cdcb50f04c0b38200665f81ef33fa04eddd39078ffdfee5cce049e18b5307c6ea32b33493abd6d7072b16c1be1f58bbc145a1157c48c48454645e52c6ee8db34a5e9d3d3618cafe63fbd2615fd832d3ab6ab90cf2f0f346d19b6a1152d0862c9865b99a4108e56cd4b2813455c96725a26d9fb2753b98ed4dee9f2c0d0d932af910edfcdc439d4f061f8f8679cb4a3abed9517f026384f77fba19ca8be6b46db25e087aa438f99091e4d10979db47b80ef39116fe5d19d2a6bd47151bccb40c8964858ebb849c27c28d2d55a6a31d8608150d10b9278bb5dc52a300ced3edbebd54fa2fea803870bf2a6a1f4cadf90ecd1ac8bd1708784c0298f6adf07ffd098d4a9d387ed1e510781e5585c4c5464b4a69e778cd4d43e60f489fedc7f17d9e021ba4f07e5a9e9b9714b4d1668036ee9d85a018cdcb130ac6a3c2619a22cc43d6fc93e9e167e0a5f74d40f15e1b06bab9f6e734a473d350dacf3a7c21793d0256355ad71a36be8ba60322b90d397e2eec55205d195958ab75ae4048a73b2b7e8153a51541b660d79bd736243ce72ef184ba806ae06790b9a405d42c223b2c26f07722e1b62af9014b42f00962862e27a5f0ceb3054d62c3fe818ebd3e17caaf6b204a9efd5c26501a00e21d9fb455cf33450743b995a6ad40d03e3d43c52de2de33ef4921a19d4d966711c6ed23fac799f247c319e4788d02521d3671d4fbbd809878a1cd94d81edb3189852c279bab984540b33d5ac156da7d66c431ac8e2f7fdbae4a291fe957602138ff13235f80a6e6175e7cc3143cab4818f591ced30a4ba032ee5f8634bb480673cfee72f28c60a4cba42b119edcb7f21d9492668b83876f05ec055805e3c72a4119a02b7e1aef8f3ceac5191c7b6befe4be1c14f06851af89b408e0306fc69c6ffe67fff6ecb43279216e064ac54bea29fd12ca9a794011f7c673ab6604a576773404551ec791a97425d2a971caaca56452e9f8a2b68d6f8b0fcbeaaa47867fc5e486a53d6d0cfad2bfda8649743ea93cef1b32df4ed01131d38746d4bcbcc056b7f88e0fecbc2fb2f1bcb6a91d07b021adad9181ad15927cfc0c03850a71d35b9701b98d0aa111bf66fed7677edc22569551e03415101dd40c47d97293d407d30d674f3296348cec5771a4c5b5ba0321f7d2f0d21cd5a115f1eeb533c357155755a38acda015146d49f6d7c237544eb4643fdb8c7fd70ddd4a86002216fb0d87d269e8b1d5ed7ee7c339494e152c5708091b2db98026847fab4563a4b39010c15b3e52f39dee01b8578f1bfaf71398fe0a113443ddd9581c5a994ebc32028cdb78182975547037569c4a570c30c3dff2a8ad316764fcfc809e5c812517cdcf6e718c68179f1badbc946914862fe5c6c86d1fa213cec0b21014764b6dd55b64dd69d64a245409607c124b27f078c290fc78f324b4f053df11692ca5b7e46ff7c99423cbb9279de45572397ed7738bf09099aecc68f82afbb0acc32537718fbe3ff659a7af675a17618d91fe86c364123516d2952200da23d61e091ef26c2ccbca0f1bcdb87c1f2d642fe657424615744bd92aaf485e389db62aeecee6ad78ddd3e09bdb7714c517c11a9557570c8451c1b98ef80831a19bb900d32290ae3665461b3c7abe60763de2fc7e69a05ffba916631a115d3f3e10e007ddc3d7330d8758ec7d13b0a472a4fc376c382b643410217deea3ad0cc376135bf4404ac36b04e992c42b4c46f439e755a7790c90f7b28d8a868cc0793ea561c58499845cbc55b5205d168ed7f139d37e1ab100a9cf8561f02470b37475f136ffcda12c95ff39c4f0de11ec56a9c7f4e4129e706489c44ab39cd682469728e4116cd3b65df341e5173eec3965e2fdab8b230ca9762abcad5b2928f68f789bfae8d476ab3393caca3da4e281ee308d2592c4a23bfa2c5e7977a09d7ade1e8c1960336b61b429710b4d92b7a9f96fe7f4a502464db2cfcf7325c0cfe08732991408f417eb2b04576f3908b18e4cbe8df33ab5a8e84ab1585d191c1e8794018c3fc1a1b22d0765530ec9ede7b17243ab6ad6df900b2b81b1fb98b8098387b25fef82379e4f1661cbf29b5690799a8bbb395ca54255771a631df4c08dbb62cb63e52cc95c3caca1c90dbbdd8a1aae7ae6fcf0d18f90a7e5ba5fd65490e468200d947cbcd6f31326ab0d8a19a9d523abdc9d7146c9d1f065e1d272223586e9432a66517547a63e23f835d71ba465adf453cd4d35f2f3ac14bbbf04c952ab0a27f31211d690e1f4eb082683955d8cffeff4ae3600d4ff6dd48033cd5259e2d5cd41f642388c969bf1f5f29d4369c518b2640e52ef0adefdb4ba1aaa19fbafd4e51084c7da39c7dd710fb5ccdd60274aae9426b5f77adb5f4d397e41de3ab2586094ac0c59a81505168b6778e6c48a96b175bc2628e3f8427bb13c39693635fa0009c649a5c63d7409559287642f2e48af737a74d99e9ac378338581835ed22b3644bc0c7f45f08bdc608e4a6f7efe21a0be50539669ac3855c82c27dd2ac958cf5afecab4b8c29e3f4e96dcea1545670d2c3500a4e84b37335818ec2db6b69899cf3891c1918d3cf55f84a6ad6e7c7b232a700dda81d8adf0a172cc63dc1da443661941d6ca5fbff1d9ec1c1b582a9c440b613cd2e73aa17a4cf85af448c202f126e47be73d7afe97997101ef4eb08cadc22618e0a8e120604ca98f63c5b06c41a6857b269568d24f99d14986436192d280abb7c473afe28a2e5adef5d3efafd6998d0b3ca2be7f1a176713246018e2e5962760be859231639ff3080661a29922dae5ef818aa821c376fedee1032f95d1f4bec89ff3890d337c7b534754b650e1c37c7e9cc4c99f2ae77c9e3311ab80b9297f71a55790a1e429799a58e702ffbde08ff205b371118b54821a0f9ead0cf43a31de6acfe444502ad74ed66dacf7c6fcbe7c9ee40ebd0335e37f8c97513253d32bddd4270897a01344452458d147df0857496a072e8a76fe6d40468e8a989b3066fad17762356a41a030ad0fc05a99d8782e03d5fa614c87d701f4d97c7ddf8ea83b35c92de0d07a1b93476627f3a70413ab5096e715189613b523d963a4b00ee3cb7e2a318d8dfe9eaaac214d9995e4a27248b6689afab0a1e02c82c3df9484df8dd02c0082ff4dd4c97d0735e634ebe86ffd45118437562d4661db8a650d212dc4fa60785cdbde2118ca90d814b666bc890257c0fa0ce94958c37e9f54f5088973e9781406d899295ac8f57a1c155525c929ba6cb73d4b92606ed38dc4756a06a8e5fe05ecc2c78deb504588ad16ca73d794a833ed99f2b23b7df4b64c7b4171881b29b776e088b766de5ec287770fef9b23c4a2d4f83e5db3c4127a82063f5bedb3063f8442a3164a7b0fd759f38a0b2ba394059183f11e8a55f266dd86ded91d164f1d2db96563f4c8bb62163d081069ccb30623306510b901b3d15b2e2711fe2d3944d7c68b528791418db71ac950ca85cdda23576a0aa9f2d6cd8195071185598357e2e7e30ba638fe85b4f678f0b6cb5b1aa82c605fb4082d93e3a86af4c918de6571c0d0d5911c2237cc0b82ac546a5332b7f6bd27b76c3c23aa95621e80fcb24a5b1857c83f2cb31e75fe7203b61697e72e1e4fac5dbae1e85fa24cc5e45ecbfbc0473043a8c2117ca09f5b4998339e2fc963c1967efe7feb5aa1c59313db1da5676a155609803445981fe7b00e4cd4c8d25e6d26ea7879ef7bb0f68184f2c476870e71d8a6aac7122862863151a12427cad8c4bd25d9181cdb84255875eb87d855cc5b4fe4f7f2fa3dca355b780114fc1053acd4f09c29fdc5e968e5e480c55aa63684ba4501e30bb05385707bf55bb3bbd633dba9061bea18f771c42231f803447507aab7b9b321676a1d25256b53e4a46abef5007fb56c6543aea4268fe93bfb223803972a30214af7cc06556ffff38e17e496097a8fc6f7daef5e1a7e9be0c0b68557c3e8b9d5b99f47d46737654d350bca47e4b1f082b2fade0fd3d2271ad66be3091964f0c81d4b8aacb329604033167476b5845c7679d1cd204408bbfafba794da610fc68491b16e4e956bac17a250a36e19782f894a90996b997660db916b87b7efeff28604c0e6cfab4f5f7ba3c2a0d2477a289b9792793f175fb710fa8266b00d7b61b523a6080cfd1e1f6572487e86cdd9be3f657f3bcd6691b88beca90b13a35f5ea174c5c8f419b933aa42e484487e678c6501efaf73238234d0065efc7e56a8ed65dd92f29d67b6479987459391f0ba1080b836433a4b747871ae09dea3cdbbe1cba96d541870e9861c7c6042f1155afb8b5f7219ba338ea4d", 0x1000}, {&(0x7f0000007500)="a6c275af836e1c12bb9d3d858cf382ab13f5a9aa64b7ab0a0dd3dea9f483528176080e347a5892bfb94cb3efccf4746cba71146f8e460784053c5919ee21113fc0582d7bc1871fc7e118a973b73e1b8cf4c36ad337129f3dfef4b57b5f7300a18d29acc969587bca44f72215d8ef9e117e562c1c852284196cb1e674b2fc2e41649a458e6d92af2dc0162b57f8084bc28e5ae8d56bc66e", 0x97}, {&(0x7f00000075c0)="81ac386df879e8816f22f465f5d2f4a1a4b6a8aea49cff58c1d3eca626b575df6de2fcaa8c4bd3ae34589b4d79811012447ba0fb8cb1c12577ec5c8e9f2fb7888ef9e9d55d2858899e3b44e5aed91bb3b506098d55aeedf4fe7b9a7b50253489936343c53011a8802f856abeed4e285af9f4260adedf513723335012e53855debdbebad444fd578f90dcc0da587e5d63881df33b2e42d17d0364face927d02214b46af7f18f5a52c1416843dc2c8e95c3c5851d0a05aa058002aeb3f36", 0xbd}, {&(0x7f0000007680)="026780c2a77108324e072ed0071d83726a0a4fb02fa520d948b5f499055cc8821a94c23fadcea315a6399d59db2fe49e48eb77e46375acbde00df60baf813200a2ea07b28199689e57511f87dc537de87742265e772f5b7f0368b69e6e8b8b8b07b170bf67201bfd58ea3439b493b02ad5649d39d49a6f4b51bd1114740c73860db68a4922f8b299db", 0x89}, {&(0x7f0000007740)="ef1fff886985500e5d4695352482319cd86ab085cf3ec5d31d06e822e9e992e7d8e066d01dcd08710e74998ff8f6d689b9e480617e46f48cfaa6c4bcc5fa902a420f0ed04236562e2d7d1c623099e7443e8b2214aa1038d4d8197ead3b201f91a4d30411b9ad0e75834706fbdccdcb7b5e5d6b224789d3f267471d150783980751f2b2e60f8e5b026ebb66b2ddb2bedadb14f1611c5f", 0x96}, {&(0x7f0000007800)="a6ae2e0dc20d8759bfa97b72e78937ab683fb7853480a5dde16b3db7370d882324ceda38a19516d2909adb745928233e76", 0x31}, {&(0x7f0000007840)="8e64bec5da473363d7b627ff055808496cc03268b383aea1dfa00ed1e6774819dd0cb9522044969f8760cb7a3935b4dabf50c04235229ede05bed60bb0529a6e4759ad2a80dd8034a2f2498845c7045819e7276ecc0c543b78ed6a701d5faaf6d68fb99c4dc0d3b21a7895bce123cde793", 0x71}, {&(0x7f00000078c0)="c9f6c5dc1e42813ebccb3c", 0xb}], 0x9, &(0x7f0000007a00)=[@cred={{0x1c, 0x1, 0x2, {r3}}}, @cred={{0x1c, 0x1, 0x2, {r3, r6, 0xffffffffffffffff}}}, @rights={{0x20, 0x1, 0x1, [r2, r0, r2, r1]}}, @rights={{0x2c, 0x1, 0x1, [0xffffffffffffffff, r2, r0, 0xffffffffffffffff, 0xffffffffffffffff, r2, r2]}}, @cred={{0x1c, 0x1, 0x2, {r3, r5}}}, @rights={{0x14, 0x1, 0x1, [r0]}}, @cred={{0x1c, 0x1, 0x2, {r4, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r4, r5, 0xffffffffffffffff}}}], 0x108}}, {{&(0x7f0000007b40)=@abs={0x1, 0x0, 0x4e23}, 0x6e, &(0x7f0000007cc0)=[{&(0x7f0000007bc0)="a768cbbfbfb9799d2798e382b370d7a4f97aac04f4e2cb6cbe709ab9b39acd04e90893b203c60b3c5cf2f230e98e9ab5dd325d07c94cc27b98340a5747ec0cd00d0410b513cf1355fb42906f4d5bf3588cc4e882c988", 0x56}, {&(0x7f0000007c40)="541bc02f07b0515b4e82633cc9001a6993f8ef2e336a1268b477b9725dc2e8b25cfac37c52dcfd6b73b31903b6634b06bb71f0c89c287b876147c4d7d974dca4574de11d8e8126a744ed0870fb584b70af7b1d2cc0512e47b67a54bd49d062", 0x5f}], 0x2, &(0x7f0000007d00)=[@cred={{0x1c, 0x1, 0x2, {0x0, r6, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r4, r6, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r3, r5}}}], 0x80}}, {{0x0, 0x0, &(0x7f0000008100)=[{&(0x7f0000007d80)="fb76ea769ba395d6b15f2c8ea822b83a9e2029a24476a5e7ab2f2eac6a6be60d53828a71a8559b2b37ab983c666f868481123427d322f659e1cb78ac9901b76b8dd75f0d594742d206f1406e66b2231cd83a667a968dcb15147b25cf1118ac465916f417022d2b3b0ccff357c780994b497968ac6f427f404d3288a6ce5592e104d4ca24273a216c432ebf4fb69d18de5aed85167d3348cd1b2e43edfee4eeb252f0a749ba32b8ba42455af1734a2c97692a2eb5b5bc01c96378b10c103f7ddfeba03e61", 0xc4}, {&(0x7f0000007e80)="45ad8de36729a10cd6b48e0b0f0fb205a41e3a569cbdd1c571a333fd268886345dcd3354e4ceea3ebc1e5135dfb20f2648100d2d49227486680a2da2d9c3ba339f27764a010f9a21b225e8cae79a408379c446a07c0044e21d62fadc37b2b8dfdc8fcdab2c27c8cd7108d3951ded4e0512a53d9806856e6ba9cdf5b86b1f37fda6636c755b1d3589e1a6e022edaaefd532027f833f", 0x95}, {&(0x7f0000007f40)="81b19a1001504839980eee69451e2846e151ffeeffa50f5fa3feca563d6ec186489394c42e6a5a3a3fff6d7d2841f67d3d0b02ee3124f4055abf6a61881edb47c71e4fd98aa45fa51ddc306f396ca61ddc2cecdbf03aa6afd0614d76d35c24f28f77a233e35cd8a1efee2b7396951bb43ce871d4c51809b352ae2a669944932843fd858863d5a380a11c9ab566f5f38945fc628206162042", 0x98}, {&(0x7f0000008000)="29cb7f772025312ae8a20c0e3cc46d7e78e8368edc68d7c32c9136c3763d2430ee52bc53648281611ad141fad41d048b5adbcde1f95d5f3f65b06a64c24dc1f6db3bd328ed1cf5fd4747fdaa76d6d8b372ae47f2425fe9bfe60c034013bc415a819c4e87f984579156be903fc686912e0d2c6152c237a2da3d5477f62fe5c73c9ad3b004369a64dc719654413a974f9707bcacf545085ac8bb985061b31c0b781a1629f8620b156b0f487866c60b4cffd9c83b42f71f0184c970186f75ba990200805119ec77f22004dd1e65e2bb9f54a53e54f66f3c09e04fa326db", 0xdc}], 0x4, &(0x7f0000008140)=[@cred={{0x1c, 0x1, 0x2, {r4, r6, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r9, r5, 0xee01}}}, @rights={{0x10}}, @cred={{0x1c, 0x1, 0x2, {r3, r6, 0xffffffffffffffff}}}], 0x70, 0x80}}, {{&(0x7f00000081c0)=@abs={0x0, 0x0, 0x4e23}, 0x6e, &(0x7f0000008340)=[{&(0x7f0000008240)="470737cc8d755bbe577ef5e4ed9fa6c8c1d0cf5c4ea5bb62a38a50ed79b6b5af2a22626570f19df5972f85f5cb5f43d43ee1447a7df4bef6054708c084659ca72f4235b591e222b470a4d17553deda1857b4e77428d80df92c21e106de3f4e663523f85f224e514f9787989768c7a276b3cba1cc6be337d8963aac261e06d7cc2c4d317a786c23f5e5cffad913f0b06c40876c07b9ec77c982f97ba5c191ae1f4aa4da95ed1f6a4b5b804a1d2e0bf10c19b7d4cf33c0216c990394e571cdaab66b9a85a204b2586f2c0e33", 0xcb}], 0x1, &(0x7f0000008380)=[@rights={{0x1c, 0x1, 0x1, [r2, r10, r2]}}], 0x20, 0x4000}}, {{&(0x7f00000083c0)=@abs={0x1, 0x0, 0x4e21}, 0x6e, &(0x7f0000009700)=[{&(0x7f0000008440)="f008f5dd6a9fbf0deb81d4da26bdf1d9b25e1f6b8b4c689d1d5a8d1cc81426234a0b48429a1836569b3241be6dec6d75ab525e6fc3acac8ad4e496be4adf68a9ba40a3f167d86f2663f4f8dccb65a045fea840ae77f00ec733fa5dc2879b7657e8e728b6c0025c8cbb8015cf2fe8df70f4bcd418548a2ef5f7822fd0045c4a5155d5cc673ac0ca5bf7a6f85c00381d808ddaabed5194c42d9a4a932b42277621be62faa1957c2457a2debfa08a7e6c67c5d22776b53595b1c719cc16166170d1fab111cb5a48bf90cfd6019516d8a235375dad", 0xd3}, {&(0x7f0000008540)="653a34365a6c464bc6b533265a6713c660363b7a2a5657742f89d77f6379b35023376a4ccc0ba58ce4fbad08949ed8d9c29a88a14b0b3fe66d46cd74243b0236796b67e81655c97f8ba53e13a55c0a462a71c41a257238d8c588929d4fbda11d0920404f7f7d46f1eb511238bb5c19df040ad61a88e8549cf4629352dc9231d0134141d5fb2e9e6131e74c3574ce24d2af664629df4a622d643828b844775c7b6500160b6e8f86e498099357214509f27fc6247d78fb2f473088339a50495bf8b7cfa3ff677dc29ddf4cc3c3903dcf744935d26cbfe6091d40cdd6a5d62379a276517983e49de0d8e5a18abd314ac4", 0xef}, {&(0x7f0000008640)="3960b134ddb8d8249d9c7e7daf922155ca7cfc209beda5cf8369d7250547e54a0a2e106e5df7b7063c9f2e14548b90e230b61fece70594792eb11551991d9b7fb8b56afa169799fdb381d7c75ffa465b9e1d86bcd9ea9cd9e1521628ed5aa6a57a9abe1e3a9d45b4f13a25418749a47350af66e33cf8bd4fc7e14a5712f5f3bde595ddd0934628e8ab42d46adfff43e40434d19c9c0870e46a88304731f1defec315af22ec8f3be3466ca1001a8c6fcdc0d346a85f8850af4370fbde8f92d2003357cadc7bf35a435cf2c6e5f2b67d54aabb3cc2da98f1eac9d78cb722cb1f1bc49bdc32e262e2ddf50372072ddd3b7d302f2e4682807529b4fe5c9625d62e32d18f531386bf49d95d5fd685d5f3fc7caf6830b691270a2ef96903b83e79ed18fcc2d8fd59b86e32a92f5629b9ca7580959d595396d31976693741c39179e232c0ef57a96a546c594b5628ca862ecf639985cb805b876994aec44d31101e8a1531a0b29d947b20d3368eec3a185395dadcd5070569015278db4ca23808e87fdfed7b05bd586b8295ce2a5bfea8362d5d09b1a0835d7742d5f03fe1ac44a19acb864f3f11830a4e3d598ec4327daa5c45e7da6a78a277357af7ebc8ac4645fd34e8e6467f802411e34c922e9cfb99b21250b2d511cd66456c7c174edfc4f6c1546f33a4905fcefa843da569be88350d60f0c0ca7a94134a2f1369d196d9dd4382a70f7a9173e6b89ee1f975b8128ad6453d3f1ed9f81ec9dc307fa4e07855b920f9276460fdc2b03ec94ab35fb0f5489018633f9a9a390531f6396d259d03ff47ec54846370d6a1ef1b405abb6607ca5cb6a58d37824ff6f6177da0edd10b5b08c34abaa93934ab06302d42fcc3327454922dc04e8a3a4ba43c0db4f282fd9e9f08ec4365fc21088f180960f6c499cab0960d67edc2ece5b046e47cf41b43b5e51273acb71af774f557c70640a68afd38e80d0c19fc1fe330906fb7b18c2350d6e0a5f9a516e72b27b41c2d50f5ebb84ee34ec982773e74d8c597ea4092ef5697a72d836958f54ff0af51822ddf07dbc58840d04098cfb3650d6dec9baf6634d175a0e3b609cff7909bb46aec2249aa0455edb7d919c495fc0fddd7054be58b099d081b99cd526225cef7649796f5a964dc35ea3ef9efad79155501d62bcf110d2bf49622fd7d17fa3568419b03d9802f771b9a06e3e5ff56ee78d7809295399beb0ab913f7cc295455ee0376e0cd870a71f4004bd14462a57034e58075b14a70166a2b1a07f96df4d392366ebdf77d333d613fda39169771c65acdf16fa7cc8b9feeb7d2412e4631066eae0222eba74e6bdef4083f848c23baa4f9175f30a02fd99baaa1941cf319bf7daacf0422fe2cabe8ab62a6696006735655363875cba6cae80b21eefad3430320959277bf4c5f4948b1e1e09bcb3f2f90fdc7cf7ad36114c852534ae70c0cf4d19f6a450bac1babe019852d3f4aef7c6461b031207e92ea9c6c38fe2ba7e28f3d79d39efdf7dbbc721ae90012bfdf3d1dd7288e5749189781c6490df473ea1f7f13890b383001cdf56f2a3647ad2a6406ff1a78475b283ff1ac0eed9b84e14f6431e90a68da77db09e4b131c0a859d3facf7bcb99ee8032d555fe0d82822bef24949576be4163d45605c6530c2a632e40b8d471bd3f6a87c1c4b0b0db05b3b1211634f93a0382312c263a2238255daea75c800d7df46a9925882d3cea8abd50fda1bd49d385325d99e935dbdfc02f26736f2ccda2c20619ef4ea783aef87910d44d19cf032562b332d6ce0ed4aa1a277f14b1d62ca55fd3f500cbd95590220736d1a451cdb3bc6e21e14e722d189d7bab832c8e95e28cc1a4862f6d5966ff0ca9b4569cea6e61e1b31fe47672036c79797e0bc2f4959f093fdb72df4f026e30c04ab730691dd25967c15f19c05981580b1ac3ef3f09a5a2b1ac79fa2dd749dfcef32164234bb115f6cbc6d2839893be27697ccf96727780f7a3b10494e677503891667175fecbb61c144eedbafb9881f0fabf6588cc6d576a7802436a075fcc6fdbbbadf2ecb78522c8a796b031f42b88a1a8c933d6ee8dea401e064b8c7b4bb11e3c4e9588655f99715a719f8ffbf1367d166e93449c9d88b4f02e48d494c25de12c234ad6006858c987bd1b20bfd3dec5bdc09b8ec5506b0418fd5405b189d681379c7d7b103ec7fd3f5065794bc7014d30d5f2fc8c5819c2766a6547d71a257b884f8cc28d746c072d99cd60d95662e0eccb6790384aecd50dde4ab5068c0e0cf5ca0f54dee8ba21bc0daa8f32ff2e871a3104cb329aabba0c4626cd4e52f77660ce9b5bb586bb36d9b8451d53ad407f2441e3b5364d822fc9f1796527d037fa077ac6722eea9ae0a3c4c3732494286120bdba8e954c7c66214361a51a18957a06f8958a9a63c795ae49f1956c66cb8a5403081b2ca93130b5431a9181585c72ac10a4abb1f8719c72dfafa560751a2de9d63c15ba3bb4987edcf4ca7e2086ccf4d5b0f618f2a24e07fd59ab02d57b3153e962b55dc12c097a62c3ced2e0ea12bff7c41ec0f276a5275022b1b459e5568fb590e81435c7f53a2adb22ca4868be96045a245926424f9d562b956d4b8952482b1193db7dc2148bdb924770ada37cb0eff1c9446a20bae3e18965dd86a1bfe322706f27d4e986fa119ecd9668d0d2d9e7e5c03a537bfa2d395ec348fad62405f6b1281c5a0ceba3331e442afd62971c0fb04496e96c66a63cf5e6dea14f43bdfc0f667b6ec50cbff3c731a3e4dfdf9189db8de7fd0ddb7fe29de2fd2633586946013b2c4b0f408fdc213a753b59edee774eb9302374f80239dbad4fa8c1d84142821b2ba89b23a006c526f69807809e5fd407ccd750038639d78c4b6f1a45317da11e4e826aa5927105b97871b137de46d61a3a7f8ad49f2a93b7f95b8292dcf2cdb2acaa3295ab72dc9dfd657722a4253df647ada62266d9b9d61300c67f2d2a48f6c2e1da327de52abf58cb2db675920111fb34464ad2a6fc511b4e73613c4b7eb61aaf6f38b738e73aefb307d95c609d80aeefd228d65f8be99497b501a7cbe371195315f265938036dd6e755e7391b83e9d886139847193f89a8cdd4e6383a0e98e83bf7a45f1f2fc73edad450ecc50ee4f368f811a6a82f931d26cf91195540d2793b81a22a5afeb35f59c932705a1b7c1e781e9f9d5fa75e7b6bb7e2ed12d8407b3523ee97c9f027e87a3c85208b4b105c5883ee86379eaba1ef9283e59ed6a54483d7146e7dff2a4e1e9203dc6073b82161719ef8307721f43fb053ae76102c36cb04017b9bc150f56ce2ea3809d7faa1d46a4745a173c13189360f9e13bf832fc4625fe5aa438f79accbb55de2160e96530e130d6912d95d4fd809fd8475ed3273dc1ac9575ce3205613259006a3ea991549688daea9b93d50a978eec49b1b675e40c52b1815531fd564c0502f03d985ba2b2489029ee8167c0bd89852494bc609ced1516b803a63e088e3e544de454af3ea35cce56cf8214ac91687e3c78c338bc8a93b837196201a445e1f6568e60a38de6538169f8e50b19a289d82ebe11ced3251c181924724a4dd78edf16c46e741af87adf627ea4ceca60c5626f301fcbd24832b0951dcf9afd8bb11ae6d4676224bd53e3cadc18d83953673f0233879157a7f610be7cad7930ea7d88a192263764b279f3670bbdd12e0f71259cc8274541e7ab3a810dd51b9931c0cde907cf8200cc5fd6e7f0a711e0fd38f6fc6fe2530ac2533690099e460bdcc776f282de5dfcddeb404d6e29b5973d2527b316e9db2600096578e8010af7aebccce3915f5d974be60c889556f87c83764ef9b21e728e2098735082baaf1e364f8a9720c802aa6e941268d452199f45dabd0b6794c3e8f6db430945828aee09f105feea67956093e91fa2ccb118034c10c5686fe27695f379d4b746665a6a8d334217cb4773085797af10b3b05098b05a2ec8a5472eee5fe96cdbd56b5213d82c1551be67af4decb3b5aac6ef3b930e2006270d68207c476203cfffd1b56b605808f6c41b097b507cf085a63268f85fbcaa0da94578764e2e18f6e592609a811826d6a28eb0842a4b24243c2bce43d88aad383d500cf447b33ec88cd190321ec43f327f86fa6df9465aa8b04d9a4c45c32879b5ee8e9d1496a7565dc4f7e33abec2a65e342f809a5b20a28c00946bf090f7c4158ea58396a68a447f2f22b3d16297aee2174f54bef72aa0e556b730be0f97e96b4806c4570409e09bd9b34cf84644fcb4f8e85a71d0c6de9d1a8c217cf63ab7b5604ca2b3456b74d4c90864094f010200ca8529de83a3c339ae2717f31358d6d32ce4601e6e50b3b6f1b9ded705a8121c5486d14a2e403c5eea18e37b57688d14f62439e1a5ebce0cc894817504779588b1f606cfc8cbfbb50beb57f53887c70892ee2e98689514f8f07f3f873fd36f7b1d5f71f02e91969ab35dc05fe48856f66e60f9b7dae042f3d62959f5f270829f9f74ca482b7d1be1640ae3995dc772b911dda2630a04d755297e85c9cd65ba9fed53411707feb3132ef30754605b8bcea375172bd4dc2e77a89ac52bbe79ff94988e72e26d8e7aec911020ebc85f0791362211a8f587333f2c6193f734011ac8c944e3af9852780a5071f1721ff95be5901fe963d648aedfa8123bde49d9cb1712f79888c0ce01cbcc468cdb5748b3207359a5c151d4257ee9cc5e677eaa789fd6d753f530d72405d090feb895110fa6e03d7cad8e3786bee8d821d4815653f5aa57b7230047b78024cce1923a6de777fe47e980a48012620f75798b374f530e7b9b46c6477949c54989309ab50778b6a90cbf5a5d096602ecc736ed743546bda5bb4cad6c84a7322f6a387afb1a7eeac01bbb2c4bfb7ded97d06b3ca2548b8ec38c2d1f5490ab09be22de4274db0e001763a27030fde34674c875207a37fd197f289f6250f8195a46ec724fd9f6b74fa18cd62521361b4b7cc42ee52bfbde8e8ca31c58039d3781da78a6571fadb7095767587a57eeaf3d0233acaf34c5578f1510b9458479d5ae731303ccd60eef59bc968e3e12054d30ed8017ee73062acb6fc8179cebf2b97b35198a6dff6716c1a69a2f57fe1172fdb91d8944fd91c6088f3bae4c736cd4a04b0a55170f390bb3414c0bfd8e751b818889d530933c6eb2fd27256d2368b9c18a27698ce40e227ff2dcb9b416fa303b7e28603642c824cd3d7b57ed9a9968501acf438d56024a3de1de9cf1983a906fc7b0df3df291a230e1c6eb84f846b089d8eec5a3c3695d1e7cf12496b468889b89fda91d15e3572b608853d001c79897a3342759a9d12e6c7a79e07a4293ecf083a0e165c05b9656c46bd402a33c177f55605d8cac9eb4ef6b2e07265ddaf361005f4174bcda4ca45d987e60780b02de300a880c0bc094fb052c3c8e22b2f07d49393fc146d5a287e82e78fbcd8cf1df400e6be04d7a9760ed81700d241f31e6dcc910bf6e192fdf058a4ac7aa761cc26eaa8fa83d58775628799e8495214b7cb387e89ef1db501eabdad5cbcb8f40e7c714642d79475de10ba2e9774148e6a38ee88fe7375daef85645d1ea8b3a0001cb7ce1f647526f050ab928e524215850293ca4c4e3a9f7f49499019d3950bbb53ca9421aaad909a177ca4690040ba8bfff527afdf31778d8b9c2b86c0ffe2ed224b9985f258e9a01381fe3b1da68c9845ed9229c8fef8fd327c7264cfeeac61b5abc1bdd7f3f45567280981ffd2129af50c09c00e0382c777b984d5bb708e3b184dfa2b64fcbda6407", 0x1000}, {&(0x7f0000009640)="5e5c5176a6a62b034409da1324a58e8c873b0d3769bb8818b8fd244314bf549a32e8d86ea36ff497db2b08240409da74aa5ba10c5e8321306d262387465e04a1b659b31ff61a5ac46e95e88c67e2be23fe6bd0bf90b654bface0350db43918e7d018c02e67df02d733a312cdaa233fd56afbc41cbc19558e4f4d1e5492859dd0fcfe38b91e55c0de870d5bff40b9069c51f9c2a45a7e53b4feee5d39419f96901254d1433e11dad556188b06c89d38", 0xaf}], 0x4, &(0x7f00000097c0)=[@rights={{0x24, 0x1, 0x1, [r11, r0, r2, r2, r1]}}, @cred={{0x1c, 0x1, 0x2, {r3, r16, 0xffffffffffffffff}}}, @rights={{0x24, 0x1, 0x1, [r21, r1, r2, r1, 0xffffffffffffffff]}}, @rights={{0x14, 0x1, 0x1, [r1]}}, @rights={{0x18, 0x1, 0x1, [r0, r2]}}, @rights={{0x2c, 0x1, 0x1, [r0, r2, r2, r0, r22, r1, r2]}}, @rights={{0x30, 0x1, 0x1, [r1, r0, r1, r1, r23, r2, r2, r0]}}, @rights={{0x14, 0x1, 0x1, [r2]}}], 0x118, 0x4000000}}], 0x9, 0xc000) 00:28:47 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0xa}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:47 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x40000) getpid() getpid() getuid() getuid() 00:28:47 executing program 0: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) r2 = openat$smackfs_ipv6host(0xffffffffffffff9c, &(0x7f0000000040), 0x2, 0x0) write$smackfs_ipv6host(r2, &(0x7f0000000100)=@l1={{0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x7f}, 0x20, '@Mm+[%/'}, 0xa0) (async) bpf$MAP_UPDATE_ELEM_TAIL_CALL(0x2, &(0x7f0000000640)={{r0, 0xffffffffffffffff}, &(0x7f00000004c0), &(0x7f0000000540)}, 0x20) (async) r4 = getpid() (async) r5 = getuid() sendmmsg$unix(r0, &(0x7f0000001980)=[{{&(0x7f0000000040)=@abs={0x1, 0x0, 0x4e24}, 0x6e, &(0x7f00000005c0)=[{&(0x7f00000000c0)="b717bb341df74e9d6777781b0feb8659a05825a132cee03c2cb9ce8db16b232db8a87b160706c91b2c4435ec49d70c3aeef358b9b6333f3e7b9d2f4c220bd2c81c0975017164fa40b7625daf60ae8de69a11aed22171090c5b4ac689855c7c7a352be8660316bf97aaa49ea70c43afdcfbffb07af9abe5c6dbc752a91a044b27d1cf2101722387de490d2e6b4c285af79358ab2e47784df3785a8c23b51bd5ad898d19e5aa4376992b8f3816f3674bf2a06f318e9ef694addb309a9965d438a454701075ce2d2e888186b0a8263a4bae72028c3a2248023cf98ed081d9b5590b47cda6a183", 0xe5}, {&(0x7f00000001c0)="1c7eb0d6fdef006522ca793a7adbc3467ee2f1f2a63d8010bea062491aa13d51fec04b35453dce6785ba201f7259a75f41633a43e6e77afcf4a92b3bdaf2d72f5466a758f496ea94ef0448581609d14d122a8c0e8be313d798e2e4fc4791", 0x5e}, {&(0x7f0000000240)="bb7764035e5e3f0b6fc7d0a948caeb1aa98600d029a3f0dbb1e873df0db00926e5de2032334c5cb2d6d20b7e6f9ca07688dbac381cdc8d1356dde70387d838f4de0101c03941561e136b373ba9bb88bba45b4ba5147873767295e43e960adb177add5a90dff50cc8801f5db8411042a8bb87e6cccc02b2d00cb35b2d0fa893258748ed37549103afe60e23645acb7672aaa6fbd706113e4b4428b8a4a06b03bb5925b85da12e58e4ae50650c8e393c4bc048f5170d4aebc4b315c4d854", 0xbd}, {&(0x7f0000000300)="f65351bf52d2313687035d03caed565df5f2062b6748c84887baeddc0fc88327af9c4b73f0f9153fcb1c8b421580aebccbd2fd81d29c990f2b50690a24be204b450d31c868565183c65c192d84004a95569fde", 0x53}, {&(0x7f0000000380)="1657c6a3e4c7f7e5316393c595ec5782434800531a41c04adcc3f4738f8c566105042decda11488bc8e5ab7425844557cb806b0f941b43c553d5242b230d92015d5f94", 0x43}, {&(0x7f0000000400)="35ea73eb1953bbf932fa13842445b75647a26526dd0462b70d5a2b37b0e6227378d8bae64762932a7ca8feb18087ac0613d5835a783162f2827c63d2d1e88046ec381483a016f22432fda496f7737030f294e35e477cdbd4324fb8e6fd93faa9822764cb83d73c5c62ffea5bf641b4bb6c01072bed8fb925f8f627a147fdd28b31536edd0e9e18417615bacad5c1dd36df4cc212f8317601ccfb40e3", 0x9c}], 0x6, &(0x7f0000001a00)=ANY=[@ANYBLOB="1c000000000000000100000001000000", @ANYRES32=r1, @ANYRES32=r2, @ANYRES32=r1, @ANYBLOB="0000000020000000000000000100000001000000", @ANYRES32=r0, @ANYRES32=r1, @ANYRES32=r1, @ANYRES32=r3, @ANYBLOB="1c000000000000000100c7bd011c6e92e6e9b19c31b828e729fa2ea9edadbfc7154bc4c2794caa137c72e26567a79792d8c6d4b0fda86ee15dad83a07864864d872c8a283ca088648f", @ANYRES32=r4, @ANYRES32=r5, @ANYRES32=0xee01, @ANYBLOB='\x00\x00\x00\x00'], 0x60, 0x14000004}}, {{&(0x7f0000000700)=@abs={0x1, 0x0, 0x4e20}, 0x6e, &(0x7f0000001940)=[{&(0x7f0000000780)="a10f5a30325487d250c4bc57e0648929ffa2eb02e0f9a4f78de17ffb92216c6526944d71755f215f3981", 0x2a}, {&(0x7f00000007c0)="1651d4eec4090a2a3c4382d70c12a652ece34aab27ab47ee0511e6b28c68d3095835dd0e1ed1c6afab230885cab82a8b7df985692fdedc2885ba20fd5b01f77719e8833e339d2d7b2926096b71092281c7f89d084d87d33231693535076f277e542edf43ad1a993edb58923cb21a7dc3c30cde45e1d12b13a1e7d03dd04b649ef409f0aa7a76911817c0466adc08", 0x8e}, {&(0x7f00000008c0)="81dca8cecc10d73455ab3c5a3d931ead7dbd174dc93df93cbd216ea454bae74d6a51f383679c7fcc2545ad25d5b2586af492498fc7bd0ffb95dc2b00b20b544359e17d1478548a68d1fc", 0x4a}, {&(0x7f0000000940)="301ddef419afce3f365a8fd662bc5a0108b5b31d43ae7c68fe33a021b6367802edea31d9059b79f295b80be9bda4b195de80c4d10d928fbef8a82c4855bf9932be1501bfcf83daf86654fa2379e534bacf22c915fa8f40b9b6cc2fbd576aa92bd6215b1e43c1b45fab01f1d096904124f469e55a4ec3b42444e040107730a55ec71846b728b9ef16d2f9c5f1e5421f0b27f6d8db3be02b0c1d794122297fa147b00769516172fdde79bf5a70f41b98c48fd73494828a5d404cc252696ab70807080aaf9daa5030ebba782aeab4cbb192a6cc464de71f8f6da53de65315aef43eaede70e4e3bd86bdb36e014596e946b67e479929e8877c19138dfe3241f9a73c3c7d7dc4a439d8ea6b7a31bed1c2bc6c310842fa72b7440268e444a2a01879e7e3b353d285975b186b3b375dd1c01ff8920b831c4edf8a4e74e00164477f5e36504f3292de590cb499aa25861c8d37b10ec8942c60dfdd0a2bbbdbcd733ea18b9b33abf3159d7a439ab070264b39eaa4bdf78ec144d80e26d50a225109172458b20bbab94912ccbe4f8123c3e883424cb1b4fb91624f653705d70aaba371abf3157f2cb5e95ab2f2bcc99c4aa7f337cbd32e00eebb9150d906fe7a7a6d501d8f3ca9c0b8d3b6288abc76e85bd5dbc12b89cf024ca6f7d8041dece2601fbb3ad911fbfa2690953928ab80967990c7c8710271eb4f2ac58fb2d15455d2826634aa5473508ad362fe0d322b4d459c1e01c60487ff596d7205e8c1a0cfdcfa2cff4f8adf850011bf8182fca4a26610446d33ce311da4c608fff89737617c0bf696d150fc001665dda6485d7075de904831698b94a8140ff3d2613f027c221998f068cfedbdaddb2f1f1365ba913e8ace3fa9414f50efa2910678763ad1275dc43cd78fefd1935cbecba4dfe1f3e0f5bbad02d977dcc84bd8d1c4677bc9798381b60161b373cb143c48cc8c00fa8491d82c7a15de9527452a1368beeb86f4b5501f0ded2f4c44af6ff0fbdaf8ce100aaf7511f821040e7c88da6c6bfc86608c7b398844cf862a06950eaafde3ddc7a33fc762d0e9d63606bca157a2eda3c0f8310f79786bc34cc227a308722bc54aee1b2ac4828908d31430a34bd08495b8fda5260ad0b424a9aa9d5ebf61ec8a83a7acbea61bd02d6d1fa12d26c0e3f8b71404821c123f1501624cc6b366997cf76651ac1418ea26e89ee2f6bad65ca3b7d737846ee708584d4de7dd3ddb71c04ee4d505227112e72f64203287f8a95c315703b5225d9421d687fdaa71b4ae98a2e77cbccee3120b1f72e70c3f84e0453b6054f2fe510b09bb22588d6a6f8d02e606bb33c4c9f5dfd3abf147738768ce6ec23363c7aec1d2aac5bf6dae182f12d12d9ff5dd047bf0911dbe95c606fd49d8282f07281db3db4d2148472aee8fc06fe3cec7ea3c4061e97a6ccd1381078e81384e32dce1fd9fc090b8ae5d203f374ab430fde2a779597edd1300a907eca6339a0baa666d4342db893cbb5fc1b9e06a93bed0afd9d301cba082a27cac434ca93fe285cfa115f7be0259c9d8a7575c887edd76848efd8f3fc6f1fe9708bb6bff5d9afb01876974f94a9df029e624a23fd5d9fd88f6e661c5bf53d09fddaa4ebb7be3248f01f712f9efe5a32e2c9a106d76c53684537a39e7df87f541a9152aa3987d4e0dcbe32971fd192f6ab0fabfd58392be7d92fa1f19aef4bacd6decaf1b7ff72303b187a3bb03ca05ac2d19a6583c02dca516ffb0743bcb38e4e8ea566f96697784ccd60e4d4fe86275a7f9dd3f94d631e09ffa7b2626f20dafa5b80943f099d0299eaf0b7526c2385a137c8b205964c848bb7624e4eaa4216c48a733ec9bd33deb12f4be09eece0edba1d6fdb1e2101da2887251be6464a9275c48196ce6cd46147f60fe0861ae2c9f5f8a0af0e9064556b9774bfd7aa4b06cc54a1755ba8e57eff0514a516669fa70c3ce065a1a885f7e8b903c892aeee8c0ad7bdc9fe4fbe3ce3d1b900b631e12318c672ef667398a3f602f64329a1c0c631124512572f8e149c4697c9b34a53b5cdb260d91f4459267f7f1cbc753f4d306c653b1960c0800a685eb476d008c32963e2d3b94f08419941599b708dc3552fd568a4bfe5bf2243a3a38d53872e53ded92818c3eec52385a97d7f581355c28e755ed353d9486d483eca9b59d7fb35d7d3ca79e10ff245e9a1f7e83b7676d35c0c2e2eb4e53bd2b76098d20eaa45752399c5a79608b90b042660385dbc8839515d8ebae25e0ea383a5663bf7207ab8e98b61cb7fc3badb88255e04978742ec2aea7ee0b433e1ce226535937dcb62901d0abfd94cc85b871fe694acf27646712015d6901a065c5eecdc1cd6cc5c6e64da97a8ee66f5eb5abbda13ee6308d059abf74d9116dd8ed092b5557ee678b64f5b3bb6eadd002e87d6f4753a8428c323d431847266a2cf0461ad06ceb1664441db92a6992876f0a91f68de6b107c0d12c10408de9b67715131b8c2792a6ee7f637e9866f8c1c3927678b5731fb4bf6827e603bd2425bc3b5f043dc048f39a65c6d8e115c2b9cf5203677e0cc263923981ca6aecc31d0723270f21c23986816c1a63d9693a275030634276ba015cb6dd3c3b23f2ca0fca02157d187550d4c42bf83d0fb62a42fc12cbd834d189e665155a3350bf60a8960cd6563260566f3aac88d52a320e65a707649e9d783d2180d7954a78dee92fa0c2a383f32754bdf67342b4d75a1ae4b18368763f138cd0296f0da9f193facc3da4ca2b1da04647dfbaa5f9497fc9fb741d43c3a2a155680b3cd0bbfdd3c3685e5ad76ec091a50fab87a300121ea140f28b6fd2aa0df616044c05cb350609778353d8ce111df21525d302b8786f8350f00255744bf05a2dc3fae375a3ea8b6abaa5db3ae2310f545afac4030c104e196b4fa257f509879ce76e6b8d2b92abcfc1d0f88852c67ac3f0c4cfbff8a580b4a121b83ca9d8ca16bbd46a70ed2567951a756a9e7a4bf37441f3cacebcb031b761035dac540ef78530ec124e2a44f4bfc8d095b882e96f099ac29396dbeecf60ccf4a6f5d4b517bddd6f6c4bb816816123d37e6e2bbe62e279d31131a5f13f75c050050a8678031c418b881b9f8680fbcfe9bdbed9ff86947bb808c4c6a990367af5271f0c6718ba76a1684ce755c208175744971cda7d0b910d0d7bb8919f4820cf5940b890b90976be89006a14bf0706e15830a016d3bcfc52f67f591c76672d2e107a0cf74d801bbf92779189dd5b3ef92daacb3958e591b9ad9fe451ba8e533814238e8a739c86973edc4a327b7b09b26973c3e11fece0042895f6be1a1ad953c7856f03cd3a941cd4104ce15cf8c9b53c44643b7fb9961f5303d6e6e72a2790ee1f129affb3dc69d9cd8ce1d6eef2c0079a6d9d0192ebf7c4f364ecfdc411cb3713c5444d01431b8cfd839f0bd783f304c7b80e3c992af7d246b4f23dd1918715f4292d1a18c0b60e352470000978d6cd1e80d64dc7473ac897d1cac97c0d9bc1f28f63e5eba9d6ec588427512742514b5a8062d75366fe77937ff5ba6a0f7bf5e20f5afcfee327588559a2c13f0bc4534c680c4eb07db972605a84f8708a6e5898442ce7fd2b0afce81353e9b30bf75a2e1b533e2c3ac0b8dbf75a85517982bccf4b20a263216174aa72a7b919e82ba6d5b5339a1f23299750fbec736e0678841480b9f72439e5d05e9bff369d6db221d8a26119f30d97d54c80d0324cfa1f298734be6843b37a95f7ed6150a2d003cbd05021ee59ada06ce6dedafa413d9417fca3f539e9005684f0860fbcbf9409548ada3f083ec513db85a30353adf83b30ada08324f76d9567c06b1b875746fc2590401ca15c0c8a84cbe65627610cd582222d785c7db45a89f3dbc689914da25443dcfe6d7fad3d6dc74653871d1be8461f0f9820b4c3d34a2fa90056de3d8dea21a0723c75d1f559fd769918829011cfd9c46c29903b05d411e7db7fc859482c3298cac162b533cec391a38a63a833176695340db5cea30da030b9629ca9be19d9a9c92c33712edc2d3c11f0dae9416b5e666f3d10ddac2a0164baf3794f5fa9242b5375f4772f65297a126663cc7e2d2ed0f0e234b3a00e66d65d2b704627c9f4c18da4f9fc60197948a7f0a4dae427595bc4bd75f1ec7db718595504f3a708dbdec3e11aee29eb7aaea12a96125c8fc38ea8799d5a9a242866fffb1d62347f94e9203cba17c1983da65246afe8341b372b40cf7ec5e55810adbc4519221470f0bf2e0a4775794eede0d98a9a5d84b619882a00b09b3a3bf067522b0a1120b0ff2d4b324bab9619fe21e7d1e8655382134fbb8bfc2be02b7e6d76a391585e2f50b0113471cb049da6dbb0578d641b3bbc957db56b38a18c52eae3caccb4f38d98c96c8774b345368c95eb2372189ea02933a5c7105bc9b336d0fd51fb9f3c8dd46bb51fbac9500d85865d2a7ee8a8f8ee511250a1fe7318787d63c5725b26ac49820f1d1cf56c7b06c15bafa5f203f7151563f1c65b710a8b2102c1d8cf80badc8df5e3e09f883c78a38cc85c7d77bc8895d6e81affbe87877ae43e984934343124a79b873430e6a4f8984cf53d8bbcb4535163ef34acb8859ff7f416b3b3046c58199a7d1e32cb37eba8fe82ae3a1013e5f9180b9ef969d5fd9c981fa4b10e90589168c5c376bfdef3bbff234542cd51825243a0e012568419a93f18e48be4c230f1189787b0ab11d595d8b067db805cc80c8fe8cb54f7796c6542a42fb85f17bb02fd63a8a569d2a67201e6b5fb139d30ea204b47fe74df8cecca610003c0419c6f1b9805b5c72ad0e980e10d6f15524094455bc09e62f560062684c714da90b8127506d5cc37aa1b07c76da6738d427d8fc4ebbf020cacbf24f859640cdca0df429662f4389f479c11ddaa193bdf214bca06c18dc98aa133970618bc269ce8dc0bca5f4a64e70299e0480a167762052206db4f274396bd022409921d06807d41f74b60504d26012d909ccffbda0d5f30158f8b6a9be9e49141acf22ab716efd7c7130634897317762732ab7b33749c39d22ec35fe79e26deefd86158c5d7b671be0c5b617950fdd76059274c00c20f02006fde7de0c158ce1aaffac4d7cb408bd35404611a42d753f6ff9f29e6d2e842bbb3989de396d025d17c120666fa41e96235cce160fc95c75ed430b5821610fd276c4779303335a0737ad8403cbfc5637405105f5e443ea5b7938aa002d1700e6e7daea7611f218de6d47b61340b57ae05047830f68ea298e39610f805360f46a652f22ca86f6f44ee0d7de7a7104bd66a85d0129a4e11ff5e16eec0b414edfa2dacc8d363dd64665e53a331c3c0e4153f23d1bba3495819b2c9fc95141c96e032bda7d25ad84703ef686b50b6b69b5514f62e14a7148d71599d296ae740abc5bfe212c4bc11eefcb5bf7cdfb0659944d55a63f3468b96be2be415ed15792a08d5ae410819fff0f5cc770cdb516c1b299c7d8d104918ac00cd9c5e35d5ebb139a3def2fecaef3f74eb16a45137772a79aa81760f5dc1898b24f5c255d4b77d932d0316df38e5163f0676b9bb9945470115bff4f8244208ca7823e1b3c09988d2f35dd16557d7b0a3e85cc343ff198c730bd125541034c6938fcd897675d13b41cf504b76a5d61b8e62449713b846d27f623f58a7f3d23f59c5cf1aecfcf950df1cc5466e748fae1725e2dd039a44cdff703382950a0ef86c2dccfa16be053e74d92be3d3c850097c110b2e0871b5fbd87795b2df71a27558e19981a730d702655bd683f89f48472577715a5006cf1c", 0x1000}], 0x4, 0x0, 0x0, 0x40000}}], 0x2, 0x4000) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:47 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0xb}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) [ 1727.761381][ T6180] Bluetooth: hci3: command 0x0405 tx timeout 00:28:47 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xc00000000000000) [ 1727.840253][ T5839] workqueue: Failed to create a rescuer kthread for wq "nfc4_nci_cmd_wq": -EINTR 00:28:47 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0xc}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:47 executing program 3: bpf$BPF_MAP_CONST_STR_FREEZE(0x16, &(0x7f0000000080)={0xffffffffffffffff, 0xffffffffffffffff}, 0x4) ioctl$sock_ipv6_tunnel_SIOCCHGTUNNEL(0xffffffffffffffff, 0x89f3, &(0x7f0000000140)={'ip6gre0\x00', &(0x7f00000000c0)={'syztnl0\x00', 0x0, 0x4, 0x2, 0x4, 0x1, 0x20, @mcast1, @private2, 0x7, 0x8000, 0x2, 0x9}}) (async) ioctl$sock_ipv6_tunnel_SIOCCHGTUNNEL(0xffffffffffffffff, 0x89f3, &(0x7f0000000140)={'ip6gre0\x00', &(0x7f00000000c0)={'syztnl0\x00', 0x0, 0x4, 0x2, 0x4, 0x1, 0x20, @mcast1, @private2, 0x7, 0x8000, 0x2, 0x9}}) r2 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000180)='/proc/sys/net/ipv4/vs/sync_threshold\x00', 0x2, 0x0) r3 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000100)='/proc/sys/net/ipv4/vs/am_droprate\x00', 0x2, 0x0) ioctl$IOCTL_GET_NCIDEV_IDX(r3, 0x4030582a, 0x0) r4 = bpf$MAP_CREATE(0x0, &(0x7f0000000340)=@bloom_filter={0x1e, 0xadab, 0xe89, 0xcc47, 0x30a, r2, 0x6, '\x00', r1, r2, 0x3, 0x4, 0x1, 0x1}, 0x48) bpf$BPF_PROG_RAW_TRACEPOINT_LOAD(0x5, &(0x7f0000000440)={0x11, 0x11, &(0x7f00000001c0)=@ringbuf={{0x18, 0x0, 0x0, 0x0, 0x401, 0x0, 0x0, 0x0, 0x3cb}, {{0x18, 0x1, 0x1, 0x0, r2}}, {}, [@map_idx={0x18, 0x4, 0x5, 0x0, 0xb}], {{}, {0x7, 0x0, 0xb, 0x2, 0x0, 0x0, 0x1}, {0x85, 0x0, 0x0, 0x84}}}, &(0x7f0000000280)='GPL\x00', 0x0, 0x2c, &(0x7f00000002c0)=""/44, 0x41100, 0x60, '\x00', 0x0, 0x0, r0, 0x8, 0x0, 0x0, 0x10, &(0x7f0000000300)={0x2, 0x3, 0x4, 0x9}, 0x10, 0x0, 0x0, 0x2, &(0x7f00000003c0)=[r3, r2, 0x1, r0, r4, r0, r2, r2], &(0x7f0000000400)=[{0x1, 0x3, 0x7, 0x9}, {0x3, 0x1, 0x10, 0x3}], 0x10, 0x6}, 0x90) bpf$MAP_CREATE(0x0, &(0x7f0000000000)=@base={0x11, 0x6, 0x800, 0x401, 0x81, r0, 0x1, '\x00', r1, r2, 0x5, 0x2, 0x5}, 0x48) (async) bpf$MAP_CREATE(0x0, &(0x7f0000000000)=@base={0x11, 0x6, 0x800, 0x401, 0x81, r0, 0x1, '\x00', r1, r2, 0x5, 0x2, 0x5}, 0x48) 00:28:47 executing program 4: bpf$MAP_CREATE(0xc, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) 00:28:47 executing program 0: r0 = socket$unix(0x1, 0x5, 0x0) (async, rerun: 32) r1 = openat$nci(0xffffffffffffff9c, &(0x7f0000000600), 0x2, 0x0) (rerun: 32) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000b80)=0x0) (async) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) (async, rerun: 64) r6 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (rerun: 64) r7 = getuid() (async) r8 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) (async, rerun: 32) r10 = geteuid() (rerun: 32) r11 = getpid() sendmmsg$unix(r4, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=ANY=[@ANYBLOB="28000000000000000100000001020000", @ANYRES32=r6, @ANYRES32=r5, @ANYRES32=r3, @ANYRES32, @ANYRES32=r3, @ANYRES32=r6, @ANYBLOB="1c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32=r7, @ANYRES32=0xee00, @ANYBLOB="0000000018000000000000000100000001000000", @ANYRES32=r4, @ANYRES32=r8, @ANYBLOB="1c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32, @ANYRES32, @ANYBLOB="000000001c000000000000000100000002000000", @ANYRES32=r9, @ANYRES32=r10, @ANYRES32=0x0, @ANYBLOB="000000001c000000000000000100000002000000", @ANYRES32=r11, @ANYRES32=0xee01, @ANYRES32=0xee01, @ANYBLOB='\x00\x00\x00\x00'], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001880)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7ccb3494e59c5e8ce8431ea3af6e0829fb4e8f523b80d22325b6151c5b4964fe11e8bd5217a714d310377c0ba5b19fd7181f4417ccdc371e0ae73e42570631082b8cabb246031894a55c875641ca17c39192378dd51f40f5699222c2a3a7df70344754107a0c5f2d7aa04f10f09cb5c11aa5e53a7542bac86581894cdad0e224dc67976013a7597503b05b8845a7db5314b17f6a9ba3893db1d5f9d2de5f0352c393cc072fdf967ae68ca17f69844bcfd753", 0x144}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) (async) r12 = socket$inet6_mptcp(0xa, 0x1, 0x106) getsockopt$inet6_mptcp_buf(r12, 0x11c, 0x4, &(0x7f00000000c0)=""/101, &(0x7f0000000040)=0x65) socketpair$nbd(0x1, 0x1, 0x0, &(0x7f0000000bc0)={0xffffffffffffffff}) (async) r14 = socket$nl_generic(0x10, 0x3, 0x10) (async, rerun: 32) r15 = openat$ttyS3(0xffffffffffffff9c, &(0x7f0000000c00), 0x55d7c0, 0x0) (async, rerun: 32) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c40)=0x0) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) (async) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) (async) r20 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) r21 = getuid() r22 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) (async) r24 = geteuid() r25 = getpid() sendmmsg$unix(r18, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r19, r17, 0xffffffffffffffff, r17, r20]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r21, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r18, r22]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r23, r24}}}, @cred={{0x1c, 0x1, 0x2, {r25, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) sendmmsg$unix(r0, &(0x7f0000000d40)=[{{&(0x7f0000000040)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000580)=[{&(0x7f00000000c0)="83cf82f9060b679a7f08be646de6b08f77678861332db9ca2cbca6", 0x1b}, {&(0x7f0000000100)="dfd4db580959da40b397e0c63e8b3d87c86cf1aea7d80827be861f6822431dc856c50df9a4fbc82cc040c0f5f03136f410c8acc146402f4aab40218d9c73e56c1ae8241726118267b459599fa4bdb7973a7540cf9accb13f1877059cdce7efc1cd706e8e50629d7edfb2a87f6d0639435828d5c4ca7219a89ce2fb9d2a8b9513cb84aad43d2f329ee1db7cbfe4bc779f66e3df27de4e2b1f51085f3f6861e5d319c07eb08e187122127e44b418467f2084cb28a1742be5b35d61e74d26fae23e86c59748d42c", 0xc6}, {&(0x7f0000000200)="5258b90232dd5dc3391d28ada7927fce7e855695ed8e02aec490081ea011305d2efe1cb20c4d55c5daff2bec67a98bba1480ac24368c9be46e8185911adf35a759306b67aeedc93a8644c4a6d9647ae44356cd731148d1e6b5bb18f566b2559d4c88e1ed0eb248e0fbf8ecef1eb3fb3d8e9adcc4ce76baa91973918834fe864b500e72bdbc9c4e06e171cff227880be452fc92dafa24b1b85e7a470925f1374ead14d5f255b2543b7fab6a47a9e82592305952f259398f1b3b297b27c175218258fd8f05376b5c811ab83a7a0bf6a8393793b159a13d935bef", 0xd9}, {&(0x7f0000000300)="920e4581afed1ee67dd8dc0a367370c293d97e3feb028eb72a2815b2255a6e5b1c5d533ff3a21a85cd6be9482b4cb6d8f2ecc638bcc06a0f334fd66bf75ba5969ccb9d4951f597318baeaa56cfbfb6c488f93bae6129892d27b46abf85ba5b85eddb06b250e0408d720da2b0a87ddfda4850c921e26bb18616dd57d9486cca4db6dcbeb4b247f22f20fb", 0x8a}, {&(0x7f00000003c0)="fa0cce0f4b622ba48d4826e60cf36464ca7be5f5291b96bd97594132b527492181fcf3bb61b750780cb5603a837c22e13a7a29cd683ca2c5183b1e1781ed1daa0d2120114cafbe6d1636180f1e091a487850660fd5a86653264a84294e4a57c8fc6fbb560c2161a4e59a02eaa48545a632df4cfc73aa0292d6376f113878089cbe7fc43bd127bd864bca06d63fa54c78c4c3d26012b831e76a6f94f16e813d88ea25e174c5fa363a91cda51ae1", 0xad}, {&(0x7f0000000480)="ff89e60b90cda2b57ba52787d3b6b94b1f39f4c1bb1ca5a8905ef8165fc76bd47f4dce35aa0cb9f31642d0ad473836e8262f866aa03968999e073b6fb0946747f0510c25e0291b7a9c6de36474a090f7701971df35aa3d3909dddfbe23aa103db519ade03fa8902ac14054733cd1ce648fb7e38f81e42f166abc8fcdcd9995cba61316d3f1091181e373644e03be0267a6e0548e484c7ff9928c4c67e968a2d9c20388d266eafc464e0344ca58ff051235686e1a0b13370130eb18fbe3f763ca5a85a3b9dd58717017a2569f39aa67b6762507d50857fdff25b03193c2f1d682e3c8a4b57320d4d15b3e1fb82a7a24f8e30b", 0xf2}], 0x6, &(0x7f0000000640)=[@rights={{0x14, 0x1, 0x1, [r1]}}], 0x18, 0x44044}}, {{0x0, 0x0, &(0x7f0000000b00)=[{&(0x7f0000000680)="43a9cc03073ccb92c777dcd1f269fdde4542e1a8b3e9f7bb81cd6df8c2fa5e9056814ce4bedf2c0bb323d4a659f9c0b4fec943f8ec19e6dce4aba0333adf98ec9c07b63c7c7efc33d03b698e687979d0f479a1a61bed0abab3ee6708ddfb5789bf15a5bf9633350f57f72e9c6f7d4112ab815b679c12ec7c326a8a42a6dd076f31f396304dfcffc7ebc2ad48b296159de59012d334918e8a8e74e950bebc", 0x9e}, {&(0x7f0000000740)="1b5d9835a1d0cf5a711dcd222d18e5c74e4005712dd07d2678ddda1c5c8ca1319ff36f11ab1c153a44c8b380e09be4a404d9f165318eea07744d3ec144c39d484a1b4264edf710d4bf861818d752b9d6bf68f378f8ec37cecc81a996bc09830779770887b65ddb6ed6e67edc22cb641e07e59e2ac4c3175cce4d1b763fe1c1a218a3f360e4d1622b16bedb7a0082", 0x8e}, {&(0x7f0000000800)="f611def51abed9593db5dcc8ebc03ee4d4d4c68ece61810b3373e9f42bb82691c8715922278c91c951c855d34ef91ec3af4cb2d4e120fd83", 0x38}, {&(0x7f0000000840)="13e37db8468046f8feccd3169267df31d7c53163b7a63b91b88c88e916cefee236d47889a2d3cbdb8a3609a4cb80d1f3c3d4cab69371517cc5bc4fcae2e825228c", 0x41}, {&(0x7f00000008c0)="72aade7d64b73a8be93eeb66b8f8bff8dd21310e83afb3f18ac6e1991fc33381852e3bbc91a4aea287abb11c23b95de72cd640db6277ddee4c3fa63dc757f905b76427559bee28fbedba690f7ca2ac3e7a79454be96904da256bba54fd737fa80f4e7f5f84b5828ba06aa86fb3a5af2cc3bd59bffe951e557c8677969b37c5cd0fd329984a6f0b", 0x87}, {&(0x7f0000000980)="3f653667865a9c8f8d59302ae454de6ab4182def183ad3c1889fcbc53636c7704da28e936e61f294d0ae25b4ef046d9a443153b2933cff6b6336aa9063ee6a7c5ea99fa439ff052a3f26250a80f6eb39c1d8f18bc3fba7b09255779571348a4909de7b23717f3fdf4f4ad7bd06440e5ae94b6d94c844583a8c3f87806f3e31ebc6207915e0082d314011b2740f32b986ae69a7caa4405444be44faffe11e7bcc6cbea08ef52717b67b38bb5890714fcf5e4831c09feefa000bd26aec0652d89dfd63a197e614634f99c8266a499919badd97286169fb7f7fea731b66907cb8008716e09ea4", 0xe5}, {&(0x7f0000000a80)="5f2a7b9af0c75453eb890b9849b306b4238a4de99d93a3bb48bc59397f0fe95237178f5a44ca450dfe16b7f03d45c38f07854668f8b0d587007022b1f792b1db483ffca6330c1d464d706ef3b2754791be94713c11bd51891183131554abdff3cc48070cbc7aad46f65f2ce85971f616083440e3e3d934a11689c3b2c9a8cf", 0x7f}], 0x7, &(0x7f0000000c80)=[@cred={{0x1c, 0x1, 0x2, {r2, r10, 0xee01}}}, @rights={{0x20, 0x1, 0x1, [r12, 0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff]}}, @rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, 0xffffffffffffffff, 0xffffffffffffffff, r13, r14, r15]}}, @rights={{0x14, 0x1, 0x1, [0xffffffffffffffff]}}, @cred={{0x1c, 0x1, 0x2, {r16, r21, 0xffffffffffffffff}}}], 0xa0, 0x800}}], 0x2, 0x11) (async) r26 = fsopen(&(0x7f0000000000)='bdev\x00', 0x1) fsconfig$FSCONFIG_CMD_CREATE(r26, 0x6, 0x0, 0x0, 0x0) 00:28:47 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0xd}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:47 executing program 1: r0 = socket$can_j1939(0x1d, 0x2, 0x7) getsockopt$SO_J1939_PROMISC(r0, 0x6b, 0x2, 0x0, 0x0) 00:28:47 executing program 2: r0 = socket$can_j1939(0x1d, 0x2, 0x7) getsockopt$SO_J1939_PROMISC(r0, 0x6b, 0x2, 0x0, 0x0) 00:28:47 executing program 3: r0 = socket$can_j1939(0x1d, 0x2, 0x7) getsockopt$SO_J1939_PROMISC(r0, 0x6b, 0x2, &(0x7f0000000000), &(0x7f0000000040)=0x4) getsockopt$SO_J1939_PROMISC(r0, 0x6b, 0x2, 0x0, &(0x7f0000000240)) 00:28:47 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0xe}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) [ 1727.847233][ T5840] workqueue: Failed to create a rescuer kthread for wq "nfc3_nci_cmd_wq": -EINTR 00:28:48 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000840)=0x0) r4 = geteuid() r5 = getpid() r6 = geteuid() r7 = getpid() socketpair$unix(0x1, 0x2, 0x0, &(0x7f0000000000)={0xffffffffffffffff}) sendmsg$unix(r8, &(0x7f0000000980)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000880)=[@cred={{0x1c, 0x1, 0x2, {r7, 0xee01, 0xee01}}}, @rights={{0x10}}], 0x30}, 0x0) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r12 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r13 = getuid() r14 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r16 = geteuid() r17 = getpid() sendmmsg$unix(r10, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r11, r9, 0xffffffffffffffff, r9, r12]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r13, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r10, r14]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r15, r16}}}, @cred={{0x1c, 0x1, 0x2, {r17, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f00000008c0)=0x0) getresuid(&(0x7f0000000900), &(0x7f0000000940), &(0x7f0000000980)=0x0) r20 = getpid() r21 = geteuid() r22 = socket$nl_generic(0x10, 0x3, 0x10) sendmsg$MPTCP_PM_CMD_ANNOUNCE(r22, &(0x7f0000000140)={0x0, 0x0, &(0x7f0000000100)={&(0x7f00000000c0)={0x1c, 0x0, 0x0, 0x0, 0x0, {}, [@MPTCP_PM_ATTR_SUBFLOWS]}, 0xfffffe7d}}, 0x0) sendmmsg$unix(r1, &(0x7f0000001e40)=[{{&(0x7f0000000240)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f00000002c0)=[{&(0x7f00000001c0)="76423d5f1885570b6227e94341cb5d7a7462015693fdeae1907e8e202b89eda40a1425", 0x23}], 0x1, 0x0, 0x0, 0x4040840}}, {{&(0x7f0000000300)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000680)=[{&(0x7f0000000380)="b3cccb746c7e8dd2f7a7a6ee49e8a28d7a6a92897aa4f01c4e935f72ca59b9e0a3ddcbe3f3ecd6b451d0f035955023b7937d7c63561dc26ff382f9607e5e0da372eb197bc02bacca1fa3005d578394c5c802b07ef94b31aa35bc7911005fe28311b42d8d3aac33804a7d22075aca3231995daef6c86d881968cc5333f4397de6eefd9f4d7d52d61dffaf9ef29ba5f3a7c4a976fa5f2e4bb2ccc5cb35", 0x9c}, {&(0x7f0000000440)="90da79fbddceffb56ba41c009fea6287abdfdbd785ae1673e6272f7d72ac19a089d71b8f382f24dafd16b8f9c5ff39aad9b2ec2b78fe09895f97e2f5bc77e733e970bf99bba1bc25608bf3c6c2ca890dad28ffb496542c9d87edf38166a987724d96696c414f4bae501f9f91d23d0975a9fa614f6becd105da5a2332cf2ae971061d", 0x82}, {&(0x7f00000005c0)="ff12d3b4d6cc3ba8b93a4df6b65c5bdc066d5fbd6dd0d11590bc49e2a23baca70bf9a54719086a55a8a4037cac5ccc7313d0cd80ba915abbfcce1fc6713b3d4ef85cde2bc46084bd0cec099b0cecdd", 0x4f}, {&(0x7f0000000540)="4943d8f7407fcd40f757d64c10e4", 0xe}, {&(0x7f0000000640)="0bf93cd7a56e715857caa3293b2dbefb044b21c5523514f7508127135f", 0x1d}], 0x5, 0x0, 0x0, 0x20000000}}, {{0x0, 0x0, &(0x7f0000000800)=[{&(0x7f0000000700)="285be241345d36945b97af0cd58d70ca268b1f2bea7ca5f48c2347bb1071931a2c2cb441520ccb96cadfc48dc32a8bfea3cf48b882cddc75d1e769aa56e23b43d834b7d78ab911892978cc74f271493c66875399e003c63543ea0ae705e7e11e23c94c5c766e24935dffaa589415d8d1e4a491374a4b793d5c79c5cd15cb0285abcbcbd8edb7ddc8fc650c8d7e39517da3c3d9bac545077a11010cdecbd6406c7d2350bec9417f043b49cc5c7141b64977d139f2a0a3c4276080a4a49e5c993004c4b7f2e9792cbe24d24e17c6cdac79d8d0d08d16c39a14c82cd5d0f0f78ad452af66011ee2118ef29d56", 0xeb}], 0x1, &(0x7f00000009c0)=[@rights={{0x18, 0x1, 0x1, [r2, r2]}}, @cred={{0x1c, 0x1, 0x2, {r3, r4, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r5, r6}}}, @cred={{0x1c, 0x1, 0x2, {r7, r13, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r18, r19, 0xee00}}}, @rights={{0x20, 0x1, 0x1, [r2, r1, r0, r1]}}, @rights={{0x18, 0x1, 0x1, [r0, r0]}}], 0xd0, 0x85}}, {{&(0x7f0000000ac0)=@abs={0x0, 0x0, 0x4e23}, 0x6e, &(0x7f0000001d40)=[{&(0x7f0000000b40)="ace2", 0x2}, {0x0}, {&(0x7f0000000b80)="c48b23f5b351c5ab75b27369c9870b2080afdc131bca41c1f8ef2d650ecdeaf7829ac4d1dd6c589f92a275c11e8d281ec3d5a11bc7275a0890d71ad0fbbfd0628f68ed4c4b777d415cec4367a6ddb04fdc460535d774fcd1ea739187125bce6c26e9b568e30af7cdd0e1c102ebe350f13079f5a5745211ba1f243b9c523f4ec230c6682958f01ffd46734edfb4d8ffc0bdd4e87cd4ba4a185e7c32f908246fb6b4c99cc4d29770df551fe6180658965e791b99afc586166b9b376c1fca9d35c584ed6158a58210faef7c72ce7a748cfd310fc16da5f0d36df7314d41e21bbdbf", 0xe0}, {&(0x7f0000000c80)="cecfe6435d750907ca81630b06f15e1ddc39a961688abc70cbfec1ff89ee6e29db835f18ca9d7cfa9b9eca2b2c1523702a7edaa82becc01be5ad14b2a239741b30fec95c7d89e511ef592afa113ab4f8549767e12e6ecae8b61bf5c30ce38c2790c746cc5bcef1d38551e7fc4e9b45b885a53ffb8780051941449f06610a6a108eb04218406815638bdecd9a9c9ba7fa0219588665cd947767", 0x99}, {&(0x7f0000000d40)="f0f4caa0f3db8620ba8d46f0101acaec4138678175d62ee34b67413621132ffdba1eae85424b6dca4b266341e2ae1817e51d739e3a2b5e758855e60fe07858db7cb42ec35b8322ef339a88cbf0975f112517cd9d31d15dc2f1e3126682ccb6ab6d7d2863316a0a54726043b11ef83b3a0077732000ce9363070e3216a1219a8e92f35f85a4eed7949313bc6034479e4db116ed1a1f7bce97042b5da00410c4b5e7799d8e82b154b10bbe916246b29a49780d1766ce62fefc1230eda124c5591fd6eb3a8771e257356f906cbc32ac53b8229542ed12c4a870ce07b109014904c8c1d9bf3b6bd76beb713af46fe3c850bf8711e67e2c58f1041130957a7259c8c41d366078d86e71f71807425596f55ae1b2c595e5681b358fb092a816ebcfeb2c7ba925b3cdec536f4038aaa9ddfec47fd90123e45cb6483eb4beb8eb533e7c853dcfd2db8144222c456917e9fbc4c9490bf5a9d437ad94dd0e57d6251e30e62141eb80f6f521f15b096531f11a5bcb76684165a538ebb422b3ee933cb94368f2eeca1438f3983e73b827f54b1cfab7f807ad8c0e4b2a04432fca7fdc2bd17d63d19e7846bfa28cc987819d2b1a799cc1e7f73a0a347284c2a93c0cd245617059d3b0a6e536b4a55b711b38b82287d05f11350817c01b43c0c02c0037bbf4abdda690be1e7172dc730d4b024da7cc7530dee44aeab9467f97d73255d91cf77bf2f9fc2c177cf5b56f8df74285d49c53daffd12fb4104050a066e2eff9708cb966c43eec25acaee7905ea1662d68d3cb39c8c162aa590de8dbf04551ecee7d0525894412b2949fcf88a7ca2244c921d81821e4cad5f5d9ca88f9dae07f1abdb2b93dd6969dafbecd4f6a5548e640e81c275aa96df6cc24053cc34b9c53044a59869417c6a8f0a6b6ff49173075ead25a4e0f61684e63510dc7c99f37d6c474d534ad3571e8926e1a38ddb8966f748ae3a27b9e3ca9181d647dad99e720f7d024093e0f5a6b743ec7f723ecee2bcb819b71d147d0a7b4ca18ad6aa431d9d9bd15625370baad0698b0d7d37d8d347e59425cf494aadd9a400ae19f6beb075d6b815caa7b73cc9c0ef88a5569959c8cb641f3001063fb797bf2c79f7b98664a2dbb2ac63958a1966876c030d29f22a6c444dbfb072bffe9551eb6980fab67776d9b8d1387fccd81c4f85915e5b535b53e3d8c3770c23e0a85a3e0b93c45949e92e5f64e815da24467d1e3903c2f7674749f647ed2bbfeb8e54d7ee1126c33b62b1498d07f5d3ae2209345756df47951c9c93388b8fb235e38e46bbe020284b46657981de23dea92867e5330e088329a8dbee029570aa895264b2990b4ddb59f5aac258efb9531ca09796de3df8a027cc3478d4ba4b5935729617ae5b628a76e33738a3be67fc43ff33ed0d460f04ed391b61c10e07f06d80bdbe0a25259b806fbbd14fb6cdb6a7f260991623c970680f92d6853e586a73e21d0a2c7613bdafdfc1014a4a7f699b01ed663e3d733baee1a8dcd5fce5119bcda2836ec31095a1752a1eb6bba269c72c6c92c029a5d5233ec9d7a715e72b614015cd08c6bbfd12c3131368964873f34bff431d7fbc21930f16ce4c77e356de3950737c58036a0022ddb64a36bf868ada075ac20473fd4078165dd5a3229b648cafb848bc75ae3f656ef53779a6b8fc22ec18c35e047cba58d7f787ce8b732c45d0e772d1160a2f04d46fbb6c0f5ba666375b57ed89a1e9dbc8206cefcf16255d0463cb87f328f4cd54adca4124ca5390d28ab91162c1379c98283eba434e3164cba07700b36b4c2a8c5635281c81174f343b99838c2cd42ea9e38b33815f388e4a4b72a4a5d8fd7a4036b78ae51fae60618f38eb58d075b996de25c754eab64c8cc410815b00c78047de13e1a7c57b16cb1fcb3796588b459af95801597c1dc7ca86898eea2babdb54948f5e52a0ec40d8bb13921074b886d253147ec5f00ba268794ad70620c47ec1b321e201fccd65955380965a2dc654708d76bcf6be2becc570dbccbafb741a30e79645700ab7b4024116940fc102234e326c33ad87515660e7bc87b74e55d8b83cdfada8bfb9995409fb35c88d4178fb1c5ce72bfae9fb4e157c7ea0c5049b25553852c37d482e916bbaaadf884581608d295a035abde1e79e6ad939ae079d161a065288ec76a949febe8f83ed0428088efbb5d2a5e38c01cd36d76ed6228ef0f2d344969e0fc01528797b54cd9521f324e102d42a8234780fdaab75cfb3bcf23a9c8dc398d9ea299a0a1f863839d2e218ec2bd2055628d4d534e0d316c83716eab025b8017e7e60d0ba52ff422b4ea1e9c2d7b8d692f95af521375fb05489882c9d83e86101f918a6b6be02b5127d5b97b016f39dca6fe90416913c0ba2e06327491926bdda888f95aa8191de3a3df1d80d7d70d15acd711b07c29122841546b6c8a4b790c350a97e6778a94a4d7ca704a4c30486e91a1cd061fe9fd8b02f3f35531c6da6d4f1560029f07bf13d2af8fb086afe27f9f7e6ea1820591efdb4869806c002ff7d9a0aaaec6f45f3f6ef84ade881d5b5225780bccfc489b2e1e897c3fafb01791c1ad759bac08948b1eaaf5de2926abbfb0433e0a07ca4599ae9dfcd6615dd37bb64cf21213d6cad6c4bd520f4dfdbf4fd958a909acd8ac262b629174108acf9548a646e58c11da745622aa21e3174e08fb8c9e385503beccc7b856974a7e76736f47893d07a07c579fb67bde3f26e9ead9ebd292801ceca355a5b6f7957ec7eb4957fb43bbcf9b7414dd9cb2892853485a51f191484ad4db99ec28126502e5d22338894ad183759a966685b259ea4c3333bc9e4557da97dc91f2fd8e6b6bfc1ec8e6dad76cc2c662d20bca5f249ca99939a6231ccd4ebe01b6e128f407865b5ec0c48fd8e27f15b423be6e2dba4476455de81b73293adabdcaf1c644b81319f5e86618bf29851db3bed04f4bb26f5fb745a3053d85b859de3d74fd7b0a61fb2ea9c6fee8b8ff4e9dafb76f92769609b29a2f33a0a6aeb744a7265db838dcc529e379449202eba9eea7b4524cd13be90755647ab45c7a730364654fdf6a4df84b0a4d1435eda1dc11ebf90767c58c279abaf04533f5649cb726c26043ba78b4878d728c748ae49f8ad1773478b5b5d32a7896ab4951ed3c1fe682899e0d0a403b2f675c97c03c6c88c6b789558ef26c1568d28b9cbfcd88bd33192f5ef0dffa6bc4bd270a25822bbeb14a6dfffea448a21d582744ea2d7cf569057f5cf2907a51213200cb36ea3262e1d362ffcb159c0835d1740b01f4a856088f2bf9783351ab4ff5cf93febedfb1883b3c13740082aa97b56d37a927101fa6b22580128ddcaecca6a5a44445cb2c8dcd24832bede09aac176881ed4aef2ae51937db93a68de35245def99405ee253df104e5d5942feaa2ef4541da9d1ac77723f402fde9ccaab5cbff26bdf00592ad5a447805160ca88e48581b00323b8baee7d068c59808dde7ad02590abf3a617478d2ad407ac88de0377c394d370bcbed63eb4403844b1bd0bf7b375b2b62924191a46cb65c133bc8cea0b19a7e0934411d4fd4d40566dc5df110923d3c242774aef86e85d9cad3579750a827122581646742037dc69091fa189ce31386a291d306b7098545d9d4a212c349160ac6cb4f09e6c801b1e7bf40be1b468f768f3ebc8d4185e3c5668568ee3a276d959a70a485caa77e711f4a90718ddaf9708ae22b7112491dbb261655f0c372f30347820db5c12b0a864ba05e3cef4b0002fe13e85447f12009ecebcc6eb908346fc3b9170335bcf9ee891751baaf6fc6f751d897c66ea7715ca1124eba0885de16ce4306abd54fda0044b56f8d31111b4fd0eb431d7a46fece4ba30d1a4da808c203e187d2300a8a06a3ac0624c97f24f626b2803fb0bd7fec8ca1804b74d7b2ce2ec6f890700fb9aea81d89dc048e53768ff42d48841461fb9f21cca32505f7d9f44bead4f2b2e739615846e40cda8cc69df7f898e0f8b951f00827a562edd04fa688135ca5c514c580083cfea9848583938c83d95e74bb6dc65091cb5bdb141734c358ecec3265de302b428ff95fa998544bbb1284c84cc6133ca2b8b578077565f3e7aacd04d9970c484956bbf95d256628556fb48d41a78e8172d347f040a95906a2144dc02572b53ce09c01b7c613204bf45316aabd7b9921cdbef15996d617f88bbc4780344e0b6eb8f35f4dcbaa435b112b2f003cba9bf1cdd80f1f5b0a79b6047af80ded8cb29de9ec13e2aef6e841d28bb7069eb9f0469b4c65a31c3ebd9ed67541ada74c97426c807131c8b760dc203c097625c6573766e1f804c18e023ddcfdb04aad6ad986cecffc2fc5313a68c3bc82332dafc6418fc31d591197928a5ae4dce47d046cc1c8a3ddbd1932e2e71ab8c85b22270049d7ddb044038eb7d4a9d2ad337a21492441a25e9918393fa972505b710b30b288c293382072cf19b8b3fd15028fdc580e543e9bc9fea90638fa9fb45249b26caf5b0d9fd20423f0b6313c37b63e2966ce408d1ad667cfe71694e39930858cf71d2a1bc36e4d6183c4d1befb951bdcc529925a98eac920b942699d05643c5b64605e7c665251122c5011d3573a87885bf92dbc231021591e610b4c6c1916d9777e0ba2f4475172afff440848b0f528d042014bbc2b9b6d175565451cb30a485ee5fd7505504897fb6788f58ce377697bc6c597f213539279455b5644af19f80de6a01b7805f1c913c23af4c70d39d15bcd6659424250384a13ca0aa25cd109a6ed0c3d7b8f8b47f8da788bc78109cb448925db6b04ba9283e41d2ee60e2e194216c766a855730c842a7d640343aa6c88f86ccf6dc4e70fc6125094f5ce537fe57fb4b72c6b19a4321a959f65c818cabba9ae0591255598b0748f90b1f2240579f89a9c692eee9b9340028016fe0eb5b73aac247b6c06a2ad3fd65687475c2f1968e248d29f602f14a1ed37a1dffc48a3adb61a76c478b83b82c0c66098e11509cdcc74ad63b97b60c076580c928bcb15581a0d5abb5b71bea3535bb2052c8e4a9d01f2a55302b265a311acc6500115b717d95cc3d1fa8bf20c92b7f6c89e125a499ed4b64e1ecc487ec05463323bcb96f7d3b4fd27fbd90ae64df2fa66c62a17f71acbfd5f810dd620880cb84aa8684c136d7e6f4347ec2419178201ec476002aa6b1de02c93bdd060e2cd78947807a010e5c4966d87ae7448e6009830f271ae18bc89287f5c7a9920f72d7973d3c041ae239174459efb8c4cd25dfb88ef181393865f4994b40c6d59463cc42b146e29ba9bf83581672d4405905ef2212fc03d7acde545ed3c205a613dec581624b43e6ea77d12d04967680f40a28aeb9269f8feba56ec7ae7686e6dbef87c9fc148e4495be64b1433a191f4a41089911e82e84073dd7ba81cdf368f614ff1cee2f1b8dcf799dd6e74bbf4cf41171102ab657615933faf14e326dc03e0eaf17d50f90c8705c2e5d8a83c2e54ca2d6d7a0cb249a299c7fc4f99aad94fee95a3ef8f976c1843d6f798cd99633b65a4bfc293064e498873ceb8da4253ac0fa3667956c327c8425e3134be0d022cb6277c918c86ee402c5ec987369b70af3cae0dd1014e3f1472ffcb79b7a78d642b99894a7c79ad4eeb9e21008d96d605d799df055334a2396ef163042737f18a882d4bf2445685f021434ea98deb58c26a93a19d8eb7b9aa4a3baa2dfc0233abf533baac960a34f225d3e92b757b81fdafa66aa2b64d12a6044b7b411298ddfdfe6feba1a53a417ee5c41539b0c6e19475ba0f98", 0x1000}], 0x5, &(0x7f0000001dc0)=[@cred={{0x1c, 0x1, 0x2, {r20, r21, 0xee00}}}, @rights={{0x1c, 0x1, 0x1, [r0, r1, r0]}}, @rights={{0x38, 0x1, 0x1, [r2, r1, r0, r22, r1, r0, r2, r2, r0, r1]}}], 0x78, 0x880}}], 0x4, 0x4000001) sendmsg$unix(r0, &(0x7f0000000200)={&(0x7f0000000040)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000180)=[{&(0x7f00000000c0)="a38885aa15e75c4726a46f09eae15a9710a1d3ef71b6bb9e779b32ac723b9d7a6392f91a2ac844faecaab889db2873b2d39d0f57acc25dc5e52d25a4b2b23566abb083dfd40b523b7d1edae4b52c9cdf40a82e2e2115b34379525cd9626ae9ec132203333f81fca22b6ac15b4ee1e664665e8028b77bae6ceb2be81e47b047a53c65310e65f4c9496358f2905411250922e8f099b326cbc8c3f1e025f0e21d4084c35315390473", 0xa7}], 0x1, &(0x7f00000001c0), 0x0, 0x40}, 0x8000) 00:28:48 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xd00000000000000) 00:28:48 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x100000) getpid() getpid() getuid() getuid() 00:28:48 executing program 4: r0 = bpf$MAP_CREATE(0x0, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) bpf$BPF_PROG_RAW_TRACEPOINT_LOAD(0x5, &(0x7f0000000240)={0x11, 0xb, &(0x7f0000000000)=@raw=[@alu={0x4, 0x1, 0xc, 0x0, 0x3, 0xfffffffffffffffc, 0xffffffffffffffff}, @printk={@lx, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x3, 0x0, 0x0, 0x2}}, @map_val={0x18, 0xb, 0x2, 0x0, r0, 0x0, 0x0, 0x0, 0x7}], &(0x7f0000000080)='syzkaller\x00', 0x7, 0x60, &(0x7f00000000c0)=""/96, 0x41100, 0x61, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, &(0x7f0000000140)={0x9, 0x1}, 0x8, 0x10, &(0x7f0000000180)={0x1, 0x2, 0x0, 0x80000000}, 0x10, 0x0, 0x0, 0x8, 0x0, &(0x7f00000001c0)=[{0x3, 0x1, 0xf, 0x3}, {0x0, 0x2, 0xa, 0xc}, {0x3, 0x4, 0xd, 0xc}, {0x5, 0x1, 0x2, 0xa}, {0x3, 0x4, 0xe, 0x7}, {0x3, 0x4, 0xe, 0xb}, {0x0, 0x5, 0xb, 0xb}, {0x4, 0x5, 0x6, 0xa}], 0x10, 0x2}, 0x90) 00:28:48 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0xf}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:48 executing program 0: bpf$MAP_CREATE(0xd, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) 00:28:48 executing program 2: fspick(0xffffffffffffffff, &(0x7f0000000000)='./file0\x00', 0x1) r0 = fsopen(&(0x7f0000000240)='virtiofs\x00', 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, 0x0, 0x0, 0xffffffffffffff9c) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:48 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0xe00000000000000) 00:28:48 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x10}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:48 executing program 1: r0 = bpf$MAP_CREATE(0x0, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) bpf$BPF_PROG_RAW_TRACEPOINT_LOAD(0x5, &(0x7f0000000240)={0x11, 0xb, &(0x7f0000000000)=@raw=[@alu={0x4, 0x1, 0xc, 0x0, 0x3, 0xfffffffffffffffc, 0xffffffffffffffff}, @printk={@lx, {}, {}, {}, {}, {0x7, 0x0, 0xb, 0x3, 0x0, 0x0, 0x2}}, @map_val={0x18, 0xb, 0x2, 0x0, r0, 0x0, 0x0, 0x0, 0x7}], &(0x7f0000000080)='syzkaller\x00', 0x7, 0x60, &(0x7f00000000c0)=""/96, 0x41100, 0x61, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, &(0x7f0000000140)={0x9, 0x1}, 0x8, 0x10, &(0x7f0000000180)={0x1, 0x2, 0x0, 0x80000000}, 0x10, 0x0, 0x0, 0x8, 0x0, &(0x7f00000001c0)=[{0x3, 0x1, 0xf, 0x3}, {0x0, 0x2, 0xa, 0xc}, {0x3, 0x4, 0xd, 0xc}, {0x5, 0x1, 0x2, 0xa}, {0x3, 0x4, 0xe, 0x7}, {0x3, 0x4, 0xe, 0xb}, {0x0, 0x5, 0xb, 0xb}, {0x4, 0x5, 0x6, 0xa}], 0x10, 0x2}, 0x90) 00:28:48 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x1fffff) getpid() getpid() getuid() getuid() [ 1728.027914][ T5890] workqueue: Failed to create a rescuer kthread for wq "nfc4_nci_rx_wq": -EINTR [ 1728.263875][ T5940] virtio-fs: tag <(null)> not found 00:28:48 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r1, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000840)=0x0) (async) r4 = geteuid() r5 = getpid() (async) r6 = geteuid() (async) r7 = getpid() (async) socketpair$unix(0x1, 0x2, 0x0, &(0x7f0000000000)={0xffffffffffffffff}) sendmsg$unix(r8, &(0x7f0000000980)={0x0, 0x0, 0x0, 0x0, &(0x7f0000000880)=[@cred={{0x1c, 0x1, 0x2, {r7, 0xee01, 0xee01}}}, @rights={{0x10}}], 0x30}, 0x0) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) (async) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) (async) r12 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r13 = getuid() r14 = socket$inet_tcp(0x2, 0x1, 0x0) (async) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) (async) r16 = geteuid() (async) r17 = getpid() sendmmsg$unix(r10, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r11, r9, 0xffffffffffffffff, r9, r12]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r13, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r10, r14]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r15, r16}}}, @cred={{0x1c, 0x1, 0x2, {r17, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) (async) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f00000008c0)=0x0) (async) getresuid(&(0x7f0000000900), &(0x7f0000000940), &(0x7f0000000980)=0x0) (async) r20 = getpid() (async) r21 = geteuid() (async) r22 = socket$nl_generic(0x10, 0x3, 0x10) sendmsg$MPTCP_PM_CMD_ANNOUNCE(r22, &(0x7f0000000140)={0x0, 0x0, &(0x7f0000000100)={&(0x7f00000000c0)={0x1c, 0x0, 0x0, 0x0, 0x0, {}, [@MPTCP_PM_ATTR_SUBFLOWS]}, 0xfffffe7d}}, 0x0) (async) sendmmsg$unix(r1, &(0x7f0000001e40)=[{{&(0x7f0000000240)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f00000002c0)=[{&(0x7f00000001c0)="76423d5f1885570b6227e94341cb5d7a7462015693fdeae1907e8e202b89eda40a1425", 0x23}], 0x1, 0x0, 0x0, 0x4040840}}, {{&(0x7f0000000300)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000680)=[{&(0x7f0000000380)="b3cccb746c7e8dd2f7a7a6ee49e8a28d7a6a92897aa4f01c4e935f72ca59b9e0a3ddcbe3f3ecd6b451d0f035955023b7937d7c63561dc26ff382f9607e5e0da372eb197bc02bacca1fa3005d578394c5c802b07ef94b31aa35bc7911005fe28311b42d8d3aac33804a7d22075aca3231995daef6c86d881968cc5333f4397de6eefd9f4d7d52d61dffaf9ef29ba5f3a7c4a976fa5f2e4bb2ccc5cb35", 0x9c}, {&(0x7f0000000440)="90da79fbddceffb56ba41c009fea6287abdfdbd785ae1673e6272f7d72ac19a089d71b8f382f24dafd16b8f9c5ff39aad9b2ec2b78fe09895f97e2f5bc77e733e970bf99bba1bc25608bf3c6c2ca890dad28ffb496542c9d87edf38166a987724d96696c414f4bae501f9f91d23d0975a9fa614f6becd105da5a2332cf2ae971061d", 0x82}, {&(0x7f00000005c0)="ff12d3b4d6cc3ba8b93a4df6b65c5bdc066d5fbd6dd0d11590bc49e2a23baca70bf9a54719086a55a8a4037cac5ccc7313d0cd80ba915abbfcce1fc6713b3d4ef85cde2bc46084bd0cec099b0cecdd", 0x4f}, {&(0x7f0000000540)="4943d8f7407fcd40f757d64c10e4", 0xe}, {&(0x7f0000000640)="0bf93cd7a56e715857caa3293b2dbefb044b21c5523514f7508127135f", 0x1d}], 0x5, 0x0, 0x0, 0x20000000}}, {{0x0, 0x0, &(0x7f0000000800)=[{&(0x7f0000000700)="285be241345d36945b97af0cd58d70ca268b1f2bea7ca5f48c2347bb1071931a2c2cb441520ccb96cadfc48dc32a8bfea3cf48b882cddc75d1e769aa56e23b43d834b7d78ab911892978cc74f271493c66875399e003c63543ea0ae705e7e11e23c94c5c766e24935dffaa589415d8d1e4a491374a4b793d5c79c5cd15cb0285abcbcbd8edb7ddc8fc650c8d7e39517da3c3d9bac545077a11010cdecbd6406c7d2350bec9417f043b49cc5c7141b64977d139f2a0a3c4276080a4a49e5c993004c4b7f2e9792cbe24d24e17c6cdac79d8d0d08d16c39a14c82cd5d0f0f78ad452af66011ee2118ef29d56", 0xeb}], 0x1, &(0x7f00000009c0)=[@rights={{0x18, 0x1, 0x1, [r2, r2]}}, @cred={{0x1c, 0x1, 0x2, {r3, r4, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {r5, r6}}}, @cred={{0x1c, 0x1, 0x2, {r7, r13, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {r18, r19, 0xee00}}}, @rights={{0x20, 0x1, 0x1, [r2, r1, r0, r1]}}, @rights={{0x18, 0x1, 0x1, [r0, r0]}}], 0xd0, 0x85}}, {{&(0x7f0000000ac0)=@abs={0x0, 0x0, 0x4e23}, 0x6e, &(0x7f0000001d40)=[{&(0x7f0000000b40)="ace2", 0x2}, {0x0}, {&(0x7f0000000b80)="c48b23f5b351c5ab75b27369c9870b2080afdc131bca41c1f8ef2d650ecdeaf7829ac4d1dd6c589f92a275c11e8d281ec3d5a11bc7275a0890d71ad0fbbfd0628f68ed4c4b777d415cec4367a6ddb04fdc460535d774fcd1ea739187125bce6c26e9b568e30af7cdd0e1c102ebe350f13079f5a5745211ba1f243b9c523f4ec230c6682958f01ffd46734edfb4d8ffc0bdd4e87cd4ba4a185e7c32f908246fb6b4c99cc4d29770df551fe6180658965e791b99afc586166b9b376c1fca9d35c584ed6158a58210faef7c72ce7a748cfd310fc16da5f0d36df7314d41e21bbdbf", 0xe0}, {&(0x7f0000000c80)="cecfe6435d750907ca81630b06f15e1ddc39a961688abc70cbfec1ff89ee6e29db835f18ca9d7cfa9b9eca2b2c1523702a7edaa82becc01be5ad14b2a239741b30fec95c7d89e511ef592afa113ab4f8549767e12e6ecae8b61bf5c30ce38c2790c746cc5bcef1d38551e7fc4e9b45b885a53ffb8780051941449f06610a6a108eb04218406815638bdecd9a9c9ba7fa0219588665cd947767", 0x99}, {&(0x7f0000000d40)="f0f4caa0f3db8620ba8d46f0101acaec4138678175d62ee34b67413621132ffdba1eae85424b6dca4b266341e2ae1817e51d739e3a2b5e758855e60fe07858db7cb42ec35b8322ef339a88cbf0975f112517cd9d31d15dc2f1e3126682ccb6ab6d7d2863316a0a54726043b11ef83b3a0077732000ce9363070e3216a1219a8e92f35f85a4eed7949313bc6034479e4db116ed1a1f7bce97042b5da00410c4b5e7799d8e82b154b10bbe916246b29a49780d1766ce62fefc1230eda124c5591fd6eb3a8771e257356f906cbc32ac53b8229542ed12c4a870ce07b109014904c8c1d9bf3b6bd76beb713af46fe3c850bf8711e67e2c58f1041130957a7259c8c41d366078d86e71f71807425596f55ae1b2c595e5681b358fb092a816ebcfeb2c7ba925b3cdec536f4038aaa9ddfec47fd90123e45cb6483eb4beb8eb533e7c853dcfd2db8144222c456917e9fbc4c9490bf5a9d437ad94dd0e57d6251e30e62141eb80f6f521f15b096531f11a5bcb76684165a538ebb422b3ee933cb94368f2eeca1438f3983e73b827f54b1cfab7f807ad8c0e4b2a04432fca7fdc2bd17d63d19e7846bfa28cc987819d2b1a799cc1e7f73a0a347284c2a93c0cd245617059d3b0a6e536b4a55b711b38b82287d05f11350817c01b43c0c02c0037bbf4abdda690be1e7172dc730d4b024da7cc7530dee44aeab9467f97d73255d91cf77bf2f9fc2c177cf5b56f8df74285d49c53daffd12fb4104050a066e2eff9708cb966c43eec25acaee7905ea1662d68d3cb39c8c162aa590de8dbf04551ecee7d0525894412b2949fcf88a7ca2244c921d81821e4cad5f5d9ca88f9dae07f1abdb2b93dd6969dafbecd4f6a5548e640e81c275aa96df6cc24053cc34b9c53044a59869417c6a8f0a6b6ff49173075ead25a4e0f61684e63510dc7c99f37d6c474d534ad3571e8926e1a38ddb8966f748ae3a27b9e3ca9181d647dad99e720f7d024093e0f5a6b743ec7f723ecee2bcb819b71d147d0a7b4ca18ad6aa431d9d9bd15625370baad0698b0d7d37d8d347e59425cf494aadd9a400ae19f6beb075d6b815caa7b73cc9c0ef88a5569959c8cb641f3001063fb797bf2c79f7b98664a2dbb2ac63958a1966876c030d29f22a6c444dbfb072bffe9551eb6980fab67776d9b8d1387fccd81c4f85915e5b535b53e3d8c3770c23e0a85a3e0b93c45949e92e5f64e815da24467d1e3903c2f7674749f647ed2bbfeb8e54d7ee1126c33b62b1498d07f5d3ae2209345756df47951c9c93388b8fb235e38e46bbe020284b46657981de23dea92867e5330e088329a8dbee029570aa895264b2990b4ddb59f5aac258efb9531ca09796de3df8a027cc3478d4ba4b5935729617ae5b628a76e33738a3be67fc43ff33ed0d460f04ed391b61c10e07f06d80bdbe0a25259b806fbbd14fb6cdb6a7f260991623c970680f92d6853e586a73e21d0a2c7613bdafdfc1014a4a7f699b01ed663e3d733baee1a8dcd5fce5119bcda2836ec31095a1752a1eb6bba269c72c6c92c029a5d5233ec9d7a715e72b614015cd08c6bbfd12c3131368964873f34bff431d7fbc21930f16ce4c77e356de3950737c58036a0022ddb64a36bf868ada075ac20473fd4078165dd5a3229b648cafb848bc75ae3f656ef53779a6b8fc22ec18c35e047cba58d7f787ce8b732c45d0e772d1160a2f04d46fbb6c0f5ba666375b57ed89a1e9dbc8206cefcf16255d0463cb87f328f4cd54adca4124ca5390d28ab91162c1379c98283eba434e3164cba07700b36b4c2a8c5635281c81174f343b99838c2cd42ea9e38b33815f388e4a4b72a4a5d8fd7a4036b78ae51fae60618f38eb58d075b996de25c754eab64c8cc410815b00c78047de13e1a7c57b16cb1fcb3796588b459af95801597c1dc7ca86898eea2babdb54948f5e52a0ec40d8bb13921074b886d253147ec5f00ba268794ad70620c47ec1b321e201fccd65955380965a2dc654708d76bcf6be2becc570dbccbafb741a30e79645700ab7b4024116940fc102234e326c33ad87515660e7bc87b74e55d8b83cdfada8bfb9995409fb35c88d4178fb1c5ce72bfae9fb4e157c7ea0c5049b25553852c37d482e916bbaaadf884581608d295a035abde1e79e6ad939ae079d161a065288ec76a949febe8f83ed0428088efbb5d2a5e38c01cd36d76ed6228ef0f2d344969e0fc01528797b54cd9521f324e102d42a8234780fdaab75cfb3bcf23a9c8dc398d9ea299a0a1f863839d2e218ec2bd2055628d4d534e0d316c83716eab025b8017e7e60d0ba52ff422b4ea1e9c2d7b8d692f95af521375fb05489882c9d83e86101f918a6b6be02b5127d5b97b016f39dca6fe90416913c0ba2e06327491926bdda888f95aa8191de3a3df1d80d7d70d15acd711b07c29122841546b6c8a4b790c350a97e6778a94a4d7ca704a4c30486e91a1cd061fe9fd8b02f3f35531c6da6d4f1560029f07bf13d2af8fb086afe27f9f7e6ea1820591efdb4869806c002ff7d9a0aaaec6f45f3f6ef84ade881d5b5225780bccfc489b2e1e897c3fafb01791c1ad759bac08948b1eaaf5de2926abbfb0433e0a07ca4599ae9dfcd6615dd37bb64cf21213d6cad6c4bd520f4dfdbf4fd958a909acd8ac262b629174108acf9548a646e58c11da745622aa21e3174e08fb8c9e385503beccc7b856974a7e76736f47893d07a07c579fb67bde3f26e9ead9ebd292801ceca355a5b6f7957ec7eb4957fb43bbcf9b7414dd9cb2892853485a51f191484ad4db99ec28126502e5d22338894ad183759a966685b259ea4c3333bc9e4557da97dc91f2fd8e6b6bfc1ec8e6dad76cc2c662d20bca5f249ca99939a6231ccd4ebe01b6e128f407865b5ec0c48fd8e27f15b423be6e2dba4476455de81b73293adabdcaf1c644b81319f5e86618bf29851db3bed04f4bb26f5fb745a3053d85b859de3d74fd7b0a61fb2ea9c6fee8b8ff4e9dafb76f92769609b29a2f33a0a6aeb744a7265db838dcc529e379449202eba9eea7b4524cd13be90755647ab45c7a730364654fdf6a4df84b0a4d1435eda1dc11ebf90767c58c279abaf04533f5649cb726c26043ba78b4878d728c748ae49f8ad1773478b5b5d32a7896ab4951ed3c1fe682899e0d0a403b2f675c97c03c6c88c6b789558ef26c1568d28b9cbfcd88bd33192f5ef0dffa6bc4bd270a25822bbeb14a6dfffea448a21d582744ea2d7cf569057f5cf2907a51213200cb36ea3262e1d362ffcb159c0835d1740b01f4a856088f2bf9783351ab4ff5cf93febedfb1883b3c13740082aa97b56d37a927101fa6b22580128ddcaecca6a5a44445cb2c8dcd24832bede09aac176881ed4aef2ae51937db93a68de35245def99405ee253df104e5d5942feaa2ef4541da9d1ac77723f402fde9ccaab5cbff26bdf00592ad5a447805160ca88e48581b00323b8baee7d068c59808dde7ad02590abf3a617478d2ad407ac88de0377c394d370bcbed63eb4403844b1bd0bf7b375b2b62924191a46cb65c133bc8cea0b19a7e0934411d4fd4d40566dc5df110923d3c242774aef86e85d9cad3579750a827122581646742037dc69091fa189ce31386a291d306b7098545d9d4a212c349160ac6cb4f09e6c801b1e7bf40be1b468f768f3ebc8d4185e3c5668568ee3a276d959a70a485caa77e711f4a90718ddaf9708ae22b7112491dbb261655f0c372f30347820db5c12b0a864ba05e3cef4b0002fe13e85447f12009ecebcc6eb908346fc3b9170335bcf9ee891751baaf6fc6f751d897c66ea7715ca1124eba0885de16ce4306abd54fda0044b56f8d31111b4fd0eb431d7a46fece4ba30d1a4da808c203e187d2300a8a06a3ac0624c97f24f626b2803fb0bd7fec8ca1804b74d7b2ce2ec6f890700fb9aea81d89dc048e53768ff42d48841461fb9f21cca32505f7d9f44bead4f2b2e739615846e40cda8cc69df7f898e0f8b951f00827a562edd04fa688135ca5c514c580083cfea9848583938c83d95e74bb6dc65091cb5bdb141734c358ecec3265de302b428ff95fa998544bbb1284c84cc6133ca2b8b578077565f3e7aacd04d9970c484956bbf95d256628556fb48d41a78e8172d347f040a95906a2144dc02572b53ce09c01b7c613204bf45316aabd7b9921cdbef15996d617f88bbc4780344e0b6eb8f35f4dcbaa435b112b2f003cba9bf1cdd80f1f5b0a79b6047af80ded8cb29de9ec13e2aef6e841d28bb7069eb9f0469b4c65a31c3ebd9ed67541ada74c97426c807131c8b760dc203c097625c6573766e1f804c18e023ddcfdb04aad6ad986cecffc2fc5313a68c3bc82332dafc6418fc31d591197928a5ae4dce47d046cc1c8a3ddbd1932e2e71ab8c85b22270049d7ddb044038eb7d4a9d2ad337a21492441a25e9918393fa972505b710b30b288c293382072cf19b8b3fd15028fdc580e543e9bc9fea90638fa9fb45249b26caf5b0d9fd20423f0b6313c37b63e2966ce408d1ad667cfe71694e39930858cf71d2a1bc36e4d6183c4d1befb951bdcc529925a98eac920b942699d05643c5b64605e7c665251122c5011d3573a87885bf92dbc231021591e610b4c6c1916d9777e0ba2f4475172afff440848b0f528d042014bbc2b9b6d175565451cb30a485ee5fd7505504897fb6788f58ce377697bc6c597f213539279455b5644af19f80de6a01b7805f1c913c23af4c70d39d15bcd6659424250384a13ca0aa25cd109a6ed0c3d7b8f8b47f8da788bc78109cb448925db6b04ba9283e41d2ee60e2e194216c766a855730c842a7d640343aa6c88f86ccf6dc4e70fc6125094f5ce537fe57fb4b72c6b19a4321a959f65c818cabba9ae0591255598b0748f90b1f2240579f89a9c692eee9b9340028016fe0eb5b73aac247b6c06a2ad3fd65687475c2f1968e248d29f602f14a1ed37a1dffc48a3adb61a76c478b83b82c0c66098e11509cdcc74ad63b97b60c076580c928bcb15581a0d5abb5b71bea3535bb2052c8e4a9d01f2a55302b265a311acc6500115b717d95cc3d1fa8bf20c92b7f6c89e125a499ed4b64e1ecc487ec05463323bcb96f7d3b4fd27fbd90ae64df2fa66c62a17f71acbfd5f810dd620880cb84aa8684c136d7e6f4347ec2419178201ec476002aa6b1de02c93bdd060e2cd78947807a010e5c4966d87ae7448e6009830f271ae18bc89287f5c7a9920f72d7973d3c041ae239174459efb8c4cd25dfb88ef181393865f4994b40c6d59463cc42b146e29ba9bf83581672d4405905ef2212fc03d7acde545ed3c205a613dec581624b43e6ea77d12d04967680f40a28aeb9269f8feba56ec7ae7686e6dbef87c9fc148e4495be64b1433a191f4a41089911e82e84073dd7ba81cdf368f614ff1cee2f1b8dcf799dd6e74bbf4cf41171102ab657615933faf14e326dc03e0eaf17d50f90c8705c2e5d8a83c2e54ca2d6d7a0cb249a299c7fc4f99aad94fee95a3ef8f976c1843d6f798cd99633b65a4bfc293064e498873ceb8da4253ac0fa3667956c327c8425e3134be0d022cb6277c918c86ee402c5ec987369b70af3cae0dd1014e3f1472ffcb79b7a78d642b99894a7c79ad4eeb9e21008d96d605d799df055334a2396ef163042737f18a882d4bf2445685f021434ea98deb58c26a93a19d8eb7b9aa4a3baa2dfc0233abf533baac960a34f225d3e92b757b81fdafa66aa2b64d12a6044b7b411298ddfdfe6feba1a53a417ee5c41539b0c6e19475ba0f98", 0x1000}], 0x5, &(0x7f0000001dc0)=[@cred={{0x1c, 0x1, 0x2, {r20, r21, 0xee00}}}, @rights={{0x1c, 0x1, 0x1, [r0, r1, r0]}}, @rights={{0x38, 0x1, 0x1, [r2, r1, r0, r22, r1, r0, r2, r2, r0, r1]}}], 0x78, 0x880}}], 0x4, 0x4000001) (async) sendmsg$unix(r0, &(0x7f0000000200)={&(0x7f0000000040)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000000180)=[{&(0x7f00000000c0)="a38885aa15e75c4726a46f09eae15a9710a1d3ef71b6bb9e779b32ac723b9d7a6392f91a2ac844faecaab889db2873b2d39d0f57acc25dc5e52d25a4b2b23566abb083dfd40b523b7d1edae4b52c9cdf40a82e2e2115b34379525cd9626ae9ec132203333f81fca22b6ac15b4ee1e664665e8028b77bae6ceb2be81e47b047a53c65310e65f4c9496358f2905411250922e8f099b326cbc8c3f1e025f0e21d4084c35315390473", 0xa7}], 0x1, &(0x7f00000001c0), 0x0, 0x40}, 0x8000) 00:28:48 executing program 4: fspick(0xffffffffffffffff, &(0x7f0000000000)='./file0\x00', 0x1) (async) r0 = fsopen(&(0x7f0000000240)='virtiofs\x00', 0x0) fsconfig$FSCONFIG_SET_PATH_EMPTY(r0, 0x4, 0x0, 0x0, 0xffffffffffffff9c) (async) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:48 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x1000000000000000) 00:28:48 executing program 0: bpf$MAP_CREATE(0xe, &(0x7f0000000f00)=@base={0x0, 0x0, 0x0, 0x0, 0x24, 0x1}, 0x48) [ 1728.355451][ T5956] virtio-fs: tag <(null)> not found 00:28:48 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x11}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:48 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r1 = socket$unix(0x1, 0x5, 0x0) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = socket$inet6_tcp(0xa, 0x1, 0x0) setsockopt$inet6_tcp_TCP_FASTOPEN_KEY(r3, 0x6, 0x21, 0x0, 0x0) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r7 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r8 = getuid() r9 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r11 = geteuid() r12 = getpid() sendmmsg$unix(r5, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r6, r4, 0xffffffffffffffff, r4, r7]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r8, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r5, r9]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r10, r11}}}, @cred={{0x1c, 0x1, 0x2, {r12, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r16 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r17 = getuid() r18 = socket$inet_tcp(0x2, 0x1, 0x0) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r20 = geteuid() r21 = getpid() sendmmsg$unix(r14, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r15, r13, 0xffffffffffffffff, r13, r16]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r17, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r14, r18]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r19, r20}}}, @cred={{0x1c, 0x1, 0x2, {r21, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) sendmmsg$unix(r1, &(0x7f0000004840)=[{{&(0x7f0000000000)=@abs={0x0, 0x0, 0x4e22}, 0x6e, &(0x7f0000000380)=[{&(0x7f0000000080)="567be99646fcd5e3d7bd0a56649fcbc6607ccb5695a7d70bcdbd632e141d3c35a98b15e23be5b6b7312f6c6bc3010f077d94b84602d495c0630b82d7d371fc3c32c67e9bb715c943c6e29f0f5e2e07347693fece84f0e441c34532a8dd4f0d02ceb6786a416044f162e37f1fda571b2e7c2ece7359e08e5a15b448bce18a70198107680241399975a59eea78f6be13da05af61ba57d96875d6a870db68785f35b0e07b9b5d8d8824111ed2a8fef5186da3e9c9fceaa3487fbfcb64d5ee7daa3f4276be66398aae248e729464683ad8156eca99003d4a9e3d916dcd2c556500bd9e34e608c4ec53cb3604", 0xea}, {&(0x7f0000000180)="976c3657fa7ad6ded2c0f95e42bdd5ebf506c44719f0c0bd2f4d1d30116b0fc7eea02c7c8102d961440fc4d8f6f4f8ca9d796c7a9cbb3337e0166da993a6d94e", 0x40}, {&(0x7f0000000300)="b2c5907cab8f8d396ecf6efe7d45fe48965cf01aa1bf92b10d798c05645a5033f019af96d862bc5da6b6fe8192420d40f638bdc7114edca1a44100fc025f673f76ce03861a1d730e9927e4e172c13d67264bcdc2053bde31c8a195c77a302d51847518f0e05439224a61d2da590d02d584e19420", 0x74}], 0x3}}, {{&(0x7f0000000200)=@abs={0x0, 0x0, 0x4e20}, 0x6e, &(0x7f00000002c0)=[{&(0x7f0000000280)="8ef9180266f972b2736860f93102e171cf02135609f38d07195b76", 0x1b}], 0x1, &(0x7f0000000480)=[@cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xee01, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee01}}}], 0xa0, 0x20010080}}, {{0x0, 0x0, &(0x7f00000007c0)=[{&(0x7f0000000540)="31b43cb1440685f7cb8612cc780b34d84ccf89f77eb0c1e0a6f74d16bcdee455926e7329c934fa51f77f6d6d3f93df477efe2fef649431dd3f15441e6166ebe207c3d3f1e9c0719023ccd385b4b5ee0de797d5089466066d5ce8873f891234d384f55c297a7f0aeee3c5e2ed37beb0ac071deaa58650306f38ce8a4c4605b72bb85973847c9595ffa4993eddd422c958507becbfea0aab7e3f5ead3eb0d244f604cea649e215910b5ff53d1a48946e9e91f2fdf2a00124006c8f27fff4f0b9132461549a3c1e9fc3751870e1ca8a428c7b62e30d3d4d9bfcb253c9f274c286ac8b4c79de4be48ae8519d671c710d7bec5d07e443ddd50dee4b4ab8", 0xfb}, {&(0x7f0000000640)="2cbbc7b130950ca26ad95ef01d97bd720fd10f4aa7b95bc23353543efc622993b80d5bcb7cf5d0491c818cda95daa5d4d90edf99004b7a513c60adda5db3093dc0bda37cc13a3f1daf667814433e2fb09cdb541375cd8d5dea668e90caef0f81cafbd84cac0bc11a8d2e93e0b96e7993a719fd8cda03286d0c13c5cad1cf3623db15f436722e11329a0f74fb1901c0d171ac757869b0278fb502a59780c76c7ec13445bdd9c0d2840a6a45ba95409b6b4964fe066b80cfa25f1f9276976509f393da7444baf915c15ee2ddc5fe3c4a59344ade9012256a464311f7eda42898429b060b56d848", 0xe6}, {&(0x7f0000000740)="460d06bc31a26fad376105f055ed6d9676577ce844ee47b01af0b97360788990d04e08f1b538d94cccb3db477b890a0e2bf8b7ddf72984b200e848bcad9f5b4e4359fb157324c50f4dcb30638c3beb586a026f0f08a697b60123a964ae53944fccbb7e515c12c16b93a9761f78fe4cbefb5ee5a888b388ab", 0x78}], 0x3, &(0x7f0000000c80)=[@rights={{0x14, 0x1, 0x1, [r2]}}, @rights={{0x10}}, @rights={{0x28, 0x1, 0x1, [r1, r1, 0xffffffffffffffff, r1, r1, 0xffffffffffffffff]}}, @rights={{0x28, 0x1, 0x1, [r1, r0, r1, r1, r1, r1]}}, @cred={{0x1c}}, @rights={{0x38, 0x1, 0x1, [r2, 0xffffffffffffffff, r0, r1, r1, r1, 0xffffffffffffffff, 0xffffffffffffffff, r2, r1]}}, @cred={{0x1c}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c}}, @rights={{0x14, 0x1, 0x1, [0xffffffffffffffff]}}], 0x148, 0x20000050}}, {{&(0x7f0000000e00)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001f80)=[{&(0x7f0000000e80)="d066e962a6efde319281f7f437c58a98d738281909c64833edc278187550e52d41b655a6053a3b2be889e22195e84033a35ddb5617a3835fa7235b3969ae01c63f5d90a24c412dabcc82d9fa4b2e888c127ddfed31a4a6c176061740503068c0ca740a754f800e758f6e3f3f98c678d4608efb7bf45f55a63f1aac", 0x7b}, {&(0x7f0000000f00)="4716a65aa6fc85daaa76de84e3d0b5c167bc1b4f4580dafa2ac06c09896278e0d0", 0x21}, {&(0x7f0000000f40)="718f1421", 0x4}, {&(0x7f0000000f80)="a04eea35fa5a188d967aaba32bb71edb9b1b7f20bd4eccdd0ca1387c2b042000ea48e0f028c46a2359bcb529fd1909bc8041eb148ca7fea5a49040a2dfba73b90034e4a601a11789754faa849d22d21ae889e7d12b34de6cc78c394c2542cfe1f018e431e08d42d85af949641c90c41aebccec293da18a3f5316bd89f3b99edc6448528a1b74b42e558c2b9981c0c0bfb7dd1b1b036ec88344765c2bd4cdd92a95f7c966a083103a8dd4ea26ad655045e49d44704babaf742444375a1a5b4c50585cb8b31ed9e5fb4a06c64f261d6429b021449053b2f2d29cf0728e4e17fe5756f5701ee64e424caa459e87ab0fb0a867b1dc6077476a04478d562b4b4b83bd701f6d9267fdf3f27e98b22e78fc2f79fbf3f575626d61b3be8f158f0abae99a595e3eb83966120b107d543a94a46856c7f0f9cc4ab0bca76309eb6e198a6e4d745380eee4ef726554ebd64e338e288372401b8b76fa0918d97cb1f2c25558c38d8479f8db44d94a63e2a8dd5499f36c4e95198a964ac23f66732379a6d2aa86c61760b747f23a5bd64dbb1523211963e69c316eb0f71ff8c304b505b9db5811ecaed27d35f17f360f1845d7efa680b529ae24eefd1b547724ef7bda05def7d13dd8d409fbe383e89c3e4dfda34d490fc970d208994b1a9b51f10b018b5105ee0cd1b72dd087554faef3081611adbf22b3905e1e1fad8ea0dd4c91e0b457db328262fc0be90459e7e335befbd10cc189d4b16f9efa28c868fdf4e7333ded8b9eb93a8cbced36cdf0ae398374e5b0fa0346b39d5f004f0c6cb20d032ee40b82ed3998a6bb3306cb0bede3e3e598d9f678d6d7cb60bec54c77d6e3fff879000fc8d25c7616e58e88a092851fdcaae130f3de0da43647a721e6eb9b5ed4e1759adeb52f86414f006d1cfc1c9c2092b08f2ac7b12bed5f4698dfeb17ef9bd096db1a97ca43320fc78095486d8852278cf5419ad380fdb1145e47c348646a1abe8870cd0ddf057c40f3ae706d53976ac9f8e29e290212aa8522d7ff5ac740b7412aad581cfbed7c4d7a6f277c8a8dc14b11bdcb27ecba9b990a9ac50b1edde3ebd855197df6fea130e71c75c8723f45059cc0cfc7d348b798787e2b1b8fd8571e5d19e68d648d77e024126ec3440e50a1173cbffdcbe06ee0096ae020c8d28ce40850f8b6d183ae439cb6b25094ae3d95b4ea9e16f91d029e0b1399440f2f301c9767974469335eb2b8cb0f891178f4a9790abcc7d1d9834741462fa8142997937b89521dc5dd9c5ab332633fd5f3784acb62ceaefcd0f20ca8d126e78bb5aa01fd047d479f0fddd2abeccbc2652329f81a4d5bb356b92c19e05a11f9604aa7805083c514cc5b428b5ec1296b26b9040e2ad606e69875503f7f5a2293e6fa6a154fbefaffe14a04e5e1f924ba70ea62d0a4114a898aabe49c463e73dd8244681a5dfe6b9ea522ab1ad01ab20be30d8463a002033e6bf319f6a80c0bf0a2771d3dd6da99a4d0ce2b3a8115bd79193226a02e278ab895fdd9917c75037ac4cf605590230d762b9c67c42fb85de936da7f4c06bbe156d4f1313ce379034331490519c8364c9b21e2225c85ef48573fe13bf2a56c1806e2645b261ab5bff6bc4e66804833444884bf0f5137f07fb21df9da9196c13def7cc52da9c7653f666ebafefa3705fa746a30d6bd1896b50acd238fa07ad4db49a85ed52a4982cfcabadf2438c1eca9d1c93c47b5b972df30d60a2bd8732e906a4b140363714a5d873f082e9eece0068c28e876bd5b7b76beb479aafa3f176f25ed41909dc5cd10145540813deb3cdc8dbc9a7365318eabfc2faad05e5cf8d1eb9c56370a6293621776fdcc0cd46e0fe243e8330da4197a5f870d41162074039cb67b544e11807084bad073504e864385a6124b5300a4f28e07861a71125c4bfa31c635725ef5c37cc71ac558a21f38d7547622195958417583774ea2d997fd20e3d87b9adbb00ad7309600990b8cf97d9459457f39897803050cf07dbc07311028f29f19bb18104784aadde3e28f770e722a08c95f79b6d415ac4fb2e5d51bf3926ccee2d0b1dac439fe4027071118457c07108df5580ce143af899e427db47bab4d8389fbe81bf45a6e9ea661c54b5648e76eb696d58a327ab752fa01bb8f035c1af51f8f826301745e7ba7ffbd5f0f35ca133c89908f76387f389b8a46ef3246e9b353bf42880c47d72a20e2210d9e47dee296cc8bfe89d07a0cf7e3da54a8589327b368c6ad6059f22291d51a74e5e6ba613647363be06e1ca118629810ce318bae9192748174d270a178707efbf11ddce1068c1000a6894b7d655aa65da4cb6e6cca256448e4eda0b4247c723cceef0f13fcb0982a230956372ec096448f7a6eeac1f53c043dd3656281a1ca7e5cac8fb6cfd11901472104f6e49fc3e54c00a367ec4362bde71d1bc618e932713dd6b4c954d119091aa675f0e5932018fbaad82e2b43b54fc82fa92fd1d187c5383bba9eba797b9b38ed35ff35689fb4c81c7590361bd63cf9ed78b289fffb30f93433d80db5448c312503fef57bdb06076d0af45dbf595b49565db7f9883418b96e8e9e6dc69f735e1773af9d4f738c58b24a4253e129123e91ff19f8d7ea667d91cf51832fdf8561da03ec61a84440612677582603574b9121f05a8c543e26cc68a4f0df203e70b01ea3cefd714169984755e74399ef2877a4565752d46a89a18cff7c11e39fc3d0f12eb7d599800ef24a435e257c211ce4f47da785b5031beaa2c6c2eb65e8cb0bc4eed775ddedc5af22e20115c49495c861e75d9903d79b3ac322bdc04584d8405e6cbd9fb0e7525516514fa405dbf039b1a60efc718c9dad55ef5767b4bcb954d52026ebd90298ed2bdbd35ad28ac320e2cb18e2c4e0ae27f3253ff0afd1aa51a63d2b975f1aefc2660315a9bf9009c52f652a06069420830e7d98eab6cb3a9ef9dffbfa773cb0c306f2189a56e293cdccb51a11639bc9d43f2100d8296b5c8df6e3f91ab466edc29886eee10592e1f9bf55abeb1f0854762b3ab4e1b792ce84105a8be5d244706cb51cf8b022a84ab88849c4b11f7edfb0762aecb3ce9878f8f1b04d3a2c2813e135b4b2e4d890001dc313c08cf00d65f5adab79a55bb8018b009595091ec6d24d755058545fe65c9d7c656647f3d9dee28c03e88d26c53b1f13df4db73f617c67a830ece2258e69c6e113913a12ecb54165294d003c2ffffa8fe8d5ce92b61c181c40b04da4b8f98cb938a1c7b5585f447de2aaa473d4d2e525d9b0b7a6a92daecb5b1182536cb9c86c6c5514c9d157c5772f57e6c3c8abb03a1f4bbe261f8fcb5b3da8dca005946a4828b1cd823c74eb7da5ac9215f11a8102cceb82b01864e54ad0571cae65654906a86cf816e45f21f67194a42691c306fdbe4a26c7260fc1812097733a12dc24e1c8541319ebfedffb5389bf4f9507612b37e71424abd692fbb97dcddde1d609f70ba9242493a6a4b18737e1c35ef2da21af7569d0316f8b6c05d27dfda749237b5a22ea4744923bf266307892d0977b414a62d7885d7126387446817c4e3ce0756b312122c80309662442d081e95e5af598e82abb20de0085c34eca215813f378376e8bd2ada1bedede2b16356fc4bd7a91ad7fb57a0881277ef76f39099c33cc8272b38f41e3f0d0881c56072cad17342a0b05fb1c266d17c7d35cc2b3015c56481bbd4ac4cde5bab3a45d985a5a6a8a6e148bcf1a2aa0c651ff7f1dd812fc4d1bafaf635e40570a89c106ed59b70b863140cc1306a6c5250b1b53342f56f36c28c451bbe7a91e3f1a0b4e3f75b1d504ddc2cb50c9ae05d113eddfa1e263d31746484aafed7f16b3fb92a739396ecdecf1265d8a12563f2b9f683d74a458b130cc9b947125a79bc620e68a0fcb2b5a7d063ff52fb8b14757ea165320dc891fc506fb8155680ee8502427151af4472e90c3a55e4a35aa38bd18091b618206226db524d93820158bacfe68aa8223e313890d418c51a5fb170f36328bfc920d4023e1d8a159935d95850f0c1e347529b013193ab1150ca98c21bd6b41e570f28e2ac5651df1b15d8b9765c70f8bc1ae91452c17ad7466dce9256b00481591e614a32c36cae32e5ea88271f67353efbe37d5fdd35559900d0bb21c0c1473a6212378dc460a9d6bcc1ef47cf2de1bdfe09112b8e6eb714ab2045be8f6ac54d4dff9feb5aa4408c4b52baa7ea603d1bb6465a002c8f96214d2c9e60a17400c9e83757559ea3c41ea344cb0bf448ece38ab566f7af362ba81e728bb5d94219decf00d22d11a59e02d97ac9be0aa65e8273eb556ac1a1f2c2cfa09a1cef104f6f1fd3e3a3ae8caf2f0da5bbff5facc79535c27d1e31ff50aa8ac82907e7b6651cc465f6eabf888d0e3e0ed7d3faec45c930e4f8b86b961ffcab99609cc91896c71559e1c443a5c7a19150fb1638eec2a44cbe72713f0203ecfb1e24fe62fe9689eb426c9b4b7ea5b1525bbc7518136b8d94f61a551aceb45ecb0852779e0307cf0228aba47b163973d80dd94cf73144121be564e83e14126c4d9e3e6c7d8b4921d3c36da2e7531af3e2f6db800468e06ae3f6a6812417a5e33657b7da9fdc3dce3b8b1814fb22c57a36bdd53a811059ac3aa0aa6b625fd43e8b05958cb1378be252b187e34ccae4a64a02642c7abd4964ec2f4eae49d6a238e3268c089c04b484bfb56fb9ff6ca0dac42ad0433752b3e65beea2f609872dc482f05fb28326f71d2058df46d77901ab24372a277098b327d26c465aea12741822b37d89392cfa1e5b3d35587080fc90db209ba31e1ac7db3a0e662c2a384922452d328ba7ef06f3166399c4bec2941545a5de1b890b20c4cb23e340a40088f562c34f1fadd7ec9c51a73437df2057fa2813595339714a5a5a135a3f4f9c66f40caa29bff10588533b0e91cb4a1d3f3844f7148aed7647640af8262f2dbea988c2855d9b8495b20edc2e2979fdc24a9bddbe7ec731d51b29a1b8eae54a40bbc64ae1ef960bd85acbce722e6ecc0c3690f07a4d54aaa2f589eb73ed064203df9fd69fb0f3873ffa87648c0aa9529a52fd8badde6668c9107adb19c58c49859dee07628b81c7481cac1da34431a9dd95d1095e0c53fb600a81c287408e7d0321101daffd03798a678228b489580078255d82ec61882f97bf7e09b097f6fa7353bc41754d8ac35d89d7bf2da204a8c95756c8012c82387f843cc59381cbb41f7d1edd6ebdc5d1ee27f53ab20e92c9311ec7044a974a0bf3912d714a5f9fcdef2d038388b0227bc0486fa44bd7ade5173df85507b052aec3b9c501ad7fc0cd4e0a2a5669f5778973f022b6678361adc37ffbd47efa65bef0135704e6833448a211159430261cda49942d524d5d65fb83c493d4243ed6c74fea66a1235a6c82ee68743f3ed6caefe3c4c95d4916c4696d07c296b21fc55c779aee8e1b0d3eb0f87d894543887e4296875c05e32eb0bf0f0fc4abdcad1c9210e4574cb81ea6917e0c20d6aedacd8b87426d334a6a9d4961a96b1281dc67c0b2ff53a16762d32ea2a118841ee50d8ce42c6f23696a9763ab4e17d711c874dd7743e5b66223ab9bd69dea5727b3afa92f8ca48594da5934c5ca914607962e5e85060a51711ce8b54d49317a71c3d0770601fc5ab6a2c010ba8ff8242f3bce4d33ab726e1a7801e63e39832587835841a5b6c0289b723e9361535fd024757e82ba48bb4699b8377226c8e2cf139e983d9b8c74fe392e526490ac4928fd835f9e1493fe40ac8cacf26277c13dfde43fec5f93156532ee9489b216ad7", 0x1000}], 0x4, &(0x7f00000008c0)=ANY=[@ANYBLOB="1c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32=0x0, @ANYRES32=0xee01, @ANYBLOB="0000000030000000000000000100000001000000", @ANYRES32=r2, @ANYRES32=r1, @ANYRES32=r2, @ANYRES32=r0, @ANYRES32=r0, @ANYRES32=r0, @ANYRES32=r1, @ANYRES32=r0, @ANYBLOB="18000000000000000100000001000000", @ANYRES32=r1, @ANYRES32=r2, @ANYBLOB="1c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32=0x0, @ANYRES32=0xee01, @ANYBLOB="8070ca877e8818fd9252099af9873286aac8ade610c1063c6add9424eafb335afd0f52535abc3a844053df4c6fe5c23a5e83124b76c2641ba74e9a5c5bddba5b1362b969e5d7856cd5d4d958995271dbca427dd0051afa579afd03b41606cfa615d745dd82ea0000", @ANYRES32=r0, @ANYRES32, @ANYRES32=r2, @ANYRES32=r1, @ANYRES32=r1, @ANYRES32=r0, @ANYRES32=r0, @ANYRES32=r0, @ANYRES32=r2, @ANYRES32=r0, @ANYBLOB="14000000000000000100000001000000", @ANYRES32=r1, @ANYBLOB="000000001c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32=0x0, @ANYRES32=0xee01, @ANYBLOB="000000001c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32=0x0, @ANYRES32=0x0, @ANYBLOB="0000000018000000000000000100000001000000", @ANYRES32=r2, @ANYRES32=r2], 0x130, 0x44}}, {{0x0, 0x0, &(0x7f0000004640)=[{&(0x7f0000002240)="a4ae76b6b84231362a9a33bd41b604e2494e87d8f81db0979d3cbc16f12ca0e79165b377d6b0365dae884d7bc4ab3f98994328d66b71772f8c97b75e72820753334dcaf86936", 0x46}, {&(0x7f00000022c0)="eb0639a70af5fd076d3b70a0c8929d0a40a9", 0x12}, {&(0x7f0000002300)="93217a2ca5f6404d768a7191406e601d8dd6e81b2ba89aa1bccb1f540744a4d392cabdd624cad2e2e75cf9898dded5e0c94686eeebc30fce28d64c529ff437ea80fa33e5e555b3fd667b6cc022b7d6f8e495e2f4b897893ea106474036af7a914ba086", 0x63}, {&(0x7f0000002380)="77e16df219f2437e10339a70af99450673fc3c5079f53658030dcc3b06cf9ce157157fee2551b46e166b4d6d3a392fa4119ba9e498974fbdbbd37716fda6558f42f60dfc9514bd5329a1927faf832a539e59809b124bb9782ff26696927c3d0c1b2435717cee222624439269d36488ae7bc5c7774978bd15dc5d691f9850", 0x7e}, {&(0x7f0000002400)="95e46b96e098508eff6bcc13f327b923cf772a3649e9cfbac2eb075f63dbbbb2fab69a61bb9b076282f2e3f0605b1a0c444f2b30b7b29546fc02a616bbcf6e6d7c84f16103cbeafd33efdb50c2e2140ffd094c7f38d0ae148dde23c26dbf246e9bd42a44da2ce02fc760577c453843119695c1f7a067ae83e95725c15a1679ff59ad13da55ad2e15d9b6e3dfdbf8021782a1ba7501b26916ae22f3fce6637164a3a25ddd24305fc5e07eba5b72aacf1d17eeb2e7e76dc614827073b4986c77333567a2eca581e96423a99060d159d7a4a3737d0cb434c9a35e7f156d279be75dba4caf689318f21848c8e952c747ad02acb39bb5b6dabc2561d4f3814c405c3b50c90420013a281efc0f9dfbcf7e610d37559efce92dbf5669450f7b8aaadffc5405a2c46a072b86bea43456e73bd3d8639154d71ed18c34036a8449a4d831cb42a1fdb47e32774086959be93f923bdfe4ab108818623e6d058723291abb4c4aada91c37867277b9098ae690fdde5fec0038df6aee8dd644c2e8da83223b799187210652370fc7cc11bf3f40ed02b11abc24cd68e1995101a7a474ba04a6cbe5fedf9b4402e0874ae0f5a8e1e7a9733f8c79eeb07d18684a45dbb29f40c85ecd84925b730a9f43bff91753686c30f5b6d53cafe22354fc7bb0c4304db44fd1b811958c70503345fc3ece2fafb32e0ee39a0bc5a49feb1bffe483ceb9eea99fc3819bd9ed6924f12772ce0e2cb1f052f8e535629b6fbff6fc96120d43505fc1581a33eab3ba3c5761e9e8f593a5de49397cf018ed7ee3361ded5536b193f356f80ed50d42f944f6d0c66d3de8d15253f793b9ea7672930ed5a9a52eec958a84d0c35042f12168f03dd8d29b1650d4de1bc2d7df6620562c3aa411ea496cb9915aa39b9a9b06f5ef5937a6d93ade3d47bee0b351be05cc55c3ceb8d5fa0f5b6f2d9fad86d133cb9785f182f4369d18b5c9ee4a5fd314b1f4142803b716a00c170e14b7bec8dc5e482faa9d58f8d884d9f5624589fcdab8232c161ed9ad7fe52f4e63e81604905269d63193a30b46b767535acc0b8e22b94af31f31cb15122592b285f8e8ead5feb0fc7e9ede4164bdc6b6b13483c05a30397397e682a798292488608822071d582319a52672de9d280364cc973f9fe784e0b5e1ff561c2f8c3c03862c59ab0570fb7fe6d7a7fef940f503acaa8bc31cb56958a5b9ef89292e8ea4c197d9ecb09ab20be446a0fd6bf698bb26624d2b1fe1bf5a83e74d25c030d03321659c496d4bdcebded967d1844091f4ebf872b86d932af3c3bbef7e1310088a5fde67cff07c0ddb596aa1158d4b3b54d9b0369e3371389c0d15ab0ff1e96362720b81e5468c6bfe9f0e94b5f500f91847f322ac55b3a20a1884217a34511561166997c10ba18434ec555fc6b03000eb733150f79ee254117aa64ea8b2129c6632f5100e6ea9042ae8c812b2be86651f2c73a866023d48f6d7313e1f09c28a4b490da78d609ef0328bea4eefd78ddff9c40f6b455b19d3d3ddd674272d6ee014fa59078fd22c811e8410ae0a4a3d70d60e4bd1f8962170df03b59095a89804779e8598f5e80ddcdc26b3e6a07a7b2673d9233248aba696481df18af15d2c4fb3ee111507687642a943787bf20309694d2339521a2ca468ca40158bfac4db39a0caf93a6bfd6463f06001a3c0d166079b444a54d963ff51b7808756369f111bffa41ba2570724af106ea82cfce4d005263a5271a23c72c60f0f8958006e865a9cb3e18e28a2440cd9b1f2bf00b87c0ff51467279c8e616553ecc2506f221f5da099f90284c3356845936fdcebc067db8bb192c925bf63be2c2e9fdeb3bfd69579d6eb3c84c44a6df2942f2916a4e2b7ed90d209814683cb06516598cdd407db43e2db14fb98c5f0056781e2bbdf95c23ae105e72a491d1fa0504f25c130ddf0f5cef735a7f7471dee7ee2f6f5536f4fb5ac2146f325218174c2a85b7eca9f0386fcfb2a4fa1cfc03c6f90cf281960911c4b2eb0fc8c482c24a89a95aa7f3f0def7a7ce4588b27234fa1e814843bbadee54b55bd572067bc08aa4f8accee600ac9321e370474150b36fd2085c22d38cc4261f147c239c77e32f8bcceec2cb403268bd69cdfed195ea5e9775adc41007b4694259afbbbcc38cf94a070d210c0066fd47b3718199b3cbebd8ed7ddba5e743750cf9835f6fbd2755b3f7fced524c94e9633770fb12a615d3bc354d3b4b8eedc6cf898913767558577b5aa5d68b943ce5727941423cbb7b85cb02c48fdd5f34f7433d3b2deecd66aabb52994cd7cfbf530f11356d44d4708696302e44895fc1d69775e38e53f3df51fbc01ffb8deba98c254afa69216530d5e1198330ca998c8966741866974a39e13e3b8fd96f9758b7553be83290d67b13457a4e0e8ef161b3b4b757b7afb072604327ebd1d3f327d6be34aa565fc43e2bc90f0054815d4e3ec6567e3ab0b5e28288811e2ee7debbb5376f9af387b70c1aa73afd2f695022660def1c88356e39ebb74711fb87aa0453fcf90975f1fbd98d138614239185c4bba28559b186b52bd4614c577eed55d1953f3279f59a50c401e873c639edb4a04a008d5b482783a70e794c3c25a0e55a1befb213bd64441067737b562d978af57357fc42275881ea47aac1f4e685b758c4876b896e746326ada88a1eafba908ad87f917d2b9927a76a84716c1badd0ed309a5fc23835232d22448d3a5e270cc4b6bccd6f5295e1ca472f7df6c3ae8ed7ba3ab62d94c597a0a1ff59f9893593456905d7d08a049aeb7015a19d4bdd8e5e51371f26cc297ea8508fb27d22d6cf5de9fb1d770c6e670068b180fbaa44b9cd0df17b508b26b48cdb9f7b02a427f9937e9047385f865a0195b401d2e1204857ca344377cf0a7a58587d0f9adda965feb9afc26b1fe1b2867e75c044fa243e18f69f12dc9b123a150decbd6f2b0bb18be0b9ab56067d7d704fdec746818a101d5da3ce7c2cde01e23e7417105c2d655eac658f6dd080e13d6a7e80ac523e0e783d36d98debeb247dc325289d9acdd93c65ce0acae92b9bb8daa5a56770ec5f79fe0ff08c9fd1af0b639fc1bcb7aa9b97acca137338ad9d7aa87cfa63c5f253e2472ffe0f9c0e11678559e3fdd422d3761ae765d41c623c325c0498a6c158f4e6447a3b45155e851171b033f616abaf64e5d71aad8ac008d09410afae0b0e705b8324e5769861af4cf18234b1b643d143237219e1e7d55095bb171f54724b80b0f8aa7879c935f16d4700020051d2820c1c718c0b178f7ceaab7058d16ed0c14cacfe4a28bad42fee5f8d7614d657627cdf5d1d5d8d40f4cc9d8f0aec5709b766b340449eddcd7998278acbe6afe1ac861b4cc056b18b07fd9985961297f631b94b957f056fcc5a6e50e082a43dd802ba2ac2809b14910708df0f7c978af0af6e6b9848005549efada14e6bef45258ca97a7b5b38b41388bd74bfdbf76d9cf4a7776eddc0d77cce04fe23c3180b40f7d74a337225681df85902574eda839b7b97fd8660d4587f88d9eae613386151d3cd55e5f766de4d05ef95f32c5f93e788dfff459942d2adfd2b15215b0dd507d327184084481c7a3bf66c7f790850351b8b8dd9afbea77f7602951500989a0590e9b6d07af28e64117172f89d5dede030e3332ba858249627dd7a5dfd8ba83e3eab37ca65468721e4989fd07910698014707e4aad987441ddc05c1891bf1450ad25f6fa78a41537cf0f5c295dad06f83976ca6eb49f612be3506a991073bc0df4fb691e008ef71252d8e411358cd86bf109594d9f2bad5eb614e59c7ea7933c7c801806ee4e7dd2ca3f7c6884ed696542ea4d48f78806b5f5c7319bc02824c2eb911d06ccae51c8ba28737ec0255609c9251e13c16e009c8bbb8e2ae4d9a85cc7e166084159ffe438538f804d9424b0cbd1dd467287488d48798f67bfd3fa2f7f359a12b8d1f06018a5981f95110038ea1a57165020462dd1b140506d4d2b23bb3672e3bdbc10e3b1d2072dbbc1e18eebe563f215905113858321a6eea3f99d5516ba4dc933283d66cbfe1a0bf89a2889c1d01d40fdb8266ee61ad4823f543ad337f2983c857dd4d788d76b7d03fa950195d06f7441da1695161f5e39bf8f98f455f84669cc1eee068ed0c37a8f548d4aa1f9a336bb631700a2665a260bd44ae8c056669b2c7c921b43a3e078c5b15a933fc06ca5cc1bf76ae1ba484f34df15fb6e271d9fa9031ea39f4cc7be581e668c37268c0c2c7abc370b1906230e0d58c3e3f7aebed5ff17956acafac27fda47bb2379ae189693df91d991fb7f92b9a15a464d287f26aac13acd295b6f1d70eed6f69477e3905ebe6516c96b379061f800245dd93bc83e4e76e4189f2ae0d8a6cbb8bce3b8063ba9cdacb82a561d23c55c31cb5b68f3ed66f31b765e06829b826d446c4380abd115588bfd4927d0f06d5fa20264655d77e42ccb5583dce2747fc651119246d404e3e76c9de80407098b7eb8764b6202b5eb98080ebcbc17afc07c2771a0ae9e6246bb70dcde4f22bd8338726c5ef9cded57af9072df7a8f9be8bb571af40f1f7e4a52676c359f8d165e809a2a4910caf3eb06b71b17f5562d088cb5ccebad00c4e67f1e286d4b6a660351e555271af34ecd851b2c5e830b69fe522d2196da5c24a53c7af749a255ac46c7d33a67fc96795e0f1e3226d8d907fc9ba78f34b085dc90886c58f98393e06c7fe3fd6116356918e6e88599908daa2b2d2e482f374c54c9fa051f0d78d33d908821c10e3da7fa63fd4d2cf5be7addff9729884dd2a17125e08283362200b20ab807572515811bf9baf09b826a2809ea87b6b86284bb8b49a56d2bc2c108162fa63944cf5e67c7b6c5ee9a59b9f595f570d2aa8bebc3163b7eb9befd60b5a4f4d726754ad3c6cbc912a99ae8c6d95b4745089fad5ca1f44c4dca75050edeca7f20adef47897ce455d339d2e9399a2217f45109ae9aeae188f1261544c784ef115e850c36b7cbeedb754fd1a2ec9ff7a0a54b988be27e1a4f6feb65425a574652ceb893bcc346ee4ff8d05b600f53b04f76983fe034089dd44fc956815bb7fef35ac2c52eddc394803fd534074191fa6c93a670099d80e19023ffeeb8979de66e0f14ccf7053653ea24c50fd17d431af8c3f1dfd9a4d1878c3d9f38ac98782c3835d45207647d72bbf319ddf24392ef89014c8f3ab5743f8100845bc322cc201a1ee59fe42ab2fa023dbe0c46b3a9aed1d8632d0cd58a0185537e18e066ddbae2373c3e138fc92a1956f507b57ffc381239437a05f6a4ec24450ff2c2b9c746aae9576d050ac4861117934e41f7834e80c04240362b8e60dbc27121e6f04325a0d06aac6ba140d679910fd2d16c77455b48d025aad51a029a5c521e20c59a13202311aa9259cfd5c9efe2caefd1f64001a9e9ac92dc319140b76505fafd25a230d9a0e150ab27f3949b5ad14bd778b08ba927ae623fa2ba139e58328f71a7a55c41f8dd9b0cdc4e76ab404e61b4019c489b30ab975329db80651c0ea2372999e2cbfe69cc701e2f969a7fb611759074c7efc519028d411dfb284c03518633ff6f7d0ebd95701e8d1669961dec50cce74d385704c7207f8cf4074cbb7519f14a370399715ab4cffe0ec28073d76b34cd420bbf8a10b4f0ff0a9bda2b39b62206e02990689f87029e9714d0925cfab3a1de6191b9e1b0961e2b8e33661f5a0da2f4ca1511ae9de34526ba5ccadbdc863896738af6882ec77c5533400d4387116c2c81be2af97190d32543e73ae50a10d16828c2c", 0x1000}, {&(0x7f0000003400)="e1c138454f7d9be9c6570bf9931361fb087748a8a4e4f852f4da691d35f7c4b8cb0f4bab0afa8c0b69a1f00a55e663e1e7e2f53700b1f1f69585a97ea74b009172bbe521", 0x44}, {&(0x7f0000003480)="90a545b012efd7eee3179d460c08c1e05e769008b0bde9dd50df27f5ff0f101c97483cf9aa4f3f6c621d14dd06c5344f370909315fd084b2651bea116585ea58a473af662055e09fd8469a1b9071e28b834ea0e38d373163c3c4c3bae0b2d0c5f976d03266db6fe8e37bce21469cc7ecca5bb083cb8db2e03b693af2b70a3c5291addc0b3f185d4a4aa984be98a8a19b8f8ef8b67e711e738af7898a4d34d31a987787fd2b19e765158d0c35a4b057d3fbf33a56ffbc4119e5dc5182546b9e4bc50cbcacf95c7d9cbca5f33a6c2a6ef5d0d1d35b2ffcf5cb4809c49e5ac713e92aac02e013e4ef0cd120651f49d821c100611d2ff2650337040b8ca717190c0b932898b7c427cce8fded11b912cf0549f190c0e34bdc1c9a08f82213976bb8943b82832b16c6c89c5c8929dca313cd88a8a003cf74917239970356fa602ebd4fca8cce28be388051774c859e154ca02a7269fc6371daca133d914cecdb7b955424227b60c92afe3346bd2e6ab541ca2acc45f4cc84f60444fac22d53e17cb52a2cc152816ea5e4b4c80dd0d35396d048210580617b8b8b654319e9f8894ffa40af33fe324d941c6a726be3b2918baa86676c9742f5cc2536740536cad7574bbaf36af6d45c89b4b243cf4d9f8c793a4a86e2f2d5c7987786fd9f5dd670ce1f0ecfff1d74ac4ae6d4ff9a74696b0784dc4db3143a8f7c67dc84977399201d1193d8c22a3911a1fc1d443660cb1ac66aad19aaab25e3b39a11cabe2657088f6aee10489cb9775e3e18a27f41cbb87c630a8815eba1600e8a81686bd5d5e79ea0d9f14a900e255af4532179f98dfa121fb4e550364cd00a3812c43232a6a640d831d0b7190f5248c8ae2adad758f49d5bf4cbb52a2c1eb2074f126db22e7ac65e6273093e9f1bc36366610fa0341916ccec3436f2b151f2574acfb21e35583eef74bcd06ba1b75c2b63abdd357a6a6cdb7d8f50168c697ff7e02f3525c3646250a4fd379ffd1e64b508b243277d679f441615a4edc1610f69f01c7dc105aac8ccccf762ed31ee25e3c22f815df57f126aaf9cc8b69ce5cabd71cb8925c29d3e221c53d4b03f5476e17c6990cc174977f10c86121929f653bbe2f996b9b437b8112fbad556a481605057b6bdf9928176853ac0e3a3cbb66c2a8621661e9b92cb658eb68bf5fe85501a5ba1cd8b1fae095fa03a0f7983732891f95fae2111a7071db3dd7e07fa739df336ffd6ec4476ac89cac2909b64eb6ee2d1a64dee6ec3ac3197764fe140f2e64f64d93c4a5641474f060276d8b3308dfdfa69b80a0062d7669679025eb0e9d244cf9655503481f96961ddb933dc7b816f69f4f8ca10f500c1009f0ab0dfb768b33e9d5e9fdb052f8d1de9f20323f423a8afcbe09609f91cca70d767255c6646c331fdec51159b7a11edd178c368bd798b26dc197ae745795219e29bfa043102352fddaf3bb64e48c48989eb17a60b7aad56829f5ba46f9532e5767de775b3aa06feb14736452e71039711acab1e3df9e74df613946a61a185543ade584f169d412a2196d089a316a8b341cac6857bec2c9ff31466ab3dbf9c386364498b257045d62fbf793bd711b5375d89a3eaae0eaebdf5252df5d4e19b9a125e4a12c229878450e718a7ea27b6baac61f0414e7c68b7cfae1132f4e12041519284fd0d25e11e7462460413828b4ce6af4f2fba22e40902c9fbf7745a2b3919f6c88624442f9a44cb50beae51f48cb5c936cf92bfb7c5ab508c19a26491002628911ee7578909bf6b14b020c508fea6c919774fbe9193e41c21aa538aefcdbd3d197ac5e9dc36041193e55320f7e16af828638328e264650273a0199eecc2c028e9360ac1826099a91cc243b3ab4789643f49851c9b227c734ce10dfa3dfaf9fede9a4320bc16cf5409137ccb51221cf07a3e1ad30e0db346586dc7c533fb87f2091afc5c69952fbf2cb0232ecc234b7c8e3fdfe0523f6ce873b0f4e8416f0db32e244bfe66657aeeb8af6c591b5b7eb62a007fecb0a8eb6e83ab7350b56d45a260769269d809845bcebf3b54825a51ca15f7259e1e40cc9673ed0c313569caeeaac3ccae59e4dfee0d72a4fe24edc9b55bb4f40715d6e7323a146bdfe86bdb960a29b6e1239002f5a6ce61f16a9376fedbc2ace452f3c2c96c7787875c8201057a0349948d27e83550a3a4bcebcd87f11a94fc3452d12ece2289e5142097ec1bd8dfcc36677fc305de79d0a6bbd6b5c34cdd59b3704e4030b316de0e2483afafc1a58f494e990fcd88e85c1ee70f73bd679b13ce469830493b4e8dcfe533be17f4ab404f03ebbdeffca7b58c4ebce69f981594a9b3eb358a639b45551342a132f241eae6df0486a32da122f780521bb8d16b1e334c5d790f591a51f1ae5639d606409dc117d13ebbf726c11b03f7370b27dc63be08cb1dea070b7ed78ce6bf9a563a686b744dff46d14bffc16893f26dd959d17b6b4e0a0de4e1533f98364c14c81dde7b8b65be00d9253b229994cdccd859a51e98f33f13279aaf8132bb26fbfbe9e78d03c75608ce7ff913d34332503435252e9d8ff0e42cc4fdcff504c450773fd82ccf969d7f8dd0dadaa0d66123ea6a9f83fc12cf4f1e80dd61a6e1ff46e23335e3bd30ba54e1d3dce70dca4cc826106420dc32b53178db724beca54d8c382c85f0054ba9c32f8e6f84b14f2ce67a78d423b468ee30901645a24e0ee5d0d6a3d617094d94514ca74ef19d9da7e07b775e85095b77d2843b9c60f27950fa53158bda491bb18222d6e5a214cb47b4142013accfcb179729bfebf5168bea40262338cc54388b4a9a2a3a4afbd7624eda361bbaaae13673f7c533403175462092255837e0e63dc3bc503a6f9a16ec7614470a8f2477d441ab4d75f18fde372d426ac27d271d4e0ed363ee32dc83863cf3218805ced05dcaf6dab02f7c9d31dc1301e37e127d7efe56d0c221bdbfba1f046c47ed02b7c9d793f8afc2457c6d6b4ca04cc214ff08761d807248f51de0db088f1888d67ff926cc685b38690ed0c6cc70a4f148c46705d41c975e19739668203461d7c13419cc4d15a38f1c9e6a04c647af3ba3776dcdaa60bc017803d232ae305747664a7d48672bb672744c4c68860bc1ac1f1829ffbae15983706536b89666e4d2727a65ecd1b0a6cad5d2e23c60547016988cbb143edf4b3ab1fd2b8355bbcdecb889c3efe5e0804d8d9c0a5590c51cfabe3085641519bdb4934d2e188bb058a49559f596dd0d41e8c5d4ecf3b6b20921606232cc686d6879e9c6ea5f09a629cb4945085519513b7b189c1edd775b8cfd53ce172f15f59c19c26f0b0b36e57d4ecd1c8a62dfa9105b5809d3823ce3e8d253c0b7a3fccf1b4957124845c5f1bee5b554123010cd01a4041c18dbc4f01d1c502c87c7a5c67b2e7c9b5f05fbf19e428139a994cad2ec8d63ddc30ea8f960316e5eb3a69c27ef43657dc1556f65caac4e72c68c4d87c263e3774ab466e3131477209f5e1c4312c33dc1157379ea2c9ddfd449a1da2d2d4628bef26ee51b56e91d1c3e947b4034571a01b85cb53578247f87bd4b2a84819a4897cd01272cd2d7394ddb109733d21eea357c85cdee4b255635030a9fb17ab02de447812d178ba2bf46dc186dc2b055b30c890113ca5fd8ed78fa5b8b54627f05d307108342737e04d48fa280539c36527beaf913921536de89420ef54936cd8e098f9d308e1d5e350adf47b21a18aa714580e8c947720c4a280262f072c0ab278f408bb0886fe6020362e593ff00de288b580bfd4cf5ebedadf67da256f5f8ccae787206df70b0bf0037ff658df4a94c1dfcc87aefe762e83836479f4cea35c0d3e90df5bc27dfab57e9e40cbfcfb780c45253447e6ce99471466bc7777c7bd10d88e1e31ddc746247da1ebe728c46b9bf2613a347f3181d37e88baa1ed051f685067fa464711b9cec5467ace786fa9e1bc97b317e733921b795433b4da5466a379d02908e9b6a3e7f35beebf934ea16649cdde811eafe1c2533adc8138d85289649e95583e1f765a88c393ad78e147c30575dcb502f2b07c7c3748301a12ad25cd17abe308dd46764ed5728b2fdc91abd211b4e8833b2b82ba1bc55f403de7f5e38379bd369d36f7cf56acf59ecaa7111ee4a31da53b82c0a597b37ce7a197a4768d066894948ba05dd0c733e6fb867d7b9ae58842e3dcd4a6217a876e84cb3209f39ac63c23b239fdb3316bb9541470818aeffb4a1dd33be87f9b20d64d3d3989f1743619bbc161c38ec3143a5f916301283d382605dae0bdca4f0aaffef247dcdcf0c03786176d1de85b2996236ceb7da75e56f11bf8e6278a1ee6e4a6a71e8d02f5f4592eb59c016b1874d10e30880ab16ecbd2b7a3e88210fec183f5da808d382f83efb081a5abaaeeefd6ba887e0f3fe1ecda8607e22f3b7bdbc71d7ab4a327e899e286878e3d75aa8551f8c5f4fe29e690a56e40a6a6e8c08a3632bad623f4e33d92af3f894440e7f054ab7ea085f4c01faac564b69d881be32d8f17d7921718da558588dbbaddacfd305cc9464407c637cc028d1fc14d255a90cfec1f775e93d75e26348ac921532c1215992330969811f6f0c642601cca6d766bdd08793c9376e9ba4e195a41bf8b2ec8d5e2ca39d63384d896a0d99db43263b71a78ea319634432205ee50a4131e4544c0d204648bff4666099ad187ed29ccf2209b8c4d0310d4ebd1677478461be26cb2ec43b14aecb2e275ca02cdd2803c9e2f190628e34c8318dc119a6df82785ff91ccf1b76c9eb56193f03e564ccbe4e612e8d6097638fe9444cd289a09e44b60ff275c664fcfdd74a200d2613a84e9c8ac3fc52732c047494633f295a853d8d798f4a3f27601ed846182b420d220d8b724d33f1b5d0e3f211fed7a213584d465521d4d6edc2bcd0601f503031dc70fed82965dfffacf050278cbf533c85537327bce213c37170289a8fea69e80a359fa0b8cb6ba43deaf2de895991d1259281d8cfbfd2c5a38a27c7af3b11dbe972cb66d4c7fc0a26ad81078e47190d1ab3dea349603dfc9e9182006bc0fdb3b81931086dad5af2ab812f8aff75e02e3967de162a2267c385d3ea827b9e70f992c692f963cedb7f754a884543958d5d21cefce013031436119426df999dbe978a82688377ea9e287a414dcdea1e8998fe2ef51d1808024506744f77a7c574780ee9ae28f8c2a526a98fe8310a36bfad61b831a40a6e133f30587eb31d854f1572c59a368263cc38665b26f3a0d2da310a8db512534032108378437033f28f60cfca24f1a05353608bea254943f7abd887fbfc5d270c04490903d45a92fb4b5cdfe87e2189fb31479d9c63212fd3f9e8ed55ed8aa917186c9fe9413d4861da5003fd83ff51f1305f6afbe3b4825d15e0afee3f60cb14435f86e7d55303a21a6fb726b1a09bbe4a7a7180f5caacfdd62f75d973c43c7716db111b8e93b001428605a025af9ebeeb9e322077cb171fb969ac7e53f126076b3fb8c2d3f252dd70d0845d40a46fb71d777a43b584c5a3dfef7bdecf75a4bd325e1ce0a1e77b039c5b35cd51fed6ddd93452543166eeef4ad63634b4ddec34d654f2d8b644d13714e8b12a9b22346dfd480762e640ccc364d18dd2bfb7059dd99624bd7466f78637a6f059efe4a0b4d795a06cbd9043bd922ea486da176766b74f5d62344557690a53262385c05a797f314d7d97389abb55d1996d620d6cc923926575e27c4d72e01bf679c83abe9acce85091fa4015eb75339d4952a4d427def31b3b5b0205b38ff151adc769c6aa", 0x1000}, {&(0x7f0000004480)="d70e621257df9fc22b4e6e83af82bceb9a1b84b516e5e16514792c8fa71b54c604fa17ec41d367188778f46e70838e295ae7270629fba2db77e527b8231292ec39c9613a6cea8a8363018905acb3bb08", 0x50}, {&(0x7f0000004500)="45c2579709ade63c23feb0dedf4b6aa15d5b32d6c313e47b203da15f073cd6ad", 0x20}, {&(0x7f0000004540)="dc9cc92e18abc8b34c9e7b1df9d95f9436b6ae0aae2648a30edc2e427fdd8a5aef54f55536a2a4c653d7728a5f84801e35d5e57707eda3f9dd09f0e36455b835a11853aed2da0c8d61de46933f7c16176b0c58c8902e6db8b08acc31b5bee9d0c89734a3e470390289ed8b69834f27a3eba9ac8c1c7bf0ab506193022e32a54b45fdc23bf2f8278769dd40849ff9b08b16ed4e0aee75caf41df32531a15e5eb3bf87404cd86ab4f71f584fbb4b58aefed337529f2243a713ba3d07414279c792d7659c86c9e59fcfee890c0e6a851c0aaa911ad7b03cfc839879c28dcd0454dba2b135039d475027", 0xe8}], 0xa, &(0x7f0000004700)=[@rights={{0x14, 0x1, 0x1, [r15]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee00}}}, @cred={{0x1c}}, @rights={{0x2c, 0x1, 0x1, [r0, r2, r0, r2, r1, r3, r2]}}, @rights={{0x20, 0x1, 0x1, [r1, r1, r2, r1]}}, @rights={{0x1c, 0x1, 0x1, [r1, r0, r1]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r11, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r2, r2]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r17, 0xee00}}}], 0x120, 0x4c050}}], 0x5, 0x1) getpid() getpid() getuid() getuid() getuid() 00:28:48 executing program 3: fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) r0 = fsopen(&(0x7f0000000000)='hugetlbfs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:48 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x989680) getpid() getpid() getuid() getuid() 00:28:48 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x12}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:48 executing program 1: fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) (async) r0 = fsopen(&(0x7f0000000000)='hugetlbfs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) 00:28:48 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x1100000000000000) 00:28:48 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:48 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x13}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:48 executing program 0: fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) r0 = fsopen(&(0x7f0000000080)='fusectl\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) r1 = fsopen(&(0x7f00000000c0)='pipefs\x00', 0x1) fsmount(r1, 0x0, 0x72) r2 = fsopen(&(0x7f0000000000)='virtiofs\x00', 0x1) fsmount(r2, 0x1, 0xfc) fsconfig$FSCONFIG_SET_FD(r2, 0x5, &(0x7f0000000040)=':\x00', 0x0, r0) 00:28:48 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x1000000) getpid() getpid() getuid() getuid() 00:28:48 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x16}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:48 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) r0 = openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) (async) r1 = socket$unix(0x1, 0x5, 0x0) r2 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r3 = socket$inet6_tcp(0xa, 0x1, 0x0) setsockopt$inet6_tcp_TCP_FASTOPEN_KEY(r3, 0x6, 0x21, 0x0, 0x0) (async) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r7 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) r8 = getuid() (async) r9 = socket$inet_tcp(0x2, 0x1, 0x0) (async) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r11 = geteuid() (async) r12 = getpid() sendmmsg$unix(r5, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r6, r4, 0xffffffffffffffff, r4, r7]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r8, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r5, r9]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r10, r11}}}, @cred={{0x1c, 0x1, 0x2, {r12, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) (async) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) (async) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) (async) r16 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) r17 = getuid() (async) r18 = socket$inet_tcp(0x2, 0x1, 0x0) (async) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r20 = geteuid() r21 = getpid() sendmmsg$unix(r14, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r15, r13, 0xffffffffffffffff, r13, r16]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r17, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r14, r18]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r19, r20}}}, @cred={{0x1c, 0x1, 0x2, {r21, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) (async) sendmmsg$unix(r1, &(0x7f0000004840)=[{{&(0x7f0000000000)=@abs={0x0, 0x0, 0x4e22}, 0x6e, &(0x7f0000000380)=[{&(0x7f0000000080)="567be99646fcd5e3d7bd0a56649fcbc6607ccb5695a7d70bcdbd632e141d3c35a98b15e23be5b6b7312f6c6bc3010f077d94b84602d495c0630b82d7d371fc3c32c67e9bb715c943c6e29f0f5e2e07347693fece84f0e441c34532a8dd4f0d02ceb6786a416044f162e37f1fda571b2e7c2ece7359e08e5a15b448bce18a70198107680241399975a59eea78f6be13da05af61ba57d96875d6a870db68785f35b0e07b9b5d8d8824111ed2a8fef5186da3e9c9fceaa3487fbfcb64d5ee7daa3f4276be66398aae248e729464683ad8156eca99003d4a9e3d916dcd2c556500bd9e34e608c4ec53cb3604", 0xea}, {&(0x7f0000000180)="976c3657fa7ad6ded2c0f95e42bdd5ebf506c44719f0c0bd2f4d1d30116b0fc7eea02c7c8102d961440fc4d8f6f4f8ca9d796c7a9cbb3337e0166da993a6d94e", 0x40}, {&(0x7f0000000300)="b2c5907cab8f8d396ecf6efe7d45fe48965cf01aa1bf92b10d798c05645a5033f019af96d862bc5da6b6fe8192420d40f638bdc7114edca1a44100fc025f673f76ce03861a1d730e9927e4e172c13d67264bcdc2053bde31c8a195c77a302d51847518f0e05439224a61d2da590d02d584e19420", 0x74}], 0x3}}, {{&(0x7f0000000200)=@abs={0x0, 0x0, 0x4e20}, 0x6e, &(0x7f00000002c0)=[{&(0x7f0000000280)="8ef9180266f972b2736860f93102e171cf02135609f38d07195b76", 0x1b}], 0x1, &(0x7f0000000480)=[@cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xee01, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee01}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee01}}}], 0xa0, 0x20010080}}, {{0x0, 0x0, &(0x7f00000007c0)=[{&(0x7f0000000540)="31b43cb1440685f7cb8612cc780b34d84ccf89f77eb0c1e0a6f74d16bcdee455926e7329c934fa51f77f6d6d3f93df477efe2fef649431dd3f15441e6166ebe207c3d3f1e9c0719023ccd385b4b5ee0de797d5089466066d5ce8873f891234d384f55c297a7f0aeee3c5e2ed37beb0ac071deaa58650306f38ce8a4c4605b72bb85973847c9595ffa4993eddd422c958507becbfea0aab7e3f5ead3eb0d244f604cea649e215910b5ff53d1a48946e9e91f2fdf2a00124006c8f27fff4f0b9132461549a3c1e9fc3751870e1ca8a428c7b62e30d3d4d9bfcb253c9f274c286ac8b4c79de4be48ae8519d671c710d7bec5d07e443ddd50dee4b4ab8", 0xfb}, {&(0x7f0000000640)="2cbbc7b130950ca26ad95ef01d97bd720fd10f4aa7b95bc23353543efc622993b80d5bcb7cf5d0491c818cda95daa5d4d90edf99004b7a513c60adda5db3093dc0bda37cc13a3f1daf667814433e2fb09cdb541375cd8d5dea668e90caef0f81cafbd84cac0bc11a8d2e93e0b96e7993a719fd8cda03286d0c13c5cad1cf3623db15f436722e11329a0f74fb1901c0d171ac757869b0278fb502a59780c76c7ec13445bdd9c0d2840a6a45ba95409b6b4964fe066b80cfa25f1f9276976509f393da7444baf915c15ee2ddc5fe3c4a59344ade9012256a464311f7eda42898429b060b56d848", 0xe6}, {&(0x7f0000000740)="460d06bc31a26fad376105f055ed6d9676577ce844ee47b01af0b97360788990d04e08f1b538d94cccb3db477b890a0e2bf8b7ddf72984b200e848bcad9f5b4e4359fb157324c50f4dcb30638c3beb586a026f0f08a697b60123a964ae53944fccbb7e515c12c16b93a9761f78fe4cbefb5ee5a888b388ab", 0x78}], 0x3, &(0x7f0000000c80)=[@rights={{0x14, 0x1, 0x1, [r2]}}, @rights={{0x10}}, @rights={{0x28, 0x1, 0x1, [r1, r1, 0xffffffffffffffff, r1, r1, 0xffffffffffffffff]}}, @rights={{0x28, 0x1, 0x1, [r1, r0, r1, r1, r1, r1]}}, @cred={{0x1c}}, @rights={{0x38, 0x1, 0x1, [r2, 0xffffffffffffffff, r0, r1, r1, r1, 0xffffffffffffffff, 0xffffffffffffffff, r2, r1]}}, @cred={{0x1c}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c}}, @rights={{0x14, 0x1, 0x1, [0xffffffffffffffff]}}], 0x148, 0x20000050}}, {{&(0x7f0000000e00)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001f80)=[{&(0x7f0000000e80)="d066e962a6efde319281f7f437c58a98d738281909c64833edc278187550e52d41b655a6053a3b2be889e22195e84033a35ddb5617a3835fa7235b3969ae01c63f5d90a24c412dabcc82d9fa4b2e888c127ddfed31a4a6c176061740503068c0ca740a754f800e758f6e3f3f98c678d4608efb7bf45f55a63f1aac", 0x7b}, {&(0x7f0000000f00)="4716a65aa6fc85daaa76de84e3d0b5c167bc1b4f4580dafa2ac06c09896278e0d0", 0x21}, {&(0x7f0000000f40)="718f1421", 0x4}, {&(0x7f0000000f80)="a04eea35fa5a188d967aaba32bb71edb9b1b7f20bd4eccdd0ca1387c2b042000ea48e0f028c46a2359bcb529fd1909bc8041eb148ca7fea5a49040a2dfba73b90034e4a601a11789754faa849d22d21ae889e7d12b34de6cc78c394c2542cfe1f018e431e08d42d85af949641c90c41aebccec293da18a3f5316bd89f3b99edc6448528a1b74b42e558c2b9981c0c0bfb7dd1b1b036ec88344765c2bd4cdd92a95f7c966a083103a8dd4ea26ad655045e49d44704babaf742444375a1a5b4c50585cb8b31ed9e5fb4a06c64f261d6429b021449053b2f2d29cf0728e4e17fe5756f5701ee64e424caa459e87ab0fb0a867b1dc6077476a04478d562b4b4b83bd701f6d9267fdf3f27e98b22e78fc2f79fbf3f575626d61b3be8f158f0abae99a595e3eb83966120b107d543a94a46856c7f0f9cc4ab0bca76309eb6e198a6e4d745380eee4ef726554ebd64e338e288372401b8b76fa0918d97cb1f2c25558c38d8479f8db44d94a63e2a8dd5499f36c4e95198a964ac23f66732379a6d2aa86c61760b747f23a5bd64dbb1523211963e69c316eb0f71ff8c304b505b9db5811ecaed27d35f17f360f1845d7efa680b529ae24eefd1b547724ef7bda05def7d13dd8d409fbe383e89c3e4dfda34d490fc970d208994b1a9b51f10b018b5105ee0cd1b72dd087554faef3081611adbf22b3905e1e1fad8ea0dd4c91e0b457db328262fc0be90459e7e335befbd10cc189d4b16f9efa28c868fdf4e7333ded8b9eb93a8cbced36cdf0ae398374e5b0fa0346b39d5f004f0c6cb20d032ee40b82ed3998a6bb3306cb0bede3e3e598d9f678d6d7cb60bec54c77d6e3fff879000fc8d25c7616e58e88a092851fdcaae130f3de0da43647a721e6eb9b5ed4e1759adeb52f86414f006d1cfc1c9c2092b08f2ac7b12bed5f4698dfeb17ef9bd096db1a97ca43320fc78095486d8852278cf5419ad380fdb1145e47c348646a1abe8870cd0ddf057c40f3ae706d53976ac9f8e29e290212aa8522d7ff5ac740b7412aad581cfbed7c4d7a6f277c8a8dc14b11bdcb27ecba9b990a9ac50b1edde3ebd855197df6fea130e71c75c8723f45059cc0cfc7d348b798787e2b1b8fd8571e5d19e68d648d77e024126ec3440e50a1173cbffdcbe06ee0096ae020c8d28ce40850f8b6d183ae439cb6b25094ae3d95b4ea9e16f91d029e0b1399440f2f301c9767974469335eb2b8cb0f891178f4a9790abcc7d1d9834741462fa8142997937b89521dc5dd9c5ab332633fd5f3784acb62ceaefcd0f20ca8d126e78bb5aa01fd047d479f0fddd2abeccbc2652329f81a4d5bb356b92c19e05a11f9604aa7805083c514cc5b428b5ec1296b26b9040e2ad606e69875503f7f5a2293e6fa6a154fbefaffe14a04e5e1f924ba70ea62d0a4114a898aabe49c463e73dd8244681a5dfe6b9ea522ab1ad01ab20be30d8463a002033e6bf319f6a80c0bf0a2771d3dd6da99a4d0ce2b3a8115bd79193226a02e278ab895fdd9917c75037ac4cf605590230d762b9c67c42fb85de936da7f4c06bbe156d4f1313ce379034331490519c8364c9b21e2225c85ef48573fe13bf2a56c1806e2645b261ab5bff6bc4e66804833444884bf0f5137f07fb21df9da9196c13def7cc52da9c7653f666ebafefa3705fa746a30d6bd1896b50acd238fa07ad4db49a85ed52a4982cfcabadf2438c1eca9d1c93c47b5b972df30d60a2bd8732e906a4b140363714a5d873f082e9eece0068c28e876bd5b7b76beb479aafa3f176f25ed41909dc5cd10145540813deb3cdc8dbc9a7365318eabfc2faad05e5cf8d1eb9c56370a6293621776fdcc0cd46e0fe243e8330da4197a5f870d41162074039cb67b544e11807084bad073504e864385a6124b5300a4f28e07861a71125c4bfa31c635725ef5c37cc71ac558a21f38d7547622195958417583774ea2d997fd20e3d87b9adbb00ad7309600990b8cf97d9459457f39897803050cf07dbc07311028f29f19bb18104784aadde3e28f770e722a08c95f79b6d415ac4fb2e5d51bf3926ccee2d0b1dac439fe4027071118457c07108df5580ce143af899e427db47bab4d8389fbe81bf45a6e9ea661c54b5648e76eb696d58a327ab752fa01bb8f035c1af51f8f826301745e7ba7ffbd5f0f35ca133c89908f76387f389b8a46ef3246e9b353bf42880c47d72a20e2210d9e47dee296cc8bfe89d07a0cf7e3da54a8589327b368c6ad6059f22291d51a74e5e6ba613647363be06e1ca118629810ce318bae9192748174d270a178707efbf11ddce1068c1000a6894b7d655aa65da4cb6e6cca256448e4eda0b4247c723cceef0f13fcb0982a230956372ec096448f7a6eeac1f53c043dd3656281a1ca7e5cac8fb6cfd11901472104f6e49fc3e54c00a367ec4362bde71d1bc618e932713dd6b4c954d119091aa675f0e5932018fbaad82e2b43b54fc82fa92fd1d187c5383bba9eba797b9b38ed35ff35689fb4c81c7590361bd63cf9ed78b289fffb30f93433d80db5448c312503fef57bdb06076d0af45dbf595b49565db7f9883418b96e8e9e6dc69f735e1773af9d4f738c58b24a4253e129123e91ff19f8d7ea667d91cf51832fdf8561da03ec61a84440612677582603574b9121f05a8c543e26cc68a4f0df203e70b01ea3cefd714169984755e74399ef2877a4565752d46a89a18cff7c11e39fc3d0f12eb7d599800ef24a435e257c211ce4f47da785b5031beaa2c6c2eb65e8cb0bc4eed775ddedc5af22e20115c49495c861e75d9903d79b3ac322bdc04584d8405e6cbd9fb0e7525516514fa405dbf039b1a60efc718c9dad55ef5767b4bcb954d52026ebd90298ed2bdbd35ad28ac320e2cb18e2c4e0ae27f3253ff0afd1aa51a63d2b975f1aefc2660315a9bf9009c52f652a06069420830e7d98eab6cb3a9ef9dffbfa773cb0c306f2189a56e293cdccb51a11639bc9d43f2100d8296b5c8df6e3f91ab466edc29886eee10592e1f9bf55abeb1f0854762b3ab4e1b792ce84105a8be5d244706cb51cf8b022a84ab88849c4b11f7edfb0762aecb3ce9878f8f1b04d3a2c2813e135b4b2e4d890001dc313c08cf00d65f5adab79a55bb8018b009595091ec6d24d755058545fe65c9d7c656647f3d9dee28c03e88d26c53b1f13df4db73f617c67a830ece2258e69c6e113913a12ecb54165294d003c2ffffa8fe8d5ce92b61c181c40b04da4b8f98cb938a1c7b5585f447de2aaa473d4d2e525d9b0b7a6a92daecb5b1182536cb9c86c6c5514c9d157c5772f57e6c3c8abb03a1f4bbe261f8fcb5b3da8dca005946a4828b1cd823c74eb7da5ac9215f11a8102cceb82b01864e54ad0571cae65654906a86cf816e45f21f67194a42691c306fdbe4a26c7260fc1812097733a12dc24e1c8541319ebfedffb5389bf4f9507612b37e71424abd692fbb97dcddde1d609f70ba9242493a6a4b18737e1c35ef2da21af7569d0316f8b6c05d27dfda749237b5a22ea4744923bf266307892d0977b414a62d7885d7126387446817c4e3ce0756b312122c80309662442d081e95e5af598e82abb20de0085c34eca215813f378376e8bd2ada1bedede2b16356fc4bd7a91ad7fb57a0881277ef76f39099c33cc8272b38f41e3f0d0881c56072cad17342a0b05fb1c266d17c7d35cc2b3015c56481bbd4ac4cde5bab3a45d985a5a6a8a6e148bcf1a2aa0c651ff7f1dd812fc4d1bafaf635e40570a89c106ed59b70b863140cc1306a6c5250b1b53342f56f36c28c451bbe7a91e3f1a0b4e3f75b1d504ddc2cb50c9ae05d113eddfa1e263d31746484aafed7f16b3fb92a739396ecdecf1265d8a12563f2b9f683d74a458b130cc9b947125a79bc620e68a0fcb2b5a7d063ff52fb8b14757ea165320dc891fc506fb8155680ee8502427151af4472e90c3a55e4a35aa38bd18091b618206226db524d93820158bacfe68aa8223e313890d418c51a5fb170f36328bfc920d4023e1d8a159935d95850f0c1e347529b013193ab1150ca98c21bd6b41e570f28e2ac5651df1b15d8b9765c70f8bc1ae91452c17ad7466dce9256b00481591e614a32c36cae32e5ea88271f67353efbe37d5fdd35559900d0bb21c0c1473a6212378dc460a9d6bcc1ef47cf2de1bdfe09112b8e6eb714ab2045be8f6ac54d4dff9feb5aa4408c4b52baa7ea603d1bb6465a002c8f96214d2c9e60a17400c9e83757559ea3c41ea344cb0bf448ece38ab566f7af362ba81e728bb5d94219decf00d22d11a59e02d97ac9be0aa65e8273eb556ac1a1f2c2cfa09a1cef104f6f1fd3e3a3ae8caf2f0da5bbff5facc79535c27d1e31ff50aa8ac82907e7b6651cc465f6eabf888d0e3e0ed7d3faec45c930e4f8b86b961ffcab99609cc91896c71559e1c443a5c7a19150fb1638eec2a44cbe72713f0203ecfb1e24fe62fe9689eb426c9b4b7ea5b1525bbc7518136b8d94f61a551aceb45ecb0852779e0307cf0228aba47b163973d80dd94cf73144121be564e83e14126c4d9e3e6c7d8b4921d3c36da2e7531af3e2f6db800468e06ae3f6a6812417a5e33657b7da9fdc3dce3b8b1814fb22c57a36bdd53a811059ac3aa0aa6b625fd43e8b05958cb1378be252b187e34ccae4a64a02642c7abd4964ec2f4eae49d6a238e3268c089c04b484bfb56fb9ff6ca0dac42ad0433752b3e65beea2f609872dc482f05fb28326f71d2058df46d77901ab24372a277098b327d26c465aea12741822b37d89392cfa1e5b3d35587080fc90db209ba31e1ac7db3a0e662c2a384922452d328ba7ef06f3166399c4bec2941545a5de1b890b20c4cb23e340a40088f562c34f1fadd7ec9c51a73437df2057fa2813595339714a5a5a135a3f4f9c66f40caa29bff10588533b0e91cb4a1d3f3844f7148aed7647640af8262f2dbea988c2855d9b8495b20edc2e2979fdc24a9bddbe7ec731d51b29a1b8eae54a40bbc64ae1ef960bd85acbce722e6ecc0c3690f07a4d54aaa2f589eb73ed064203df9fd69fb0f3873ffa87648c0aa9529a52fd8badde6668c9107adb19c58c49859dee07628b81c7481cac1da34431a9dd95d1095e0c53fb600a81c287408e7d0321101daffd03798a678228b489580078255d82ec61882f97bf7e09b097f6fa7353bc41754d8ac35d89d7bf2da204a8c95756c8012c82387f843cc59381cbb41f7d1edd6ebdc5d1ee27f53ab20e92c9311ec7044a974a0bf3912d714a5f9fcdef2d038388b0227bc0486fa44bd7ade5173df85507b052aec3b9c501ad7fc0cd4e0a2a5669f5778973f022b6678361adc37ffbd47efa65bef0135704e6833448a211159430261cda49942d524d5d65fb83c493d4243ed6c74fea66a1235a6c82ee68743f3ed6caefe3c4c95d4916c4696d07c296b21fc55c779aee8e1b0d3eb0f87d894543887e4296875c05e32eb0bf0f0fc4abdcad1c9210e4574cb81ea6917e0c20d6aedacd8b87426d334a6a9d4961a96b1281dc67c0b2ff53a16762d32ea2a118841ee50d8ce42c6f23696a9763ab4e17d711c874dd7743e5b66223ab9bd69dea5727b3afa92f8ca48594da5934c5ca914607962e5e85060a51711ce8b54d49317a71c3d0770601fc5ab6a2c010ba8ff8242f3bce4d33ab726e1a7801e63e39832587835841a5b6c0289b723e9361535fd024757e82ba48bb4699b8377226c8e2cf139e983d9b8c74fe392e526490ac4928fd835f9e1493fe40ac8cacf26277c13dfde43fec5f93156532ee9489b216ad7", 0x1000}], 0x4, &(0x7f00000008c0)=ANY=[@ANYBLOB="1c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32=0x0, @ANYRES32=0xee01, @ANYBLOB="0000000030000000000000000100000001000000", @ANYRES32=r2, @ANYRES32=r1, @ANYRES32=r2, @ANYRES32=r0, @ANYRES32=r0, @ANYRES32=r0, @ANYRES32=r1, @ANYRES32=r0, @ANYBLOB="18000000000000000100000001000000", @ANYRES32=r1, @ANYRES32=r2, @ANYBLOB="1c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32=0x0, @ANYRES32=0xee01, @ANYBLOB="8070ca877e8818fd9252099af9873286aac8ade610c1063c6add9424eafb335afd0f52535abc3a844053df4c6fe5c23a5e83124b76c2641ba74e9a5c5bddba5b1362b969e5d7856cd5d4d958995271dbca427dd0051afa579afd03b41606cfa615d745dd82ea0000", @ANYRES32=r0, @ANYRES32, @ANYRES32=r2, @ANYRES32=r1, @ANYRES32=r1, @ANYRES32=r0, @ANYRES32=r0, @ANYRES32=r0, @ANYRES32=r2, @ANYRES32=r0, @ANYBLOB="14000000000000000100000001000000", @ANYRES32=r1, @ANYBLOB="000000001c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32=0x0, @ANYRES32=0xee01, @ANYBLOB="000000001c000000000000000100000002000000", @ANYRES32=0x0, @ANYRES32=0x0, @ANYRES32=0x0, @ANYBLOB="0000000018000000000000000100000001000000", @ANYRES32=r2, @ANYRES32=r2], 0x130, 0x44}}, {{0x0, 0x0, &(0x7f0000004640)=[{&(0x7f0000002240)="a4ae76b6b84231362a9a33bd41b604e2494e87d8f81db0979d3cbc16f12ca0e79165b377d6b0365dae884d7bc4ab3f98994328d66b71772f8c97b75e72820753334dcaf86936", 0x46}, {&(0x7f00000022c0)="eb0639a70af5fd076d3b70a0c8929d0a40a9", 0x12}, {&(0x7f0000002300)="93217a2ca5f6404d768a7191406e601d8dd6e81b2ba89aa1bccb1f540744a4d392cabdd624cad2e2e75cf9898dded5e0c94686eeebc30fce28d64c529ff437ea80fa33e5e555b3fd667b6cc022b7d6f8e495e2f4b897893ea106474036af7a914ba086", 0x63}, {&(0x7f0000002380)="77e16df219f2437e10339a70af99450673fc3c5079f53658030dcc3b06cf9ce157157fee2551b46e166b4d6d3a392fa4119ba9e498974fbdbbd37716fda6558f42f60dfc9514bd5329a1927faf832a539e59809b124bb9782ff26696927c3d0c1b2435717cee222624439269d36488ae7bc5c7774978bd15dc5d691f9850", 0x7e}, {&(0x7f0000002400)="95e46b96e098508eff6bcc13f327b923cf772a3649e9cfbac2eb075f63dbbbb2fab69a61bb9b076282f2e3f0605b1a0c444f2b30b7b29546fc02a616bbcf6e6d7c84f16103cbeafd33efdb50c2e2140ffd094c7f38d0ae148dde23c26dbf246e9bd42a44da2ce02fc760577c453843119695c1f7a067ae83e95725c15a1679ff59ad13da55ad2e15d9b6e3dfdbf8021782a1ba7501b26916ae22f3fce6637164a3a25ddd24305fc5e07eba5b72aacf1d17eeb2e7e76dc614827073b4986c77333567a2eca581e96423a99060d159d7a4a3737d0cb434c9a35e7f156d279be75dba4caf689318f21848c8e952c747ad02acb39bb5b6dabc2561d4f3814c405c3b50c90420013a281efc0f9dfbcf7e610d37559efce92dbf5669450f7b8aaadffc5405a2c46a072b86bea43456e73bd3d8639154d71ed18c34036a8449a4d831cb42a1fdb47e32774086959be93f923bdfe4ab108818623e6d058723291abb4c4aada91c37867277b9098ae690fdde5fec0038df6aee8dd644c2e8da83223b799187210652370fc7cc11bf3f40ed02b11abc24cd68e1995101a7a474ba04a6cbe5fedf9b4402e0874ae0f5a8e1e7a9733f8c79eeb07d18684a45dbb29f40c85ecd84925b730a9f43bff91753686c30f5b6d53cafe22354fc7bb0c4304db44fd1b811958c70503345fc3ece2fafb32e0ee39a0bc5a49feb1bffe483ceb9eea99fc3819bd9ed6924f12772ce0e2cb1f052f8e535629b6fbff6fc96120d43505fc1581a33eab3ba3c5761e9e8f593a5de49397cf018ed7ee3361ded5536b193f356f80ed50d42f944f6d0c66d3de8d15253f793b9ea7672930ed5a9a52eec958a84d0c35042f12168f03dd8d29b1650d4de1bc2d7df6620562c3aa411ea496cb9915aa39b9a9b06f5ef5937a6d93ade3d47bee0b351be05cc55c3ceb8d5fa0f5b6f2d9fad86d133cb9785f182f4369d18b5c9ee4a5fd314b1f4142803b716a00c170e14b7bec8dc5e482faa9d58f8d884d9f5624589fcdab8232c161ed9ad7fe52f4e63e81604905269d63193a30b46b767535acc0b8e22b94af31f31cb15122592b285f8e8ead5feb0fc7e9ede4164bdc6b6b13483c05a30397397e682a798292488608822071d582319a52672de9d280364cc973f9fe784e0b5e1ff561c2f8c3c03862c59ab0570fb7fe6d7a7fef940f503acaa8bc31cb56958a5b9ef89292e8ea4c197d9ecb09ab20be446a0fd6bf698bb26624d2b1fe1bf5a83e74d25c030d03321659c496d4bdcebded967d1844091f4ebf872b86d932af3c3bbef7e1310088a5fde67cff07c0ddb596aa1158d4b3b54d9b0369e3371389c0d15ab0ff1e96362720b81e5468c6bfe9f0e94b5f500f91847f322ac55b3a20a1884217a34511561166997c10ba18434ec555fc6b03000eb733150f79ee254117aa64ea8b2129c6632f5100e6ea9042ae8c812b2be86651f2c73a866023d48f6d7313e1f09c28a4b490da78d609ef0328bea4eefd78ddff9c40f6b455b19d3d3ddd674272d6ee014fa59078fd22c811e8410ae0a4a3d70d60e4bd1f8962170df03b59095a89804779e8598f5e80ddcdc26b3e6a07a7b2673d9233248aba696481df18af15d2c4fb3ee111507687642a943787bf20309694d2339521a2ca468ca40158bfac4db39a0caf93a6bfd6463f06001a3c0d166079b444a54d963ff51b7808756369f111bffa41ba2570724af106ea82cfce4d005263a5271a23c72c60f0f8958006e865a9cb3e18e28a2440cd9b1f2bf00b87c0ff51467279c8e616553ecc2506f221f5da099f90284c3356845936fdcebc067db8bb192c925bf63be2c2e9fdeb3bfd69579d6eb3c84c44a6df2942f2916a4e2b7ed90d209814683cb06516598cdd407db43e2db14fb98c5f0056781e2bbdf95c23ae105e72a491d1fa0504f25c130ddf0f5cef735a7f7471dee7ee2f6f5536f4fb5ac2146f325218174c2a85b7eca9f0386fcfb2a4fa1cfc03c6f90cf281960911c4b2eb0fc8c482c24a89a95aa7f3f0def7a7ce4588b27234fa1e814843bbadee54b55bd572067bc08aa4f8accee600ac9321e370474150b36fd2085c22d38cc4261f147c239c77e32f8bcceec2cb403268bd69cdfed195ea5e9775adc41007b4694259afbbbcc38cf94a070d210c0066fd47b3718199b3cbebd8ed7ddba5e743750cf9835f6fbd2755b3f7fced524c94e9633770fb12a615d3bc354d3b4b8eedc6cf898913767558577b5aa5d68b943ce5727941423cbb7b85cb02c48fdd5f34f7433d3b2deecd66aabb52994cd7cfbf530f11356d44d4708696302e44895fc1d69775e38e53f3df51fbc01ffb8deba98c254afa69216530d5e1198330ca998c8966741866974a39e13e3b8fd96f9758b7553be83290d67b13457a4e0e8ef161b3b4b757b7afb072604327ebd1d3f327d6be34aa565fc43e2bc90f0054815d4e3ec6567e3ab0b5e28288811e2ee7debbb5376f9af387b70c1aa73afd2f695022660def1c88356e39ebb74711fb87aa0453fcf90975f1fbd98d138614239185c4bba28559b186b52bd4614c577eed55d1953f3279f59a50c401e873c639edb4a04a008d5b482783a70e794c3c25a0e55a1befb213bd64441067737b562d978af57357fc42275881ea47aac1f4e685b758c4876b896e746326ada88a1eafba908ad87f917d2b9927a76a84716c1badd0ed309a5fc23835232d22448d3a5e270cc4b6bccd6f5295e1ca472f7df6c3ae8ed7ba3ab62d94c597a0a1ff59f9893593456905d7d08a049aeb7015a19d4bdd8e5e51371f26cc297ea8508fb27d22d6cf5de9fb1d770c6e670068b180fbaa44b9cd0df17b508b26b48cdb9f7b02a427f9937e9047385f865a0195b401d2e1204857ca344377cf0a7a58587d0f9adda965feb9afc26b1fe1b2867e75c044fa243e18f69f12dc9b123a150decbd6f2b0bb18be0b9ab56067d7d704fdec746818a101d5da3ce7c2cde01e23e7417105c2d655eac658f6dd080e13d6a7e80ac523e0e783d36d98debeb247dc325289d9acdd93c65ce0acae92b9bb8daa5a56770ec5f79fe0ff08c9fd1af0b639fc1bcb7aa9b97acca137338ad9d7aa87cfa63c5f253e2472ffe0f9c0e11678559e3fdd422d3761ae765d41c623c325c0498a6c158f4e6447a3b45155e851171b033f616abaf64e5d71aad8ac008d09410afae0b0e705b8324e5769861af4cf18234b1b643d143237219e1e7d55095bb171f54724b80b0f8aa7879c935f16d4700020051d2820c1c718c0b178f7ceaab7058d16ed0c14cacfe4a28bad42fee5f8d7614d657627cdf5d1d5d8d40f4cc9d8f0aec5709b766b340449eddcd7998278acbe6afe1ac861b4cc056b18b07fd9985961297f631b94b957f056fcc5a6e50e082a43dd802ba2ac2809b14910708df0f7c978af0af6e6b9848005549efada14e6bef45258ca97a7b5b38b41388bd74bfdbf76d9cf4a7776eddc0d77cce04fe23c3180b40f7d74a337225681df85902574eda839b7b97fd8660d4587f88d9eae613386151d3cd55e5f766de4d05ef95f32c5f93e788dfff459942d2adfd2b15215b0dd507d327184084481c7a3bf66c7f790850351b8b8dd9afbea77f7602951500989a0590e9b6d07af28e64117172f89d5dede030e3332ba858249627dd7a5dfd8ba83e3eab37ca65468721e4989fd07910698014707e4aad987441ddc05c1891bf1450ad25f6fa78a41537cf0f5c295dad06f83976ca6eb49f612be3506a991073bc0df4fb691e008ef71252d8e411358cd86bf109594d9f2bad5eb614e59c7ea7933c7c801806ee4e7dd2ca3f7c6884ed696542ea4d48f78806b5f5c7319bc02824c2eb911d06ccae51c8ba28737ec0255609c9251e13c16e009c8bbb8e2ae4d9a85cc7e166084159ffe438538f804d9424b0cbd1dd467287488d48798f67bfd3fa2f7f359a12b8d1f06018a5981f95110038ea1a57165020462dd1b140506d4d2b23bb3672e3bdbc10e3b1d2072dbbc1e18eebe563f215905113858321a6eea3f99d5516ba4dc933283d66cbfe1a0bf89a2889c1d01d40fdb8266ee61ad4823f543ad337f2983c857dd4d788d76b7d03fa950195d06f7441da1695161f5e39bf8f98f455f84669cc1eee068ed0c37a8f548d4aa1f9a336bb631700a2665a260bd44ae8c056669b2c7c921b43a3e078c5b15a933fc06ca5cc1bf76ae1ba484f34df15fb6e271d9fa9031ea39f4cc7be581e668c37268c0c2c7abc370b1906230e0d58c3e3f7aebed5ff17956acafac27fda47bb2379ae189693df91d991fb7f92b9a15a464d287f26aac13acd295b6f1d70eed6f69477e3905ebe6516c96b379061f800245dd93bc83e4e76e4189f2ae0d8a6cbb8bce3b8063ba9cdacb82a561d23c55c31cb5b68f3ed66f31b765e06829b826d446c4380abd115588bfd4927d0f06d5fa20264655d77e42ccb5583dce2747fc651119246d404e3e76c9de80407098b7eb8764b6202b5eb98080ebcbc17afc07c2771a0ae9e6246bb70dcde4f22bd8338726c5ef9cded57af9072df7a8f9be8bb571af40f1f7e4a52676c359f8d165e809a2a4910caf3eb06b71b17f5562d088cb5ccebad00c4e67f1e286d4b6a660351e555271af34ecd851b2c5e830b69fe522d2196da5c24a53c7af749a255ac46c7d33a67fc96795e0f1e3226d8d907fc9ba78f34b085dc90886c58f98393e06c7fe3fd6116356918e6e88599908daa2b2d2e482f374c54c9fa051f0d78d33d908821c10e3da7fa63fd4d2cf5be7addff9729884dd2a17125e08283362200b20ab807572515811bf9baf09b826a2809ea87b6b86284bb8b49a56d2bc2c108162fa63944cf5e67c7b6c5ee9a59b9f595f570d2aa8bebc3163b7eb9befd60b5a4f4d726754ad3c6cbc912a99ae8c6d95b4745089fad5ca1f44c4dca75050edeca7f20adef47897ce455d339d2e9399a2217f45109ae9aeae188f1261544c784ef115e850c36b7cbeedb754fd1a2ec9ff7a0a54b988be27e1a4f6feb65425a574652ceb893bcc346ee4ff8d05b600f53b04f76983fe034089dd44fc956815bb7fef35ac2c52eddc394803fd534074191fa6c93a670099d80e19023ffeeb8979de66e0f14ccf7053653ea24c50fd17d431af8c3f1dfd9a4d1878c3d9f38ac98782c3835d45207647d72bbf319ddf24392ef89014c8f3ab5743f8100845bc322cc201a1ee59fe42ab2fa023dbe0c46b3a9aed1d8632d0cd58a0185537e18e066ddbae2373c3e138fc92a1956f507b57ffc381239437a05f6a4ec24450ff2c2b9c746aae9576d050ac4861117934e41f7834e80c04240362b8e60dbc27121e6f04325a0d06aac6ba140d679910fd2d16c77455b48d025aad51a029a5c521e20c59a13202311aa9259cfd5c9efe2caefd1f64001a9e9ac92dc319140b76505fafd25a230d9a0e150ab27f3949b5ad14bd778b08ba927ae623fa2ba139e58328f71a7a55c41f8dd9b0cdc4e76ab404e61b4019c489b30ab975329db80651c0ea2372999e2cbfe69cc701e2f969a7fb611759074c7efc519028d411dfb284c03518633ff6f7d0ebd95701e8d1669961dec50cce74d385704c7207f8cf4074cbb7519f14a370399715ab4cffe0ec28073d76b34cd420bbf8a10b4f0ff0a9bda2b39b62206e02990689f87029e9714d0925cfab3a1de6191b9e1b0961e2b8e33661f5a0da2f4ca1511ae9de34526ba5ccadbdc863896738af6882ec77c5533400d4387116c2c81be2af97190d32543e73ae50a10d16828c2c", 0x1000}, {&(0x7f0000003400)="e1c138454f7d9be9c6570bf9931361fb087748a8a4e4f852f4da691d35f7c4b8cb0f4bab0afa8c0b69a1f00a55e663e1e7e2f53700b1f1f69585a97ea74b009172bbe521", 0x44}, {&(0x7f0000003480)="90a545b012efd7eee3179d460c08c1e05e769008b0bde9dd50df27f5ff0f101c97483cf9aa4f3f6c621d14dd06c5344f370909315fd084b2651bea116585ea58a473af662055e09fd8469a1b9071e28b834ea0e38d373163c3c4c3bae0b2d0c5f976d03266db6fe8e37bce21469cc7ecca5bb083cb8db2e03b693af2b70a3c5291addc0b3f185d4a4aa984be98a8a19b8f8ef8b67e711e738af7898a4d34d31a987787fd2b19e765158d0c35a4b057d3fbf33a56ffbc4119e5dc5182546b9e4bc50cbcacf95c7d9cbca5f33a6c2a6ef5d0d1d35b2ffcf5cb4809c49e5ac713e92aac02e013e4ef0cd120651f49d821c100611d2ff2650337040b8ca717190c0b932898b7c427cce8fded11b912cf0549f190c0e34bdc1c9a08f82213976bb8943b82832b16c6c89c5c8929dca313cd88a8a003cf74917239970356fa602ebd4fca8cce28be388051774c859e154ca02a7269fc6371daca133d914cecdb7b955424227b60c92afe3346bd2e6ab541ca2acc45f4cc84f60444fac22d53e17cb52a2cc152816ea5e4b4c80dd0d35396d048210580617b8b8b654319e9f8894ffa40af33fe324d941c6a726be3b2918baa86676c9742f5cc2536740536cad7574bbaf36af6d45c89b4b243cf4d9f8c793a4a86e2f2d5c7987786fd9f5dd670ce1f0ecfff1d74ac4ae6d4ff9a74696b0784dc4db3143a8f7c67dc84977399201d1193d8c22a3911a1fc1d443660cb1ac66aad19aaab25e3b39a11cabe2657088f6aee10489cb9775e3e18a27f41cbb87c630a8815eba1600e8a81686bd5d5e79ea0d9f14a900e255af4532179f98dfa121fb4e550364cd00a3812c43232a6a640d831d0b7190f5248c8ae2adad758f49d5bf4cbb52a2c1eb2074f126db22e7ac65e6273093e9f1bc36366610fa0341916ccec3436f2b151f2574acfb21e35583eef74bcd06ba1b75c2b63abdd357a6a6cdb7d8f50168c697ff7e02f3525c3646250a4fd379ffd1e64b508b243277d679f441615a4edc1610f69f01c7dc105aac8ccccf762ed31ee25e3c22f815df57f126aaf9cc8b69ce5cabd71cb8925c29d3e221c53d4b03f5476e17c6990cc174977f10c86121929f653bbe2f996b9b437b8112fbad556a481605057b6bdf9928176853ac0e3a3cbb66c2a8621661e9b92cb658eb68bf5fe85501a5ba1cd8b1fae095fa03a0f7983732891f95fae2111a7071db3dd7e07fa739df336ffd6ec4476ac89cac2909b64eb6ee2d1a64dee6ec3ac3197764fe140f2e64f64d93c4a5641474f060276d8b3308dfdfa69b80a0062d7669679025eb0e9d244cf9655503481f96961ddb933dc7b816f69f4f8ca10f500c1009f0ab0dfb768b33e9d5e9fdb052f8d1de9f20323f423a8afcbe09609f91cca70d767255c6646c331fdec51159b7a11edd178c368bd798b26dc197ae745795219e29bfa043102352fddaf3bb64e48c48989eb17a60b7aad56829f5ba46f9532e5767de775b3aa06feb14736452e71039711acab1e3df9e74df613946a61a185543ade584f169d412a2196d089a316a8b341cac6857bec2c9ff31466ab3dbf9c386364498b257045d62fbf793bd711b5375d89a3eaae0eaebdf5252df5d4e19b9a125e4a12c229878450e718a7ea27b6baac61f0414e7c68b7cfae1132f4e12041519284fd0d25e11e7462460413828b4ce6af4f2fba22e40902c9fbf7745a2b3919f6c88624442f9a44cb50beae51f48cb5c936cf92bfb7c5ab508c19a26491002628911ee7578909bf6b14b020c508fea6c919774fbe9193e41c21aa538aefcdbd3d197ac5e9dc36041193e55320f7e16af828638328e264650273a0199eecc2c028e9360ac1826099a91cc243b3ab4789643f49851c9b227c734ce10dfa3dfaf9fede9a4320bc16cf5409137ccb51221cf07a3e1ad30e0db346586dc7c533fb87f2091afc5c69952fbf2cb0232ecc234b7c8e3fdfe0523f6ce873b0f4e8416f0db32e244bfe66657aeeb8af6c591b5b7eb62a007fecb0a8eb6e83ab7350b56d45a260769269d809845bcebf3b54825a51ca15f7259e1e40cc9673ed0c313569caeeaac3ccae59e4dfee0d72a4fe24edc9b55bb4f40715d6e7323a146bdfe86bdb960a29b6e1239002f5a6ce61f16a9376fedbc2ace452f3c2c96c7787875c8201057a0349948d27e83550a3a4bcebcd87f11a94fc3452d12ece2289e5142097ec1bd8dfcc36677fc305de79d0a6bbd6b5c34cdd59b3704e4030b316de0e2483afafc1a58f494e990fcd88e85c1ee70f73bd679b13ce469830493b4e8dcfe533be17f4ab404f03ebbdeffca7b58c4ebce69f981594a9b3eb358a639b45551342a132f241eae6df0486a32da122f780521bb8d16b1e334c5d790f591a51f1ae5639d606409dc117d13ebbf726c11b03f7370b27dc63be08cb1dea070b7ed78ce6bf9a563a686b744dff46d14bffc16893f26dd959d17b6b4e0a0de4e1533f98364c14c81dde7b8b65be00d9253b229994cdccd859a51e98f33f13279aaf8132bb26fbfbe9e78d03c75608ce7ff913d34332503435252e9d8ff0e42cc4fdcff504c450773fd82ccf969d7f8dd0dadaa0d66123ea6a9f83fc12cf4f1e80dd61a6e1ff46e23335e3bd30ba54e1d3dce70dca4cc826106420dc32b53178db724beca54d8c382c85f0054ba9c32f8e6f84b14f2ce67a78d423b468ee30901645a24e0ee5d0d6a3d617094d94514ca74ef19d9da7e07b775e85095b77d2843b9c60f27950fa53158bda491bb18222d6e5a214cb47b4142013accfcb179729bfebf5168bea40262338cc54388b4a9a2a3a4afbd7624eda361bbaaae13673f7c533403175462092255837e0e63dc3bc503a6f9a16ec7614470a8f2477d441ab4d75f18fde372d426ac27d271d4e0ed363ee32dc83863cf3218805ced05dcaf6dab02f7c9d31dc1301e37e127d7efe56d0c221bdbfba1f046c47ed02b7c9d793f8afc2457c6d6b4ca04cc214ff08761d807248f51de0db088f1888d67ff926cc685b38690ed0c6cc70a4f148c46705d41c975e19739668203461d7c13419cc4d15a38f1c9e6a04c647af3ba3776dcdaa60bc017803d232ae305747664a7d48672bb672744c4c68860bc1ac1f1829ffbae15983706536b89666e4d2727a65ecd1b0a6cad5d2e23c60547016988cbb143edf4b3ab1fd2b8355bbcdecb889c3efe5e0804d8d9c0a5590c51cfabe3085641519bdb4934d2e188bb058a49559f596dd0d41e8c5d4ecf3b6b20921606232cc686d6879e9c6ea5f09a629cb4945085519513b7b189c1edd775b8cfd53ce172f15f59c19c26f0b0b36e57d4ecd1c8a62dfa9105b5809d3823ce3e8d253c0b7a3fccf1b4957124845c5f1bee5b554123010cd01a4041c18dbc4f01d1c502c87c7a5c67b2e7c9b5f05fbf19e428139a994cad2ec8d63ddc30ea8f960316e5eb3a69c27ef43657dc1556f65caac4e72c68c4d87c263e3774ab466e3131477209f5e1c4312c33dc1157379ea2c9ddfd449a1da2d2d4628bef26ee51b56e91d1c3e947b4034571a01b85cb53578247f87bd4b2a84819a4897cd01272cd2d7394ddb109733d21eea357c85cdee4b255635030a9fb17ab02de447812d178ba2bf46dc186dc2b055b30c890113ca5fd8ed78fa5b8b54627f05d307108342737e04d48fa280539c36527beaf913921536de89420ef54936cd8e098f9d308e1d5e350adf47b21a18aa714580e8c947720c4a280262f072c0ab278f408bb0886fe6020362e593ff00de288b580bfd4cf5ebedadf67da256f5f8ccae787206df70b0bf0037ff658df4a94c1dfcc87aefe762e83836479f4cea35c0d3e90df5bc27dfab57e9e40cbfcfb780c45253447e6ce99471466bc7777c7bd10d88e1e31ddc746247da1ebe728c46b9bf2613a347f3181d37e88baa1ed051f685067fa464711b9cec5467ace786fa9e1bc97b317e733921b795433b4da5466a379d02908e9b6a3e7f35beebf934ea16649cdde811eafe1c2533adc8138d85289649e95583e1f765a88c393ad78e147c30575dcb502f2b07c7c3748301a12ad25cd17abe308dd46764ed5728b2fdc91abd211b4e8833b2b82ba1bc55f403de7f5e38379bd369d36f7cf56acf59ecaa7111ee4a31da53b82c0a597b37ce7a197a4768d066894948ba05dd0c733e6fb867d7b9ae58842e3dcd4a6217a876e84cb3209f39ac63c23b239fdb3316bb9541470818aeffb4a1dd33be87f9b20d64d3d3989f1743619bbc161c38ec3143a5f916301283d382605dae0bdca4f0aaffef247dcdcf0c03786176d1de85b2996236ceb7da75e56f11bf8e6278a1ee6e4a6a71e8d02f5f4592eb59c016b1874d10e30880ab16ecbd2b7a3e88210fec183f5da808d382f83efb081a5abaaeeefd6ba887e0f3fe1ecda8607e22f3b7bdbc71d7ab4a327e899e286878e3d75aa8551f8c5f4fe29e690a56e40a6a6e8c08a3632bad623f4e33d92af3f894440e7f054ab7ea085f4c01faac564b69d881be32d8f17d7921718da558588dbbaddacfd305cc9464407c637cc028d1fc14d255a90cfec1f775e93d75e26348ac921532c1215992330969811f6f0c642601cca6d766bdd08793c9376e9ba4e195a41bf8b2ec8d5e2ca39d63384d896a0d99db43263b71a78ea319634432205ee50a4131e4544c0d204648bff4666099ad187ed29ccf2209b8c4d0310d4ebd1677478461be26cb2ec43b14aecb2e275ca02cdd2803c9e2f190628e34c8318dc119a6df82785ff91ccf1b76c9eb56193f03e564ccbe4e612e8d6097638fe9444cd289a09e44b60ff275c664fcfdd74a200d2613a84e9c8ac3fc52732c047494633f295a853d8d798f4a3f27601ed846182b420d220d8b724d33f1b5d0e3f211fed7a213584d465521d4d6edc2bcd0601f503031dc70fed82965dfffacf050278cbf533c85537327bce213c37170289a8fea69e80a359fa0b8cb6ba43deaf2de895991d1259281d8cfbfd2c5a38a27c7af3b11dbe972cb66d4c7fc0a26ad81078e47190d1ab3dea349603dfc9e9182006bc0fdb3b81931086dad5af2ab812f8aff75e02e3967de162a2267c385d3ea827b9e70f992c692f963cedb7f754a884543958d5d21cefce013031436119426df999dbe978a82688377ea9e287a414dcdea1e8998fe2ef51d1808024506744f77a7c574780ee9ae28f8c2a526a98fe8310a36bfad61b831a40a6e133f30587eb31d854f1572c59a368263cc38665b26f3a0d2da310a8db512534032108378437033f28f60cfca24f1a05353608bea254943f7abd887fbfc5d270c04490903d45a92fb4b5cdfe87e2189fb31479d9c63212fd3f9e8ed55ed8aa917186c9fe9413d4861da5003fd83ff51f1305f6afbe3b4825d15e0afee3f60cb14435f86e7d55303a21a6fb726b1a09bbe4a7a7180f5caacfdd62f75d973c43c7716db111b8e93b001428605a025af9ebeeb9e322077cb171fb969ac7e53f126076b3fb8c2d3f252dd70d0845d40a46fb71d777a43b584c5a3dfef7bdecf75a4bd325e1ce0a1e77b039c5b35cd51fed6ddd93452543166eeef4ad63634b4ddec34d654f2d8b644d13714e8b12a9b22346dfd480762e640ccc364d18dd2bfb7059dd99624bd7466f78637a6f059efe4a0b4d795a06cbd9043bd922ea486da176766b74f5d62344557690a53262385c05a797f314d7d97389abb55d1996d620d6cc923926575e27c4d72e01bf679c83abe9acce85091fa4015eb75339d4952a4d427def31b3b5b0205b38ff151adc769c6aa", 0x1000}, {&(0x7f0000004480)="d70e621257df9fc22b4e6e83af82bceb9a1b84b516e5e16514792c8fa71b54c604fa17ec41d367188778f46e70838e295ae7270629fba2db77e527b8231292ec39c9613a6cea8a8363018905acb3bb08", 0x50}, {&(0x7f0000004500)="45c2579709ade63c23feb0dedf4b6aa15d5b32d6c313e47b203da15f073cd6ad", 0x20}, {&(0x7f0000004540)="dc9cc92e18abc8b34c9e7b1df9d95f9436b6ae0aae2648a30edc2e427fdd8a5aef54f55536a2a4c653d7728a5f84801e35d5e57707eda3f9dd09f0e36455b835a11853aed2da0c8d61de46933f7c16176b0c58c8902e6db8b08acc31b5bee9d0c89734a3e470390289ed8b69834f27a3eba9ac8c1c7bf0ab506193022e32a54b45fdc23bf2f8278769dd40849ff9b08b16ed4e0aee75caf41df32531a15e5eb3bf87404cd86ab4f71f584fbb4b58aefed337529f2243a713ba3d07414279c792d7659c86c9e59fcfee890c0e6a851c0aaa911ad7b03cfc839879c28dcd0454dba2b135039d475027", 0xe8}], 0xa, &(0x7f0000004700)=[@rights={{0x14, 0x1, 0x1, [r15]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xee00}}}, @cred={{0x1c}}, @rights={{0x2c, 0x1, 0x1, [r0, r2, r0, r2, r1, r3, r2]}}, @rights={{0x20, 0x1, 0x1, [r1, r1, r2, r1]}}, @rights={{0x1c, 0x1, 0x1, [r1, r0, r1]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r11, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r2, r2]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r17, 0xee00}}}], 0x120, 0x4c050}}], 0x5, 0x1) (async) getpid() getpid() (async) getuid() (async) getuid() getuid() 00:28:48 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x1200000000000000) 00:28:48 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) (async) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)}, 0x4008804) (async) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) 00:28:48 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x17}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:48 executing program 1: fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) (async, rerun: 32) r0 = fsopen(&(0x7f0000000080)='fusectl\x00', 0x0) (rerun: 32) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(r0, 0x6, 0x0, 0x0, 0x0) (async) r1 = fsopen(&(0x7f00000000c0)='pipefs\x00', 0x1) fsmount(r1, 0x0, 0x72) r2 = fsopen(&(0x7f0000000000)='virtiofs\x00', 0x1) fsmount(r2, 0x1, 0xfc) (async) fsconfig$FSCONFIG_SET_FD(r2, 0x5, &(0x7f0000000040)=':\x00', 0x0, r0) [ 1728.763926][ T6027] workqueue: Failed to create a rescuer kthread for wq "nfc4_nci_cmd_wq": -EINTR 00:28:48 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x1080000) getpid() getpid() getuid() getuid() 00:28:48 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x18}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:48 executing program 2: r0 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000000)='/proc/sys/net/ipv4/vs/conntrack\x00', 0x2, 0x0) ioctl$IOCTL_GET_NCIDEV_IDX(r0, 0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:48 executing program 3: r0 = bpf$MAP_CREATE(0x0, &(0x7f0000000f00)=@base={0x18, 0x0, 0x0, 0x0, 0x24, 0x1, 0x0, '\x00', 0x0, 0xffffffffffffffff, 0x4}, 0x48) ioctl$ifreq_SIOCGIFINDEX_vcan(0xffffffffffffffff, 0x8933, &(0x7f0000000000)={'vxcan0\x00', 0x0}) socketpair$nbd(0x1, 0x1, 0x0, &(0x7f0000000000)={0xffffffffffffffff}) ioctl$ifreq_SIOCGIFINDEX_vcan(r2, 0x8933, &(0x7f00000001c0)={'vcan0\x00'}) ioctl$ifreq_SIOCGIFINDEX_vcan(r2, 0x8933, &(0x7f0000000280)={'vxcan1\x00'}) r3 = bpf$MAP_CREATE_RINGBUF(0x0, &(0x7f00000001c0)={0x1b, 0x0, 0x0, 0x5, 0x0, r0, 0xfcdd, '\x00', r1, 0xffffffffffffffff, 0x3, 0x2, 0x1}, 0x48) r4 = openat$dlm_plock(0xffffffffffffff9c, &(0x7f0000000240), 0x0, 0x0) bpf$MAP_CREATE_TAIL_CALL(0x0, &(0x7f0000000300)={0x3, 0x4, 0x4, 0xa, 0x0, r3, 0x80000001, '\x00', r1, r4, 0x1, 0x5, 0x4}, 0x48) r5 = openat$ipvs(0xffffffffffffff9c, &(0x7f00000002c0)='/proc/sys/net/ipv4/vs/drop_entry\x00', 0x2, 0x0) fsmount(r5, 0x0, 0x8e) ioctl$SIOCX25SFACILITIES(r5, 0x89e3, 0x0) bpf$MAP_CREATE_CONST_STR(0x0, &(0x7f0000000040)={0x2, 0x4, 0x8, 0x1, 0x80, r0, 0x6, '\x00', r1, r5}, 0x48) ioctl$SIOCX25SSUBSCRIP(r5, 0x89e1, &(0x7f00000000c0)={'macvlan1\x00', 0x3, 0x200}) 00:28:48 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x1800000000000000) 00:28:48 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x19}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:48 executing program 0: r0 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000100)='/proc/sys/net/ipv4/vs/am_droprate\x00', 0x2, 0x0) r1 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000080)='/proc/sys/net/ipv4/vs/sloppy_tcp\x00', 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(r1, 0x5, &(0x7f0000000040)='/pro\xc6\x04\x00\x00s/&\x18\x89\v\xd2\x86~\x98\x02K\xa0\xc1\x01net/ipv4/vs/am_droprate\x00', 0x0, r0) r2 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000100)='/proc/sys/net/ipv4/vs/am_droprate\x00', 0x2, 0x0) ioctl$IOCTL_GET_NCIDEV_IDX(r2, 0x4030582b, 0x0) sendmsg$unix(r2, 0x0, 0x48000) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) r3 = ioctl$TIOCGPTPEER(0xffffffffffffffff, 0x5441, 0x5) ioctl$TIOCGSID(r3, 0x5429, &(0x7f0000000000)) r4 = socket$unix(0x1, 0x5, 0x0) r5 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r6 = getpid() r7 = getpid() socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) r12 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) r13 = getuid() r14 = socket$inet_tcp(0x2, 0x1, 0x0) r15 = ioctl$TIOCGPTPEER(r1, 0x5441, 0x7ff) r16 = openat$smackfs_ipv6host(0xffffffffffffff9c, &(0x7f0000000b00), 0x2, 0x0) r17 = socket$inet6_mptcp(0xa, 0x1, 0x106) getsockopt$inet6_mptcp_buf(r17, 0x11c, 0x2, &(0x7f0000000000)=""/48, &(0x7f0000000080)=0x30) sendmmsg$unix(r4, &(0x7f0000000ec0)=[{{&(0x7f00000003c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000004c0)=[{&(0x7f0000000440)="da5c9e13a34dfe063a2c25ce8170b209081a9163979c20a0c82636f33479123ade5c213c72fbafc6636fe7e1c97df2ed7a72e8d0bf636da9ad81abee515072fd3a2dc855283e9c23c6c90e3603477cf1c7e70ce2d7081697a2b0c0c06ee75f4b084086fe6828e249faaf65463329e57b05", 0x71}, {&(0x7f0000001880)="e801e3dabb66d0e9bb11720fd55cc42d4b935e0ad4b80b33437dc4091bb237268373235f955fb253f53db94adef4a0be5d8b56a72f1390c08f60f81ae4822c77c5fa38c4606cf7423fdaf80b559db60e117d4419f8ddc6bdb75f9d71f10753b18cdc21064940d18cda52600e14861632d6d08d69cc2d0fd94c2fde1e9551c9dd406e36d78e64343feaa3276ef0a5b3ebb95159e84de393c5fb697ad5a44463027cce27c9324ab724712557ed6a196585e9a92f613d798e3b9c0984cede6397b695113fc741b83650d1fbab84d6f8be2aa970e80146ecacff903039737796cdc6b2631f2c634cd54ba75750a57d93c197fc74364bd32164e526ff6056e968495519f35feea23b164ea8b6615797018aab756dfcc8e50b7c5b015f04ec30830b18dec02e9178ab603097764c7e387aba3a2a7b664b039738b81d1637cdd331ef6f88932a8ea81b1353370c60a08f2fa240df1b727c626f90daf51df59327d4f901b5926aac1823ad91eeefa060017c54aad627fb8597cef64d28b270d87294bbbc84b02fa7a239d32aaaf2b6c711b0ab127960edc9591fc6b0138cf90daea89b5612ffce3b59ff81c4666226339966debfeee0e634aca254d38fdb321c69a88b7ef5663fe774f8ad2bfc01768321801e395f723e830be710ac6c801e7cb91f5a1d05fa4658f265e60ef083a213d016aa7013b06a5a58881a57bc596159bf0ef759331c14fe720a14995f383a3eb30aa4bb51c9bac35e891ddfcb9ba3dcca80d24a0ee216aa4990f9ceac735c3f1af8d934567be4f55a87e2637814681366455c9242411309fcc357a6a07c623f0e593058976adc243d0d61477d1d8865f01e248b42c2f52de232594ceb682dcbefeb024595139c3f652356f10eb3f9bc1238a3d81efa8790e54921ec8be2fd4ecc12521948da585fbb33f9628ba20d9a493d28710db1254e355162058b0b984eab64b6add35618c71920bd392403d3173bed176b9b10e62f747d2f2e25efea46a27fdb614e83cf0c1584d1fa5bdb1f0bd5ea5d5ea8964acce83f27fa5094ef0e226d0a136f5fe3933c862be2786a5846583318263f3b7a9ca50fad8c7c1347cb732594e102637e52780046b0a2e30618a14bc139cd9190fb7b95daeeda76207ad489ab7bbaf7e35d96f69089fee5e6d4d5e7be8b9876a7474130d28abbc9bae5a728dc5906798de62a61a97da600d24ba757af4a4bf7597f60535b0afede7420677bc999f83e8bdb395e8e5aabf0545d4ffb03e006c3a515ebc554ee675b323236cd73ce5485468b271d023f02c2b0d080bc82dbac7369007ace87316feb439ccaec7e078d7f2e87bec62b5722809fbf23a4666c403e0330eac6b1ac29c463d4939098ee9296957efe6bcb57e095c75cdb5e3c60117d2b7245e8ae7bec11a330bc81fdd6bf965fa3ebaef03b8749a29db6bdc31c123d9b3c93bd4a91449e9df1ad36f20707256f263ff9e33c67af2fb074fea38948424bf34a252c42b218672898bf8421ac0b159be38b42476ff8f2d9992ea6b176004fd6c37d05800da1f9b6ea8058d23b02d1ffd05b888fe41c227af39e8110761b7800f13efa08b5e03f86564c12001306bab7097d3327a5a1ba3fa039c5d87c0fe9d46c332c3acd082f6e55282f5a5beab8009ccd29cc02cf6f9877935fb50f678e424f7fc4a2bb6501be29b1141a458997a60edc16044ef85134bbef77c0c983ad10ba171217c04f4c8fab25eafb7fdc7e685441b7c05f2280760c3768b6e81a277bc71f57563adde4946f2e77843a6adde6a63f4b391ef9aced6ee4f20a88e4e495f44db37b4620984d026d6b97732b05b45c51f66f9418b5d9dec5a18e17ee55e4a46e28082b51f5d22f4f5ff20d184e60771b1f881ed2243c12f61e2db7f6d7fddb0b8692dc5e071446a342161059948b5dee5f7dea65e751ea03a7970db69e760f7de704a22be8959f87808474f46df00137ef9cecbcf8e0d11786df92a402354de36b0d9ec51c8fc84cb580cf606c7fcbfcbd2b20b4dd055614b4c8d10232f9316f58e495bdb517c054ec662f9d4db9a3f8fc6ba38eb1a916d8fb5f119f08f7fa6ae615e3c318b8af848970ff545ee76f4871be0714ad9b8aaf06f0bf25f2692ea82f1f1c9154b7447f05bd3c2ea64f99c95e5a19eaae6450f10923f8a46a57b52e70bb25619677d9edec808ab66ba25b20a5d6fce6e0bbf269d9abd6775cf07b0efd1f33f8ec496976b33425ad6c534f6fc5099ef3aed1306baf7070388a934926970c9ac102e1bd0fcdddc3179ac3abcb460edc7efcbdd20ac05554ac952749a880a41c87569587dd0aaebd7ddcdd70f2b94eb4d2dc5c934311c8c6bfe7b7152e7aa2b553e491a0563832eba740e6c18f2adb06909fb5be6ffd2b801f3730fcd07ccacd196ac885aee7f0caa2020255bafe3c01b737a8b641646281a79e3ef89e01b24be21407f629d90f66cb5711752d918cb79e02b9f2755c057fe829ca1a9c833fc7368302c3ba0a62db78268e37577431ef8c29cc4aa4351524cb71f5ce5d6c1abf1ba9d815640c4fac8f4b7933cc33bfb37695516651adb50337c725ba6e37652648758be0dc64598854e0b03133cf5c42b46ef980d22d9c4f4d283373a6c600e782881dffa8f0eada4d1e31d29560b60547ae71c5b01a55c2a8ac8f105f468d2e74cbfe2b3ea234805960f2793f1715bcd261355a296b229c1261814c98fbe5a0a9a50cacf2351ee014205fd335b69874da8ff787bf83f029c2a41bf1482206390da5e55900850d8415ab941e5b4e662fb46b54cb2c46f4a66919117f3cbfa168e94dd73036c6d957ef979f75e83bed6f44c3e61f7c096bb36801a4f5e13b2f51d32dcb05a7f52c66ec67f5c7d7c8a3f3a53cd9177882db640d7158fe903c0aeea4a3ea20370a65b2718415edcbc80fa971c5c0b011c41deeb58745e319e11682cc2f45d6efc1ad0a621f376e2c2d3f7ad3e3003be6cc247e7abe0e596152ebee966bb379858f6acaf59cc1fcd200dfeb95e7d80f08cbe88f45a4d460435185179cbe885eb130a40c2e4a07afb63642740e2fbd6804e978abdd4adc00765f4a0da81ccbac0d5d91cf0866b1f111a199d94e7ad6785c63a6576ba86a856b41515a77bb4d07559d4607c38550e63857fe8436cb78f01a56d9b91a3df43c9f006bee864af231d5da037283478904398b73744445e01a41fe8c983fa6adcdef8af14053d6643c58aff413f34ec3331fbc41c0f189eb3a02d3764b9d1d64c543a195bd5e6dc188fe694d7e670fc712fccc3dc2264d5a2c85998c0c51b07cb3ef7d14f1bd78bbe98a202be945c7fd7375a698b4858521ba33f54195332d39d3f16bf576fdd6a7ba4548a79ca509f666ce9242771863b557ae9bb4353cea6b1f69ecc0fb946dfd3d1b577878dabfd8a0930205fbaded24da0c6734dcd4540576e71c5c8ae99a1099cc5d72cd8b8e7c0fcfd3a1fe7ca29d9edaaaacdef83bd957b712891b616b7a7e68473ef3759c7bffe521479809e881fd21ab5feeb35296b02dc686735035f6b6924fd6331a0cb63453b4bc3d4fbba41c9df87716d3049322ff1e2c1c1951f2d3505fdfa0b89e7ae28488fecbd04b5e23bf48d9b7b80f400c5626f64e66d7e3ecc010588bf6313bc014a9d14a28b238edcef30fb698798f04d19e35260ffd7d90f25ef17ba4f02664efdced59b5a49838f2f6d393ebeabcdeaea5f76be2508bc8af5f5b2efd28f4119c955f9ab2d7ed1f3c21adb50efd0786a2d881e71b0283ec61144d26ba51d836e8789631c0541e0e0239462d001fa2f281a985d12af2c3149bb97196fb0bc231a32697306ee9d5f766f1dcec8c9ae7dd61bbc1386a2f1b232e4475a27aba38ab5897fd1e033a7b834a51d684984afdbe516b39e3569247332ec6b88e39bb024de66f3a95a20d9dfaa00bfbb47ef60d2c939ec1655a1a68b2792ec77f7c85f5a5c498f77a97efc94e37f6b74a492224e524d5e275871d52b3c106509c3c08f5786354ab2c19dfc7fd429aca72c8eb6edbc07f9757941ae40ec8d16593fe227dc88f44127a6cef17be8f7ef7302251be16efe01d412f19e0f0e8e709c7e2ae1edc9aef66e9f5cd8b8e9c26abf94b61ee37b23152bd07735feb888aa015c4d50f0453df9aa834752f0886ee209ed63705c7b4fd6a53aa7b3ea82f2d228bf4fd9218981a5d0a574cadf609dd4ea5c335dfb49d999d940715a41b400e8d81f58300b570a433da9b2086cabee279b18ce5daa1981892928c007796dac2cab04370ea90c15667e6d17b6c539b63a38ce23e32e96943fa232fb3a76a5812d98742fa06dc00359d3d08f7b8425c260cc0bb55dcadb74db458e6126d329675e92de7c4220a066c6dec5e091a745e21f404efd3b8233eeac7fb610f96594c19cee57f7f1fc18f7dbde713383782063fe8399791447ce6bb612bcb5fc3681e6e1fafcf9c34429eedaced8c8d3d9f38c19ec1d622bb7403903dcdc0d8ba4c9ef4b1a22b872103e1837b7a7040612ddf9840388e54d9e1a28906aaa5b10621677769c58ffd9cf4a4a1561ff742b4319e03c10d76b0db41be8f863f5be3155d56650c4cceb8467b49fd334f6cf434e13ceec39b7ffc0aa28ef18b4edd1f0de8de0b083302651bc21a2e4a53e7aa0f7ad020f0cad0e6626b39472891d99490080c0df99f98a470d5100073680d05aa0e5b35ff65212ad76897c26d076d231ae687d1c8877af3013544b56cb7926195b8846cfa8177e48374753f4fc47ea53c493b16af2ebff1209e92787f6a09ff0a366547cd4cdc8cabb35f50e317017bebb0b0cebf64be8c370915de36f91b94aefa5ed5b868d2f38158188163340adf41fea5885ddbc9cc9b1ec9aebd57ac87d5fce669667a8d9962e18379db0f95554e2aff442b3f5a0b1c6925b4037399bd6f95a05ea88c164d138bd4c2e28639d8329c18a8caa3bc948d763024797b0de771b183015e58adbdf0f9ac9bbb3bc3cbe0bd6bdc0388708f015d822999bfc40c2894d6ffaadf6b072dd0bf8482f5ce25533ea8f2ace4c863fffa8697517fde868b99e55bb2bdd477580d6fad957a7eb121ea1ce8b6f4e821e0470cca9c0828f4edd8c80182556c5495994ed747f8b80af8707a4e1d3df2617445cadc262040eaf8d8acdea3916cf2da8ae18d9caf1db16963493ece1046e268ace8fd8eeb0ebabe47d7bd8e7aba9b3aef56c9371c0b89c7447048029f2df5fd88c4856ec13d25539f1f9a8c4cae82efda0890aa1f6742e12216b44ee5f9ac5af25f4060ee48ccd7556ae77dc2d09f88032586dfcfd35d258e03c00e77284b90b27881af6e1476776b5db669a1c5770890076ef23953c73310a535a389e0cfc0d27d28162b7b25f9bf56ac1591dc4a70a3e789d33f71ca2e8f0b40c43adc07fcd17116a47a378256f24169002295a4a4bd8a0cd7ad63f3e3e1fff2a781761af3a20aa2fc016fb3925dd8df8436f1f2c46e399819792735dccce4d13585ba8817afc8982a2ea3c885c8916a19ed2454cdad8cf26effb5d0f204e5a67d3b35dcc8ea8b55f6aab5c8bfc59848a3f03a1d4a006e5504fb0e28a046199f44096ebbd07ea4c381ca7ea0414bb16aca7a84b16a72c7144aec667a6f8fb6cec84c3ecf21825388fbb41e7179466ae3002b72a560ba424258626a337d8823294b3b4f08a89ee15c80fa6481578464bc29e16baaa4a834dbb5ba4e55741df3ba92336e6196d13ec4c69ec129c419790caf302343b1434b5837ab81ca4ff80481eb709df8b49557eafd6dfce15fceeeb3cd", 0x1000}], 0x2, &(0x7f0000000900)=[@rights={{0x18, 0x1, 0x1, [r4, r9]}}, @rights={{0x20, 0x1, 0x1, [r14, r11, r3, r10]}}, @rights={{0x18, 0x1, 0x1, [r10, 0xffffffffffffffff]}}, @rights={{0x1c, 0x1, 0x1, [0xffffffffffffffff, r9, r14]}}, @cred={{0x1c, 0x1, 0x2, {r7, r13, 0xffffffffffffffff}}}, @rights={{0x20, 0x1, 0x1, [r5, 0xffffffffffffffff, 0xffffffffffffffff, r3]}}, @cred={{0x1c, 0x1, 0x2, {r6, 0x0, 0xee01}}}], 0xd0, 0x80c0}}, {{0x0, 0x0, &(0x7f0000000a80)=[{&(0x7f0000000a00)="272e3c537de42b5794264df9612c3c75bc17db110ee2b0202e733ac53e576fd89cc5652de684f124a9bca651e5b4d3f54ea24f2abfc7715f2fc76d2266bfa9e5e0191fbe3488a87a84675dc511f8334f86dc91b3d2bb0446ce86f318c750688ba33a6679a2c60f6026d4648cf21e054f1d1a787e059d", 0x76}], 0x1, &(0x7f0000000b40)=[@rights={{0x20, 0x1, 0x1, [r1, r4, r5, 0xffffffffffffffff]}}, @rights={{0x28, 0x1, 0x1, [r12, r1, r15, r3, r16, r1]}}], 0x48, 0xc000}}, {{&(0x7f0000000bc0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000c40)=[{&(0x7f0000003900)="5f3a97f9e25015454c451722e6b070dd976bffff79ebd1cbf69f8f812cf767bb65dc6b6a6861ee8e2dbd8caecd769f9d0df9c12eea3e33e25ddaa33669c106d8bf3ecde77fba93ca1f1eddf04bacd23d2d836022ec0b8fc83592e9c28854418666fe430736c342ccbb92ac3fcae6e8711bf8757b04982c2f16ce84976369c63e9e9242ce38947bbaa8ef1732dadd6548bed2d4fd3e070559d10c487f3de6be13978021990dbeb5766ef652715ef2852168b5756fe6b7c546f452ce190dccc9b50871b66f6eef183ad50d8b7af6d44a803409566059ca79c4e82bc39bc1648005f1556e74a5c7ea4c6fdc082de041bda951c43b31b3aa4900e75ef6b139672e9d71bac445cc8f5cd6bfccf5c65f39fc04e94c887c2581102c052d6be23a6d744103b12114772dbfe2b4eb9191be3f424f611161fba1830c72b12e6961407f13365b95046be707cae44c42cdcb0ef924b04b16789b4ae8dbd6d92efd691fcccd15364fa5f112cc861e5886a7698f4dfe0c78d0756a367316c0b0e6e9406517df499d586e3293e0a93a11c6eeb20b9e2f04950e6adf939dc2613f9a14c1ee802f4afc91c8229f8e32c520426c7847aa0924d2406602b22d7ce9fe1770d6d9bbb7971a9b04fb4c786b9802f9999da001e4910571fd5aaad469ba01e46db24f5b603c9e3a59737bd1497e616b287814387a1759380073e874a35cfada797a70af5c45d4444a3c8f362bb617b3f965c4fb0157d13349ca45369837c6915c5eecbff9e0751279b04a772c1ffc267887e19c67f1eb9f0bee24b55563f6d83386c69ad76fae32fb06267755d41cc384dea50116c1f48bd42c44b51b974ae8aba85464b484ebe3105aeb602d281cf150e41990aa6480354e1edde4977dbde707d461f60864e8e43bdecb2401437d3d95b7b705b5ec435f865870c6c20715b5fd58731176ef4263f5e99a2667e17f1c8a5f6dc1bd5cce5cb3b42e2f5667580a05f4cc8831702315489dc5f5fa377cab563bbdc12cd3909017099478a781668691fc07663f8323484035ef258f12aa1816478fb4a41e0f7338977b802b1a8baaeb665fd24c11987ad10e319deb12534c4bd86df6d1c4fb49b331f4557d2009e6cd07318db00ac887a2b7b95d716d44510c5332373f2eeacc73d27ac33c144abfc239eb61a258475913b0b50e01c540fe0613c883432834ea64aeb070773d331374bd46d49e887bb8d1a1bba08bda6340ae0ae150699dd0736035dcfc8c335687d38a518ed2ad16a14ac64c9cc64c684f79c613a3466a7549e0dd61d896b372e79ed636cb3c118248ca6e5f3baa0e4ae3810c5298c1c205bbff40d275fd41107d86aa2c3bd6f4ac734ec680b54daeeeff749442cb94aa40987f7e10dbb5ab1e65ecbb539c0bca2bfec2aaea3db3b94855399c9b376bd5918d27d3d2b186ceacbeda05c8beffa0b0d796d725ad464b5df722321169f61e52fe7c833bae87ea29c4232493248eeb60ff45ffa95fee7f19a2836d08161fb67bc5e14c8c22187252b2035baf9374de69443fefa4c3429e807338d6726678fb9148ccfc25596afb9fba7876c224c8186c7f26f3e5dc278ef12d7939b8482b6fbbc58316d853d8516cc09e9ac9db31dfb04a24aaad081b895ace0eb427b97fbd4c21906e7ba200bc259bd63f60ea13f49d704f0754b7acb377f6c643d74d8c309d56740e237bd3f4f7f58bd570061caca51616eaa0ff9e19326161a03ad6a0562afb922baa18320b96f5d4d58a4b2f7c8e7c764ba3121f33e8b31da3c108894e29b2573370ac45c96f46adadad1126ad398677c1fa3a9306eb371e48c6861acdfd648cc4067b159e5663ecefbb42d919349b9b427835e603957fffe0e6c0f81e7c7709d97751e15406aa9f8ad4489a42febcfad0711c34c28b68f9921053b28351a1dbbb6fcfdecd486bd997a8be7810766256fe99116cbe48ea92e56c94e9c36120b01b4963fc3075be34abc76952f3f75d7c3c7349c11ca37914d1d209091cab231d9e6ee91c66481c08fa15ca12c71760d2457c7d6740bb2bae9ec2a41347de011f0077cbf5e0c7ebb30c1875d508e96ef72fafaf563734a5006c26daa3ed20064bf2e72d33f5a005ecb32e0c84b7f874c1dfc771659a5a10bce778d1b5a4c66f59d7233ac5911a8789546ffe1a0b96f320b76e1bbb90acf052ff456fd32c6938c74937f77ba38471e14a839d7ac6404b2a0c79532a039ecaf72eba4fec1ab2b31aab1d2de10469813b527df28a231c3e21f3172b0b2aaa3f473ac13f1e2bd0268fcd82f8eae180656ec2510ff3635be7790169b61b3cc46b9eea019fdccb0ff3e592ac20fb95826b0f260afa804ac3a71f5a7a37e316d2f82b3dc2e61fceb9d9aa095f8cbb13cea7d5c36546e77360eb12c162b0879dd980f27a608ec5a30fe35cf89fe77907fa34a171c38bbc7fc605f505dc9e5ae60d40256c5443e590cd6c84cf838701c3f7d340c94858b9533744b1f9fb03b68dc4aeba146daecad1d988e23d81c82e8a35304f1cd28bc653151576ad8fd0ee3c87e0cb61a1fb2fb7620b024332420fb28638c5c306611b79427f236849d610ad05e6121c2024f4a5c2d376d525ac5ec1f9729f827ae9cc7b9628ffbebe3c60c9b88625fd5b0641b4d5747292a05df47a7057ecf44be8ec57b247e88a0715ffd6189fad35486cdcf1c7e7ac7f7d7e54719fb15b592cdab81166aa355dc892249115814bf6e60219fcc2cd8197ac5c8ce9c1718300865c281e8ab4aa1c25944e467cc83ae51617ca39153ca78290417ed008b2189a38d99f3c0565e1f6090ba5c1f28d1a0e05ad4f1e9f77203e69d9cecfa442ebecd6a737dcee4aa9f81844a4a43c0ba412d35175657754e31fe31869ffa9cca8d0de7f75e131acce5fbae908e9b5e4cf12c0e5d3a4543516ee7884b0b5283bede3d677041825593bb2f28879445886a5ec9da94e3c7f8e720a957dccb1b3f1994029a4dbec1983b23e468a46b9d1b55a18532f8e92eaf00022f883ce7b5142da95fb7612d8a8f62be0ad5beda06ac35732631f15568f7991208282b98448f9e931d3c2321d3cf1f426d3a882f33b97e015a818f56bc48056538ba9978d1351d4c9e87026363be4c099c2b787362908ab453fa487f9451a812c151075cfacea8a8bf4097a25218e3e1039ab7d433ca69c1a3915a4ae30ebf88e1705a1282c01b3240985f713c5bfd29b764cefa1c8d080add2f675f0b598530884d457724b12ce0a59bac5d551891545d06bb543809b46095ea42384c58ea46e4733f0031d5756929d7d659aaa493af28cd26bbef64da84226bb07d285aa29be3af34b71189850ed4557fc868030886d0e4c0a905f6e0507eef375b475fcc0340774147bea21d22ca41983a3a8619d8ff3e7c0555b1ca37c912a2d74e3ab395d23bd0e73d930865d937f9958d5b3e1dbf79230be6a5a160a55c768f5f0b3c619d2316a7c65db1a5aff75948c93c33e1490ec724f14394f30cf2a25fcaaad7660f6b731f78e16fdf72e2cfc39402da4894114e68d6e5a20d3f7d0fb420dcd0dbb8e8d1d1df49700c9f4f2260a586649c0203b96b7a978f899272282bae047577ef80335b1134cd0060afb14ffa55fb5789dd31bd33b85f7e52124f4da12f80ca6874f3bbef599367fe2fde029eaa785ae7aedf5e29138655c3fed9b0512cb88a3cc5ba20f00e23a7ee95e97f61c2ea34103de98887dbdab233fbabd74da63506b85774cb45c083ee38764d5a761435b7b1d0f5ec960b933931277bed7d8be4f1614a49de99149007af3954353cd1e5d2ce69c0267f97e1d2307e2bb18e647bd7758c344e4af039b3fae31b48178a2f20da319eb3b32c0f36e8a8d087e7c93e637985e67c492eece4aa6e202cdf2d58e24bce4579cbae648601643828cb3d24d260bfa982f87588bf3952aee373301f4e064118159b4d4ada8e2e95deb8bafb9d469b7c9a9097e41e10a4d58bd9ea1e01a357f3a5eb9be910fab91f45919268bca8912a2c7bec511ecee9c491569500845520492776592b36f9586ab1014e2359cb31b1521e4445c6fe2370060c4e467feefbcfd6ae081900d745e4e4d8f703728165b51e49a4bc705de04900efdf0a921d63bbacc50a3129c87aae89a3605b1a3af0e137c20b64cd0107812692113b1dbe97f629337398b2635fb67c8fef6dd9720320e1a004a6b627b04683c496eb4cbc46353a352b40b893324e6297d39d6f982b894d55d04ed97880d31b0e01a23da1e59b3c3e0366b5f75d456e6782b32b1e1cb5830cc4144dbe0af9c573f94ad45b93cb802899d62e1bbfe1fdeebfa1ac712ce8de634cf2fac1e7676f1afcac449c1dc07dde258af110b3049026a1502d7d16777c874b6b6c03ca7cb3587beafcdde70a01121df9809bdeade57e3770c97d37022870f655ba862bc9104765da255d849d3bf9addde849264f1e0a81b635f01256d10b6b1879a08895ef3ac44e9d5fb312e8c6a9230e929d2dfa2060997307012f9b0658622f59a68d3d342bbb32b90a4f1a5b486128cbdab7d2434a9b67e378b8f637671b77485cdf284f7a238e075b8ba8bd7ac6daa1d387842a09ada3a431018b0ac909892ee08cc476dd9b4f409e082f48f9098ff0f46e4096e2f30377ebc42440ce0b05ff4b8282404d658194ee43351afe9654bbd9cc8839efd0c0608fd21680625dd11136f7f9881de07bee0fe0abfa6283db93a26f5f28496015ca728a099ccb1a356f30efd3bd8f839bfbb280332afbd17cc04be347f07d5e5bbc7779f2769add22e33ab221258d85770a854c9d75cbc0e3e92bfa71cb430320c0ca1f3dc3edcbee9d1ea219082893599df3c8427ce48c1322cfde3771a380e72e79ef8ed527b7e1f8573e848f30f4fe935038a5f4219f6b87b4679e8b577586747a6c6d6ea11f5f4c4be0c41099ca8a63a34f66fe4067e2cd3d31b81f42fca4e3bc69d55d69fb14bafad4776334abbb076fedc6853bf2d203cd6a3024d1f7c63031d5da3abdb61a87722b0061c200b1ab7e14f797e9707e65a5c81271be26d784b5cfe339c1a4ce05026d812f253c9d06a5ba1731d11ae1036f999fdd157d78a7774d467c493c0c5c66bd6d50ca87020b55bec81d03710182a49f476670aa91b8a1459d073f398f87ca0157414b51a9ff8807a29c5cc8e8d2c53890d0d739f6a3e363bfea68b692cb408b156e67879f3e8147ed635ededca191bed8ba5846006005604f9dd8508fddaf8326a36ca05ede287b01acceb843b0a64caccbfffa6225a588a99c49ec159740d4bf04658cd30b01b6560b4c8958758b59eb45d6cfbaacca844ac216ca187b631c7c7c5de04c67f6a5307e70788db1e8e504d80ca7b2cc26fae42ef76b4c88810aae6be014b0d08f5c21229bd681f54f446e5afd39e42a54645c6ea350c8cb44ce179ca53c2e33106e7d6df09c46919c86879accc8e0de0820c4e04b5a39d19024107d7864ce06af96963f75c93d00cab7545053de0eedcf89ff861819605d1536ef84cc023f7ffbe6508bad83f56944523b47120bc229c19e9814ea2dda2fdaceb0aeb15327f6948f6e126e344a352aafa36ece8d5211c89ad4b87d843e1d81757c584df431c2f0488b4b6f5520cbb0f773a3f1b86cd6482af1ac79f1ab8e62c7623a4176f4f76232fe755d2dc53191b2aff207964dad1a4f8a33c72b267727a46203d9206dae651cf13f4e29ecd8499a9e1ffd450ec657858a58266270d4e07d7435dcfb4b12f4ccff15be62e2f0e01f95f5d166893b3cecbac231fb12cdd603650", 0x1000}], 0x1, &(0x7f0000000d80)=[@rights={{0x14, 0x1, 0x1, [r9]}}, @cred={{0x1c, 0x1, 0x2, {r6, r13}}}, @rights={{0x28, 0x1, 0x1, [r14, r14, r17, r0, r14, 0xffffffffffffffff]}}], 0x60, 0x20000000}}], 0x3, 0x810) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) r19 = geteuid() r20 = getpid() sendmmsg$unix(r9, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r11, r8, 0xffffffffffffffff, r8, r12]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r13, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r9, r14]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r18, r19}}}, @cred={{0x1c, 0x1, 0x2, {r20, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) sendmmsg$unix(r4, &(0x7f0000000380)=[{{&(0x7f0000000040)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000300)=[{&(0x7f00000000c0)="dbe3c5e14986b76e473b23d2da33bfb31b5601b0f2ebf4f5356ad25a4e639b694fd5c3c9a54be9d87e1c75f3ffc4450b638aa05a2acf5f9d22c84f4ea7fbcf282efb1f411b1bdb3844a8201efa81a16e2dc17f5a23c21305add3428ce7270134e68c3860674333c7419634d6f7f6f646c8c81e8a760d1e0e9af06a6b998c4a6365ca0f73617fd565add2bf44092b563f4bdc0e3e920e241637cef8b80837524ca2d360e6e388fdf4e6282fd917d0430711bb70595e978ee0dcf1259dc168b2e6133cf7da929445c69802c73af83374b00830", 0xd2}, {&(0x7f00000001c0)="5ff568cfa2c303437a4027b1caf18f6ef9edfc6b6573111c6d5129d322dfd0312e431037dfd2af59a9de1ef4eb4e05157e45da0a3d9331553f8c8b24e57dbf687b0d6137f6213494e0702990af7aecab2646f20488bd2c168ce7dcbd2f2233af6785475b9068ebb563dfed5afde7fc38e52b72ffdc3135262c68a4fd0754b9948df6a67ac2792638f4d83c6b31759560f26828dcf2b0c4774271f25e904e253d32f3", 0xa2}, {&(0x7f0000000280)="349644b1b9b883e5cdd313480f39c45522346cfc6640ee9c3fc239f2fb80f43b0c2e14782ad3854925a59a8229c6d5ca800531943db7a6f30751ff921ed656a285d5901552d6491efb67921c5f9f2eb993c9dec164b5467b7c6ea8951a60e30c106161dd15e6", 0x66}], 0x3, &(0x7f0000000340)=[@cred={{0x1c, 0x1, 0x2, {r6, r13, 0xee00}}}], 0x20}}], 0x1, 0x44808) getuid() 00:28:48 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2000000) getpid() getpid() getuid() getuid() 00:28:48 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:49 executing program 3: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x1a}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:49 executing program 1: r0 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000000)='/proc/sys/net/ipv4/vs/conntrack\x00', 0x2, 0x0) ioctl$IOCTL_GET_NCIDEV_IDX(r0, 0x0, 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:49 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2000000000000000) 00:28:49 executing program 4: r0 = openat$ipvs(0xffffffffffffff9c, &(0x7f00000000c0)='/proc/sys/net/ipv4/vs/sync_refresh_period\x00', 0x2, 0x0) write$smackfs_ipv6host(r0, &(0x7f0000000140)=@l2={{0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x4}, 0x2f, 0x0, 0x20, '['}, 0xaf) ioctl$UDMABUF_CREATE_LIST(0xffffffffffffffff, 0x40087543, &(0x7f0000000040)={0x0, 0x1, [{r0, 0x0, 0xfffffffffffff000, 0x2000}]}) fsopen(&(0x7f0000000080)='vxfs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) bind(r0, &(0x7f0000000200)=@qipcrtr={0x2a, 0x3}, 0x80) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) 00:28:49 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x3000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x1b}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:49 executing program 2: r0 = openat$ipvs(0xffffffffffffff9c, &(0x7f00000000c0)='/proc/sys/net/ipv4/vs/sync_refresh_period\x00', 0x2, 0x0) write$smackfs_ipv6host(r0, &(0x7f0000000140)=@l2={{0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x4}, 0x2f, 0x0, 0x20, '['}, 0xaf) ioctl$UDMABUF_CREATE_LIST(0xffffffffffffffff, 0x40087543, &(0x7f0000000040)={0x0, 0x1, [{r0, 0x0, 0xfffffffffffff000, 0x2000}]}) fsopen(&(0x7f0000000080)='vxfs\x00', 0x0) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) bind(r0, &(0x7f0000000200)=@qipcrtr={0x2a, 0x3}, 0x80) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) openat$ipvs(0xffffffffffffff9c, &(0x7f00000000c0)='/proc/sys/net/ipv4/vs/sync_refresh_period\x00', 0x2, 0x0) (async) write$smackfs_ipv6host(r0, &(0x7f0000000140)=@l2={{0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x0, 0x3a, 0x4}, 0x2f, 0x0, 0x20, '['}, 0xaf) (async) ioctl$UDMABUF_CREATE_LIST(0xffffffffffffffff, 0x40087543, &(0x7f0000000040)={0x0, 0x1, [{r0, 0x0, 0xfffffffffffff000, 0x2000}]}) (async) fsopen(&(0x7f0000000080)='vxfs\x00', 0x0) (async) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) (async) bind(r0, &(0x7f0000000200)=@qipcrtr={0x2a, 0x3}, 0x80) (async) fsconfig$FSCONFIG_CMD_CREATE(0xffffffffffffffff, 0x6, 0x0, 0x0, 0x0) (async) 00:28:49 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) [ 1728.765361][ T6036] workqueue: Failed to create a rescuer kthread for wq "nfc4_nci_cmd_wq": -EINTR [ 1729.212295][ T6073] BUG: MAX_LOCKDEP_KEYS too low! [ 1729.216157][ T6073] turning off the locking correctness validator. [ 1729.217853][ T6073] CPU: 0 PID: 6073 Comm: syz-executor.0 Not tainted 6.8.0-rc7-syzkaller-g707081b61156 #0 [ 1729.220464][ T6073] Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 01/25/2024 [ 1729.223228][ T6073] Call trace: [ 1729.224113][ T6073] dump_backtrace+0x1b8/0x1e4 [ 1729.225388][ T6073] show_stack+0x2c/0x3c [ 1729.226486][ T6073] dump_stack_lvl+0xd0/0x124 [ 1729.227747][ T6073] dump_stack+0x1c/0x28 [ 1729.228908][ T6073] register_lock_class+0x568/0x6ac [ 1729.230257][ T6073] __lock_acquire+0x184/0x763c [ 1729.231533][ T6073] lock_acquire+0x23c/0x71c [ 1729.232696][ T6073] __flush_workqueue+0x14c/0x11c4 [ 1729.234069][ T6073] drain_workqueue+0xb8/0x32c [ 1729.235342][ T6073] destroy_workqueue+0xb8/0xdc4 [ 1729.236643][ T6073] nci_unregister_device+0xac/0x21c [ 1729.238054][ T6073] virtual_ncidev_close+0x5c/0xa0 [ 1729.239417][ T6073] __fput+0x30c/0x738 [ 1729.240505][ T6073] __fput_sync+0x60/0x9c [ 1729.241619][ T6073] __arm64_sys_close+0x150/0x1e0 [ 1729.242959][ T6073] invoke_syscall+0x98/0x2b8 [ 1729.244214][ T6073] el0_svc_common+0x130/0x23c [ 1729.245463][ T6073] do_el0_svc+0x48/0x58 [ 1729.246614][ T6073] el0_svc+0x54/0x168 [ 1729.247679][ T6073] el0t_64_sync_handler+0x84/0xfc [ 1729.249068][ T6073] el0t_64_sync+0x190/0x194 00:28:49 executing program 1: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x1c}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:49 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:49 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x1c}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:49 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x1d}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) [ 1729.313605][ T28] audit: type=1400 audit(1729.290:2): lsm=SMACK fn=smk_ipv6_check action=denied subject="_" object="@Mm+[%" requested=w pid=6114 comm="sshd" dest=22 00:28:49 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:49 executing program 1: r0 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000100)='/proc/sys/net/ipv4/vs/am_droprate\x00', 0x2, 0x0) r1 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000080)='/proc/sys/net/ipv4/vs/sloppy_tcp\x00', 0x2, 0x0) fsconfig$FSCONFIG_SET_FD(r1, 0x5, &(0x7f0000000040)='/pro\xc6\x04\x00\x00s/&\x18\x89\v\xd2\x86~\x98\x02K\xa0\xc1\x01net/ipv4/vs/am_droprate\x00', 0x0, r0) (async) r2 = openat$ipvs(0xffffffffffffff9c, &(0x7f0000000100)='/proc/sys/net/ipv4/vs/am_droprate\x00', 0x2, 0x0) ioctl$IOCTL_GET_NCIDEV_IDX(r2, 0x4030582b, 0x0) (async) sendmsg$unix(r2, 0x0, 0x48000) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) (async) r3 = ioctl$TIOCGPTPEER(0xffffffffffffffff, 0x5441, 0x5) ioctl$TIOCGSID(r3, 0x5429, &(0x7f0000000000)) (async, rerun: 32) r4 = socket$unix(0x1, 0x5, 0x0) (async, rerun: 32) r5 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) r6 = getpid() r7 = getpid() (async, rerun: 64) socketpair(0x21, 0x800, 0xff, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) (rerun: 64) socketpair(0x1f, 0x0, 0x6, &(0x7f0000000800)={0xffffffffffffffff, 0xffffffffffffffff}) (async) r12 = openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) (async) r13 = getuid() (async) r14 = socket$inet_tcp(0x2, 0x1, 0x0) (async) r15 = ioctl$TIOCGPTPEER(r1, 0x5441, 0x7ff) (async) r16 = openat$smackfs_ipv6host(0xffffffffffffff9c, &(0x7f0000000b00), 0x2, 0x0) (async) r17 = socket$inet6_mptcp(0xa, 0x1, 0x106) getsockopt$inet6_mptcp_buf(r17, 0x11c, 0x2, &(0x7f0000000000)=""/48, &(0x7f0000000080)=0x30) sendmmsg$unix(r4, &(0x7f0000000ec0)=[{{&(0x7f00000003c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000004c0)=[{&(0x7f0000000440)="da5c9e13a34dfe063a2c25ce8170b209081a9163979c20a0c82636f33479123ade5c213c72fbafc6636fe7e1c97df2ed7a72e8d0bf636da9ad81abee515072fd3a2dc855283e9c23c6c90e3603477cf1c7e70ce2d7081697a2b0c0c06ee75f4b084086fe6828e249faaf65463329e57b05", 0x71}, {&(0x7f0000001880)="e801e3dabb66d0e9bb11720fd55cc42d4b935e0ad4b80b33437dc4091bb237268373235f955fb253f53db94adef4a0be5d8b56a72f1390c08f60f81ae4822c77c5fa38c4606cf7423fdaf80b559db60e117d4419f8ddc6bdb75f9d71f10753b18cdc21064940d18cda52600e14861632d6d08d69cc2d0fd94c2fde1e9551c9dd406e36d78e64343feaa3276ef0a5b3ebb95159e84de393c5fb697ad5a44463027cce27c9324ab724712557ed6a196585e9a92f613d798e3b9c0984cede6397b695113fc741b83650d1fbab84d6f8be2aa970e80146ecacff903039737796cdc6b2631f2c634cd54ba75750a57d93c197fc74364bd32164e526ff6056e968495519f35feea23b164ea8b6615797018aab756dfcc8e50b7c5b015f04ec30830b18dec02e9178ab603097764c7e387aba3a2a7b664b039738b81d1637cdd331ef6f88932a8ea81b1353370c60a08f2fa240df1b727c626f90daf51df59327d4f901b5926aac1823ad91eeefa060017c54aad627fb8597cef64d28b270d87294bbbc84b02fa7a239d32aaaf2b6c711b0ab127960edc9591fc6b0138cf90daea89b5612ffce3b59ff81c4666226339966debfeee0e634aca254d38fdb321c69a88b7ef5663fe774f8ad2bfc01768321801e395f723e830be710ac6c801e7cb91f5a1d05fa4658f265e60ef083a213d016aa7013b06a5a58881a57bc596159bf0ef759331c14fe720a14995f383a3eb30aa4bb51c9bac35e891ddfcb9ba3dcca80d24a0ee216aa4990f9ceac735c3f1af8d934567be4f55a87e2637814681366455c9242411309fcc357a6a07c623f0e593058976adc243d0d61477d1d8865f01e248b42c2f52de232594ceb682dcbefeb024595139c3f652356f10eb3f9bc1238a3d81efa8790e54921ec8be2fd4ecc12521948da585fbb33f9628ba20d9a493d28710db1254e355162058b0b984eab64b6add35618c71920bd392403d3173bed176b9b10e62f747d2f2e25efea46a27fdb614e83cf0c1584d1fa5bdb1f0bd5ea5d5ea8964acce83f27fa5094ef0e226d0a136f5fe3933c862be2786a5846583318263f3b7a9ca50fad8c7c1347cb732594e102637e52780046b0a2e30618a14bc139cd9190fb7b95daeeda76207ad489ab7bbaf7e35d96f69089fee5e6d4d5e7be8b9876a7474130d28abbc9bae5a728dc5906798de62a61a97da600d24ba757af4a4bf7597f60535b0afede7420677bc999f83e8bdb395e8e5aabf0545d4ffb03e006c3a515ebc554ee675b323236cd73ce5485468b271d023f02c2b0d080bc82dbac7369007ace87316feb439ccaec7e078d7f2e87bec62b5722809fbf23a4666c403e0330eac6b1ac29c463d4939098ee9296957efe6bcb57e095c75cdb5e3c60117d2b7245e8ae7bec11a330bc81fdd6bf965fa3ebaef03b8749a29db6bdc31c123d9b3c93bd4a91449e9df1ad36f20707256f263ff9e33c67af2fb074fea38948424bf34a252c42b218672898bf8421ac0b159be38b42476ff8f2d9992ea6b176004fd6c37d05800da1f9b6ea8058d23b02d1ffd05b888fe41c227af39e8110761b7800f13efa08b5e03f86564c12001306bab7097d3327a5a1ba3fa039c5d87c0fe9d46c332c3acd082f6e55282f5a5beab8009ccd29cc02cf6f9877935fb50f678e424f7fc4a2bb6501be29b1141a458997a60edc16044ef85134bbef77c0c983ad10ba171217c04f4c8fab25eafb7fdc7e685441b7c05f2280760c3768b6e81a277bc71f57563adde4946f2e77843a6adde6a63f4b391ef9aced6ee4f20a88e4e495f44db37b4620984d026d6b97732b05b45c51f66f9418b5d9dec5a18e17ee55e4a46e28082b51f5d22f4f5ff20d184e60771b1f881ed2243c12f61e2db7f6d7fddb0b8692dc5e071446a342161059948b5dee5f7dea65e751ea03a7970db69e760f7de704a22be8959f87808474f46df00137ef9cecbcf8e0d11786df92a402354de36b0d9ec51c8fc84cb580cf606c7fcbfcbd2b20b4dd055614b4c8d10232f9316f58e495bdb517c054ec662f9d4db9a3f8fc6ba38eb1a916d8fb5f119f08f7fa6ae615e3c318b8af848970ff545ee76f4871be0714ad9b8aaf06f0bf25f2692ea82f1f1c9154b7447f05bd3c2ea64f99c95e5a19eaae6450f10923f8a46a57b52e70bb25619677d9edec808ab66ba25b20a5d6fce6e0bbf269d9abd6775cf07b0efd1f33f8ec496976b33425ad6c534f6fc5099ef3aed1306baf7070388a934926970c9ac102e1bd0fcdddc3179ac3abcb460edc7efcbdd20ac05554ac952749a880a41c87569587dd0aaebd7ddcdd70f2b94eb4d2dc5c934311c8c6bfe7b7152e7aa2b553e491a0563832eba740e6c18f2adb06909fb5be6ffd2b801f3730fcd07ccacd196ac885aee7f0caa2020255bafe3c01b737a8b641646281a79e3ef89e01b24be21407f629d90f66cb5711752d918cb79e02b9f2755c057fe829ca1a9c833fc7368302c3ba0a62db78268e37577431ef8c29cc4aa4351524cb71f5ce5d6c1abf1ba9d815640c4fac8f4b7933cc33bfb37695516651adb50337c725ba6e37652648758be0dc64598854e0b03133cf5c42b46ef980d22d9c4f4d283373a6c600e782881dffa8f0eada4d1e31d29560b60547ae71c5b01a55c2a8ac8f105f468d2e74cbfe2b3ea234805960f2793f1715bcd261355a296b229c1261814c98fbe5a0a9a50cacf2351ee014205fd335b69874da8ff787bf83f029c2a41bf1482206390da5e55900850d8415ab941e5b4e662fb46b54cb2c46f4a66919117f3cbfa168e94dd73036c6d957ef979f75e83bed6f44c3e61f7c096bb36801a4f5e13b2f51d32dcb05a7f52c66ec67f5c7d7c8a3f3a53cd9177882db640d7158fe903c0aeea4a3ea20370a65b2718415edcbc80fa971c5c0b011c41deeb58745e319e11682cc2f45d6efc1ad0a621f376e2c2d3f7ad3e3003be6cc247e7abe0e596152ebee966bb379858f6acaf59cc1fcd200dfeb95e7d80f08cbe88f45a4d460435185179cbe885eb130a40c2e4a07afb63642740e2fbd6804e978abdd4adc00765f4a0da81ccbac0d5d91cf0866b1f111a199d94e7ad6785c63a6576ba86a856b41515a77bb4d07559d4607c38550e63857fe8436cb78f01a56d9b91a3df43c9f006bee864af231d5da037283478904398b73744445e01a41fe8c983fa6adcdef8af14053d6643c58aff413f34ec3331fbc41c0f189eb3a02d3764b9d1d64c543a195bd5e6dc188fe694d7e670fc712fccc3dc2264d5a2c85998c0c51b07cb3ef7d14f1bd78bbe98a202be945c7fd7375a698b4858521ba33f54195332d39d3f16bf576fdd6a7ba4548a79ca509f666ce9242771863b557ae9bb4353cea6b1f69ecc0fb946dfd3d1b577878dabfd8a0930205fbaded24da0c6734dcd4540576e71c5c8ae99a1099cc5d72cd8b8e7c0fcfd3a1fe7ca29d9edaaaacdef83bd957b712891b616b7a7e68473ef3759c7bffe521479809e881fd21ab5feeb35296b02dc686735035f6b6924fd6331a0cb63453b4bc3d4fbba41c9df87716d3049322ff1e2c1c1951f2d3505fdfa0b89e7ae28488fecbd04b5e23bf48d9b7b80f400c5626f64e66d7e3ecc010588bf6313bc014a9d14a28b238edcef30fb698798f04d19e35260ffd7d90f25ef17ba4f02664efdced59b5a49838f2f6d393ebeabcdeaea5f76be2508bc8af5f5b2efd28f4119c955f9ab2d7ed1f3c21adb50efd0786a2d881e71b0283ec61144d26ba51d836e8789631c0541e0e0239462d001fa2f281a985d12af2c3149bb97196fb0bc231a32697306ee9d5f766f1dcec8c9ae7dd61bbc1386a2f1b232e4475a27aba38ab5897fd1e033a7b834a51d684984afdbe516b39e3569247332ec6b88e39bb024de66f3a95a20d9dfaa00bfbb47ef60d2c939ec1655a1a68b2792ec77f7c85f5a5c498f77a97efc94e37f6b74a492224e524d5e275871d52b3c106509c3c08f5786354ab2c19dfc7fd429aca72c8eb6edbc07f9757941ae40ec8d16593fe227dc88f44127a6cef17be8f7ef7302251be16efe01d412f19e0f0e8e709c7e2ae1edc9aef66e9f5cd8b8e9c26abf94b61ee37b23152bd07735feb888aa015c4d50f0453df9aa834752f0886ee209ed63705c7b4fd6a53aa7b3ea82f2d228bf4fd9218981a5d0a574cadf609dd4ea5c335dfb49d999d940715a41b400e8d81f58300b570a433da9b2086cabee279b18ce5daa1981892928c007796dac2cab04370ea90c15667e6d17b6c539b63a38ce23e32e96943fa232fb3a76a5812d98742fa06dc00359d3d08f7b8425c260cc0bb55dcadb74db458e6126d329675e92de7c4220a066c6dec5e091a745e21f404efd3b8233eeac7fb610f96594c19cee57f7f1fc18f7dbde713383782063fe8399791447ce6bb612bcb5fc3681e6e1fafcf9c34429eedaced8c8d3d9f38c19ec1d622bb7403903dcdc0d8ba4c9ef4b1a22b872103e1837b7a7040612ddf9840388e54d9e1a28906aaa5b10621677769c58ffd9cf4a4a1561ff742b4319e03c10d76b0db41be8f863f5be3155d56650c4cceb8467b49fd334f6cf434e13ceec39b7ffc0aa28ef18b4edd1f0de8de0b083302651bc21a2e4a53e7aa0f7ad020f0cad0e6626b39472891d99490080c0df99f98a470d5100073680d05aa0e5b35ff65212ad76897c26d076d231ae687d1c8877af3013544b56cb7926195b8846cfa8177e48374753f4fc47ea53c493b16af2ebff1209e92787f6a09ff0a366547cd4cdc8cabb35f50e317017bebb0b0cebf64be8c370915de36f91b94aefa5ed5b868d2f38158188163340adf41fea5885ddbc9cc9b1ec9aebd57ac87d5fce669667a8d9962e18379db0f95554e2aff442b3f5a0b1c6925b4037399bd6f95a05ea88c164d138bd4c2e28639d8329c18a8caa3bc948d763024797b0de771b183015e58adbdf0f9ac9bbb3bc3cbe0bd6bdc0388708f015d822999bfc40c2894d6ffaadf6b072dd0bf8482f5ce25533ea8f2ace4c863fffa8697517fde868b99e55bb2bdd477580d6fad957a7eb121ea1ce8b6f4e821e0470cca9c0828f4edd8c80182556c5495994ed747f8b80af8707a4e1d3df2617445cadc262040eaf8d8acdea3916cf2da8ae18d9caf1db16963493ece1046e268ace8fd8eeb0ebabe47d7bd8e7aba9b3aef56c9371c0b89c7447048029f2df5fd88c4856ec13d25539f1f9a8c4cae82efda0890aa1f6742e12216b44ee5f9ac5af25f4060ee48ccd7556ae77dc2d09f88032586dfcfd35d258e03c00e77284b90b27881af6e1476776b5db669a1c5770890076ef23953c73310a535a389e0cfc0d27d28162b7b25f9bf56ac1591dc4a70a3e789d33f71ca2e8f0b40c43adc07fcd17116a47a378256f24169002295a4a4bd8a0cd7ad63f3e3e1fff2a781761af3a20aa2fc016fb3925dd8df8436f1f2c46e399819792735dccce4d13585ba8817afc8982a2ea3c885c8916a19ed2454cdad8cf26effb5d0f204e5a67d3b35dcc8ea8b55f6aab5c8bfc59848a3f03a1d4a006e5504fb0e28a046199f44096ebbd07ea4c381ca7ea0414bb16aca7a84b16a72c7144aec667a6f8fb6cec84c3ecf21825388fbb41e7179466ae3002b72a560ba424258626a337d8823294b3b4f08a89ee15c80fa6481578464bc29e16baaa4a834dbb5ba4e55741df3ba92336e6196d13ec4c69ec129c419790caf302343b1434b5837ab81ca4ff80481eb709df8b49557eafd6dfce15fceeeb3cd", 0x1000}], 0x2, &(0x7f0000000900)=[@rights={{0x18, 0x1, 0x1, [r4, r9]}}, @rights={{0x20, 0x1, 0x1, [r14, r11, r3, r10]}}, @rights={{0x18, 0x1, 0x1, [r10, 0xffffffffffffffff]}}, @rights={{0x1c, 0x1, 0x1, [0xffffffffffffffff, r9, r14]}}, @cred={{0x1c, 0x1, 0x2, {r7, r13, 0xffffffffffffffff}}}, @rights={{0x20, 0x1, 0x1, [r5, 0xffffffffffffffff, 0xffffffffffffffff, r3]}}, @cred={{0x1c, 0x1, 0x2, {r6, 0x0, 0xee01}}}], 0xd0, 0x80c0}}, {{0x0, 0x0, &(0x7f0000000a80)=[{&(0x7f0000000a00)="272e3c537de42b5794264df9612c3c75bc17db110ee2b0202e733ac53e576fd89cc5652de684f124a9bca651e5b4d3f54ea24f2abfc7715f2fc76d2266bfa9e5e0191fbe3488a87a84675dc511f8334f86dc91b3d2bb0446ce86f318c750688ba33a6679a2c60f6026d4648cf21e054f1d1a787e059d", 0x76}], 0x1, &(0x7f0000000b40)=[@rights={{0x20, 0x1, 0x1, [r1, r4, r5, 0xffffffffffffffff]}}, @rights={{0x28, 0x1, 0x1, [r12, r1, r15, r3, r16, r1]}}], 0x48, 0xc000}}, {{&(0x7f0000000bc0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000c40)=[{&(0x7f0000003900)="5f3a97f9e25015454c451722e6b070dd976bffff79ebd1cbf69f8f812cf767bb65dc6b6a6861ee8e2dbd8caecd769f9d0df9c12eea3e33e25ddaa33669c106d8bf3ecde77fba93ca1f1eddf04bacd23d2d836022ec0b8fc83592e9c28854418666fe430736c342ccbb92ac3fcae6e8711bf8757b04982c2f16ce84976369c63e9e9242ce38947bbaa8ef1732dadd6548bed2d4fd3e070559d10c487f3de6be13978021990dbeb5766ef652715ef2852168b5756fe6b7c546f452ce190dccc9b50871b66f6eef183ad50d8b7af6d44a803409566059ca79c4e82bc39bc1648005f1556e74a5c7ea4c6fdc082de041bda951c43b31b3aa4900e75ef6b139672e9d71bac445cc8f5cd6bfccf5c65f39fc04e94c887c2581102c052d6be23a6d744103b12114772dbfe2b4eb9191be3f424f611161fba1830c72b12e6961407f13365b95046be707cae44c42cdcb0ef924b04b16789b4ae8dbd6d92efd691fcccd15364fa5f112cc861e5886a7698f4dfe0c78d0756a367316c0b0e6e9406517df499d586e3293e0a93a11c6eeb20b9e2f04950e6adf939dc2613f9a14c1ee802f4afc91c8229f8e32c520426c7847aa0924d2406602b22d7ce9fe1770d6d9bbb7971a9b04fb4c786b9802f9999da001e4910571fd5aaad469ba01e46db24f5b603c9e3a59737bd1497e616b287814387a1759380073e874a35cfada797a70af5c45d4444a3c8f362bb617b3f965c4fb0157d13349ca45369837c6915c5eecbff9e0751279b04a772c1ffc267887e19c67f1eb9f0bee24b55563f6d83386c69ad76fae32fb06267755d41cc384dea50116c1f48bd42c44b51b974ae8aba85464b484ebe3105aeb602d281cf150e41990aa6480354e1edde4977dbde707d461f60864e8e43bdecb2401437d3d95b7b705b5ec435f865870c6c20715b5fd58731176ef4263f5e99a2667e17f1c8a5f6dc1bd5cce5cb3b42e2f5667580a05f4cc8831702315489dc5f5fa377cab563bbdc12cd3909017099478a781668691fc07663f8323484035ef258f12aa1816478fb4a41e0f7338977b802b1a8baaeb665fd24c11987ad10e319deb12534c4bd86df6d1c4fb49b331f4557d2009e6cd07318db00ac887a2b7b95d716d44510c5332373f2eeacc73d27ac33c144abfc239eb61a258475913b0b50e01c540fe0613c883432834ea64aeb070773d331374bd46d49e887bb8d1a1bba08bda6340ae0ae150699dd0736035dcfc8c335687d38a518ed2ad16a14ac64c9cc64c684f79c613a3466a7549e0dd61d896b372e79ed636cb3c118248ca6e5f3baa0e4ae3810c5298c1c205bbff40d275fd41107d86aa2c3bd6f4ac734ec680b54daeeeff749442cb94aa40987f7e10dbb5ab1e65ecbb539c0bca2bfec2aaea3db3b94855399c9b376bd5918d27d3d2b186ceacbeda05c8beffa0b0d796d725ad464b5df722321169f61e52fe7c833bae87ea29c4232493248eeb60ff45ffa95fee7f19a2836d08161fb67bc5e14c8c22187252b2035baf9374de69443fefa4c3429e807338d6726678fb9148ccfc25596afb9fba7876c224c8186c7f26f3e5dc278ef12d7939b8482b6fbbc58316d853d8516cc09e9ac9db31dfb04a24aaad081b895ace0eb427b97fbd4c21906e7ba200bc259bd63f60ea13f49d704f0754b7acb377f6c643d74d8c309d56740e237bd3f4f7f58bd570061caca51616eaa0ff9e19326161a03ad6a0562afb922baa18320b96f5d4d58a4b2f7c8e7c764ba3121f33e8b31da3c108894e29b2573370ac45c96f46adadad1126ad398677c1fa3a9306eb371e48c6861acdfd648cc4067b159e5663ecefbb42d919349b9b427835e603957fffe0e6c0f81e7c7709d97751e15406aa9f8ad4489a42febcfad0711c34c28b68f9921053b28351a1dbbb6fcfdecd486bd997a8be7810766256fe99116cbe48ea92e56c94e9c36120b01b4963fc3075be34abc76952f3f75d7c3c7349c11ca37914d1d209091cab231d9e6ee91c66481c08fa15ca12c71760d2457c7d6740bb2bae9ec2a41347de011f0077cbf5e0c7ebb30c1875d508e96ef72fafaf563734a5006c26daa3ed20064bf2e72d33f5a005ecb32e0c84b7f874c1dfc771659a5a10bce778d1b5a4c66f59d7233ac5911a8789546ffe1a0b96f320b76e1bbb90acf052ff456fd32c6938c74937f77ba38471e14a839d7ac6404b2a0c79532a039ecaf72eba4fec1ab2b31aab1d2de10469813b527df28a231c3e21f3172b0b2aaa3f473ac13f1e2bd0268fcd82f8eae180656ec2510ff3635be7790169b61b3cc46b9eea019fdccb0ff3e592ac20fb95826b0f260afa804ac3a71f5a7a37e316d2f82b3dc2e61fceb9d9aa095f8cbb13cea7d5c36546e77360eb12c162b0879dd980f27a608ec5a30fe35cf89fe77907fa34a171c38bbc7fc605f505dc9e5ae60d40256c5443e590cd6c84cf838701c3f7d340c94858b9533744b1f9fb03b68dc4aeba146daecad1d988e23d81c82e8a35304f1cd28bc653151576ad8fd0ee3c87e0cb61a1fb2fb7620b024332420fb28638c5c306611b79427f236849d610ad05e6121c2024f4a5c2d376d525ac5ec1f9729f827ae9cc7b9628ffbebe3c60c9b88625fd5b0641b4d5747292a05df47a7057ecf44be8ec57b247e88a0715ffd6189fad35486cdcf1c7e7ac7f7d7e54719fb15b592cdab81166aa355dc892249115814bf6e60219fcc2cd8197ac5c8ce9c1718300865c281e8ab4aa1c25944e467cc83ae51617ca39153ca78290417ed008b2189a38d99f3c0565e1f6090ba5c1f28d1a0e05ad4f1e9f77203e69d9cecfa442ebecd6a737dcee4aa9f81844a4a43c0ba412d35175657754e31fe31869ffa9cca8d0de7f75e131acce5fbae908e9b5e4cf12c0e5d3a4543516ee7884b0b5283bede3d677041825593bb2f28879445886a5ec9da94e3c7f8e720a957dccb1b3f1994029a4dbec1983b23e468a46b9d1b55a18532f8e92eaf00022f883ce7b5142da95fb7612d8a8f62be0ad5beda06ac35732631f15568f7991208282b98448f9e931d3c2321d3cf1f426d3a882f33b97e015a818f56bc48056538ba9978d1351d4c9e87026363be4c099c2b787362908ab453fa487f9451a812c151075cfacea8a8bf4097a25218e3e1039ab7d433ca69c1a3915a4ae30ebf88e1705a1282c01b3240985f713c5bfd29b764cefa1c8d080add2f675f0b598530884d457724b12ce0a59bac5d551891545d06bb543809b46095ea42384c58ea46e4733f0031d5756929d7d659aaa493af28cd26bbef64da84226bb07d285aa29be3af34b71189850ed4557fc868030886d0e4c0a905f6e0507eef375b475fcc0340774147bea21d22ca41983a3a8619d8ff3e7c0555b1ca37c912a2d74e3ab395d23bd0e73d930865d937f9958d5b3e1dbf79230be6a5a160a55c768f5f0b3c619d2316a7c65db1a5aff75948c93c33e1490ec724f14394f30cf2a25fcaaad7660f6b731f78e16fdf72e2cfc39402da4894114e68d6e5a20d3f7d0fb420dcd0dbb8e8d1d1df49700c9f4f2260a586649c0203b96b7a978f899272282bae047577ef80335b1134cd0060afb14ffa55fb5789dd31bd33b85f7e52124f4da12f80ca6874f3bbef599367fe2fde029eaa785ae7aedf5e29138655c3fed9b0512cb88a3cc5ba20f00e23a7ee95e97f61c2ea34103de98887dbdab233fbabd74da63506b85774cb45c083ee38764d5a761435b7b1d0f5ec960b933931277bed7d8be4f1614a49de99149007af3954353cd1e5d2ce69c0267f97e1d2307e2bb18e647bd7758c344e4af039b3fae31b48178a2f20da319eb3b32c0f36e8a8d087e7c93e637985e67c492eece4aa6e202cdf2d58e24bce4579cbae648601643828cb3d24d260bfa982f87588bf3952aee373301f4e064118159b4d4ada8e2e95deb8bafb9d469b7c9a9097e41e10a4d58bd9ea1e01a357f3a5eb9be910fab91f45919268bca8912a2c7bec511ecee9c491569500845520492776592b36f9586ab1014e2359cb31b1521e4445c6fe2370060c4e467feefbcfd6ae081900d745e4e4d8f703728165b51e49a4bc705de04900efdf0a921d63bbacc50a3129c87aae89a3605b1a3af0e137c20b64cd0107812692113b1dbe97f629337398b2635fb67c8fef6dd9720320e1a004a6b627b04683c496eb4cbc46353a352b40b893324e6297d39d6f982b894d55d04ed97880d31b0e01a23da1e59b3c3e0366b5f75d456e6782b32b1e1cb5830cc4144dbe0af9c573f94ad45b93cb802899d62e1bbfe1fdeebfa1ac712ce8de634cf2fac1e7676f1afcac449c1dc07dde258af110b3049026a1502d7d16777c874b6b6c03ca7cb3587beafcdde70a01121df9809bdeade57e3770c97d37022870f655ba862bc9104765da255d849d3bf9addde849264f1e0a81b635f01256d10b6b1879a08895ef3ac44e9d5fb312e8c6a9230e929d2dfa2060997307012f9b0658622f59a68d3d342bbb32b90a4f1a5b486128cbdab7d2434a9b67e378b8f637671b77485cdf284f7a238e075b8ba8bd7ac6daa1d387842a09ada3a431018b0ac909892ee08cc476dd9b4f409e082f48f9098ff0f46e4096e2f30377ebc42440ce0b05ff4b8282404d658194ee43351afe9654bbd9cc8839efd0c0608fd21680625dd11136f7f9881de07bee0fe0abfa6283db93a26f5f28496015ca728a099ccb1a356f30efd3bd8f839bfbb280332afbd17cc04be347f07d5e5bbc7779f2769add22e33ab221258d85770a854c9d75cbc0e3e92bfa71cb430320c0ca1f3dc3edcbee9d1ea219082893599df3c8427ce48c1322cfde3771a380e72e79ef8ed527b7e1f8573e848f30f4fe935038a5f4219f6b87b4679e8b577586747a6c6d6ea11f5f4c4be0c41099ca8a63a34f66fe4067e2cd3d31b81f42fca4e3bc69d55d69fb14bafad4776334abbb076fedc6853bf2d203cd6a3024d1f7c63031d5da3abdb61a87722b0061c200b1ab7e14f797e9707e65a5c81271be26d784b5cfe339c1a4ce05026d812f253c9d06a5ba1731d11ae1036f999fdd157d78a7774d467c493c0c5c66bd6d50ca87020b55bec81d03710182a49f476670aa91b8a1459d073f398f87ca0157414b51a9ff8807a29c5cc8e8d2c53890d0d739f6a3e363bfea68b692cb408b156e67879f3e8147ed635ededca191bed8ba5846006005604f9dd8508fddaf8326a36ca05ede287b01acceb843b0a64caccbfffa6225a588a99c49ec159740d4bf04658cd30b01b6560b4c8958758b59eb45d6cfbaacca844ac216ca187b631c7c7c5de04c67f6a5307e70788db1e8e504d80ca7b2cc26fae42ef76b4c88810aae6be014b0d08f5c21229bd681f54f446e5afd39e42a54645c6ea350c8cb44ce179ca53c2e33106e7d6df09c46919c86879accc8e0de0820c4e04b5a39d19024107d7864ce06af96963f75c93d00cab7545053de0eedcf89ff861819605d1536ef84cc023f7ffbe6508bad83f56944523b47120bc229c19e9814ea2dda2fdaceb0aeb15327f6948f6e126e344a352aafa36ece8d5211c89ad4b87d843e1d81757c584df431c2f0488b4b6f5520cbb0f773a3f1b86cd6482af1ac79f1ab8e62c7623a4176f4f76232fe755d2dc53191b2aff207964dad1a4f8a33c72b267727a46203d9206dae651cf13f4e29ecd8499a9e1ffd450ec657858a58266270d4e07d7435dcfb4b12f4ccff15be62e2f0e01f95f5d166893b3cecbac231fb12cdd603650", 0x1000}], 0x1, &(0x7f0000000d80)=[@rights={{0x14, 0x1, 0x1, [r9]}}, @cred={{0x1c, 0x1, 0x2, {r6, r13}}}, @rights={{0x28, 0x1, 0x1, [r14, r14, r17, r0, r14, 0xffffffffffffffff]}}], 0x60, 0x20000000}}], 0x3, 0x810) (async, rerun: 64) ioctl$TIOCGSID(0xffffffffffffffff, 0x5429, &(0x7f0000000c80)=0x0) (async, rerun: 64) r19 = geteuid() r20 = getpid() sendmmsg$unix(r9, &(0x7f0000003780)=[{{&(0x7f00000005c0)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f00000007c0)=[{&(0x7f0000000640)="5502e05c616f8c64f3d251d340ddf300b826eeebbe4af65bd585179ab6a8fa08067da53f72f9cc830fc44a5559571ce58d68d4b1f672af2979eb2f20c7c8f50c530052b443a07311107fd7491406466461bacd318eab6185554371b13299da5865f24146450a5ad40f81f3cb1ef5bff365651d2b767640677b4ab80c469961ea55312dba574da5c5a65bfc57e05943b8b6b56cda4670054fac2f1aa1d054392c21ecd81b8745dc74da8af2768624fcf005ac54fae4c8fd4d95606c081ccf52213f559d9a4048f4ebf200de2b8e9cae85d36786bb6e77585f58e8c3156ca610459ee3880eeecdb961e9132aac2b28", 0xee}, {&(0x7f0000000740)="38c7f19bc7450d1d2b07a21e9b84a39a3ea37c4a0e4b859e570f1ca934e8cb36e919c4d292826d715a4300d3830c143e3ae4663c951ebc76b3444dd5c74938ca893d126b975f4b31c63c5fefcc4dca830dbd3683a72a7a3865bf0a4a9b58", 0x5e}], 0x2, &(0x7f0000000cc0)=[@rights={{0x28, 0x1, 0x1, [0xffffffffffffffff, r11, r8, 0xffffffffffffffff, r8, r12]}}, @cred={{0x1c, 0x1, 0x2, {0x0, r13, 0xee00}}}, @rights={{0x18, 0x1, 0x1, [r9, r14]}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0xffffffffffffffff, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {r18, r19}}}, @cred={{0x1c, 0x1, 0x2, {r20, 0xee01, 0xee01}}}], 0xc0, 0x20000000}}, {{&(0x7f0000000e40)=@file={0x1, './file0\x00'}, 0x6e, &(0x7f0000001380)=[{&(0x7f0000000ec0)}, {&(0x7f0000000f80)="8065135304d3f294f66324ccd2a095dcf2a4584dcab750bc7838d681d0e09c3a8ac933c010260406307614705613432dc92d4e667ecd0f243fae2c3a011f6977834542d46b2b2087d30d7c14dbffcb6af37e8af98ade2999c006447d77e768cdb8dce0f3e9fa45d9d9828d37437486ecb0262e12a3ad1a9b7b81e8a551d0e0839288706e933b4dde6ed5bd0a2c37f23b4aa671337c6c46d143a7ed9817873c2f3de30fd34cee91cc43c659e6892b471a5111a97cb8a05bb639e6429a16e94a463ae82781b7b0c11124ca9d77ef856794511f174203", 0xd5}, {&(0x7f0000001080)="4c60b8c815ffad7a51dd98197064359d78b728d08c9f8a4d48641b69a0e6031f923b19e609eec86341a05366d212f3bfa101bbd26531012b91e6193aeb5ceb74f0dd004672709ab03c7d6920d312e4c62cef0bc8cce5624ad8460c7d2d0bc9d1e6297d391458747d53b6cbe1766e069a7f96fe5a4f7dbf70a8afa22f", 0x7c}, {&(0x7f0000001100)="4c7d3b896799b13ab97eaecc33c996d5f5e313ed19e2541d6a02c4fdaac4711c0cfad392968eef16bc03030ba3efb04e2f4f0ce0af789e777307788103b75dd4ff5cd6f4082c60e5fda6669ae0b7c8ca6beeb13c84c8ac828da7cc4e042357ef44a6c0f86c7f54747bd464d3441e4e10504092a7da5f6df6310b24c636c128fc156694d1152d791403ab2cfb518d2bf05a0c7c", 0x93}, {&(0x7f00000011c0)="13bb43bfddc1a594e7cf652269f256f2386379b233a2f504b4925fb482bfc8dd0dab43d5c5e78c569cdaf61e48cdec330263bcbcce0fdd6021b2a38e4308aef40681c30217d653f13964e3d098066b0adfbcc0f166dd92c01057331ac9009d8ec740eb3b5a2b7e9a5b13718d", 0x6c}, {0x0}, {&(0x7f0000001280)="19524e2eb994bdfa7e8ea83336840dd0c9784c830e43e9a9d21ce0e9526a8971fed48dde7c46e2cee115fbb8494ced2ed7e5a80dd9ebd008f7cbd87b315f24bfe0628a71332eb46c317a3b6a87c038ff6930c8c734f73be0064025a2027776f238e98bd12937f9aa3ff189b97c9359720e076beb249ffd4d20b55faa1c6077bdd4fd3301fa15ea68b742f927cfb10e9426b4122c7c4f7c823a52b48b5bb05b95606bcc46f71cd5b12753e62cfc479f3d18aefec1bbbfbd719a96b5dbd410124c3c9efe04470b1eada180d3a19a3146b826d7dd9af868930b952c05250dcf7e78c32fea6e34b0098b562b82", 0xeb}], 0x7, 0x0, 0x0, 0x4000001}}, {{0x0, 0x0, &(0x7f0000002a00)=[{&(0x7f0000001400)="27eff6ed04ed460367b0e879cca2decaaac7f7cea9d0e07f3d398161a25cc966ee40b4b1199a9a5f75004757421d4f6f96b533e1f777995424ddd6965f8e51beaa5b792c6cc22362e49641848142b14987476528bb95d6459e5f4c1108899b7dff683285aa1b752b9b59ca862af6c24429ada6672640683998b2d8af9722b2f8013b7f4ab7470e6342e02ccb5d321d5cc544fd1740873febe43e7ede97ea45e9c675d4d466836a5ad324e8cfd0f7b626195ea0cfa9a6b0ac5eeed27b43a7595f0af33308c83a2e159392d1ec37869e41e1d7259e3e3c97aa7bd728873169e540b06340c5", 0xe4}, {&(0x7f0000001500)="fa468d4f5bcc6e0fdf39c803d4320b23e46d2dfd8d17379a0047731b179a6c8439fc40343e76ba6782b236cc560fb691d5476f5fdc7f4e66e0e2801be8070eb4b01744a48c405f1dbacccbd274c271cdb528769ed51f82098bcac234ce3c0be8268d32038131b2ce277aac037cfd804cd305ebf349b20f6f0f2082de50420ee304ff331c46909631", 0x88}, {&(0x7f00000015c0)="68f839df43745846d6ac9693371765fbce46b5de94a790ad1d6b424e29ae229cef1b", 0x22}, {&(0x7f00000016c0)="0495198ac0c55c56c05b6e98a65a2cf797f516709a07694484b484ed65bfeb90da19d718518721228d203d8194bfa359fd200f67358f9de982def981368592a8b66d61ca742f8ffe4f", 0x49}, {&(0x7f00000017c0)="cd3fe54a3657af0546570ecb28b93097db925e5ffdd8c6a9342d784e521b7d6187542958306d46d9bf766340d16abbd8f45c9cd831dcf9dba41e304d9d0bcb707d565cd6c96ae8679761c64b4c0934a25b31e18fbd9ba3d71d15947a492ce3c7aabdfb19cb23e4cae50d5171bc99f3015895d5d3a465e20dd41fc014caece1dcf7006b9faa55975756089c33812f9599b03b7725dc25270da6a7030ae674ca387716bd1a04bfa5c76441e7849314c0438f3235d3e8e32db2b5", 0xb9}, {&(0x7f0000001900)}, {&(0x7f0000002900)}], 0x7, 0x0, 0x0, 0x20000012}}, {{&(0x7f0000002ac0)=@file={0x0, '.\x00'}, 0x6e, &(0x7f0000002e80)=[{0x0}, {&(0x7f0000002c00)}, {0x0}], 0x3, &(0x7f0000002f40)=[@cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee00}}}, @cred={{0x1c, 0x1, 0x2, {0x0, 0x0, 0xffffffffffffffff}}}, @cred={{0x1c, 0x1, 0x2, {0xffffffffffffffff, 0x0, 0xee01}}}], 0x60, 0x24000010}}, {{&(0x7f0000003080)=@abs={0x1}, 0x6e, &(0x7f00000032c0)=[{&(0x7f0000003100)}, {&(0x7f0000003200)}], 0x2}}, {{&(0x7f0000003300)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000003440)=[{&(0x7f0000003380)}], 0x1, 0x0, 0x0, 0x24040090}}], 0x6, 0x0) sendmmsg$unix(r4, &(0x7f0000000380)=[{{&(0x7f0000000040)=@file={0x0, './file0\x00'}, 0x6e, &(0x7f0000000300)=[{&(0x7f00000000c0)="dbe3c5e14986b76e473b23d2da33bfb31b5601b0f2ebf4f5356ad25a4e639b694fd5c3c9a54be9d87e1c75f3ffc4450b638aa05a2acf5f9d22c84f4ea7fbcf282efb1f411b1bdb3844a8201efa81a16e2dc17f5a23c21305add3428ce7270134e68c3860674333c7419634d6f7f6f646c8c81e8a760d1e0e9af06a6b998c4a6365ca0f73617fd565add2bf44092b563f4bdc0e3e920e241637cef8b80837524ca2d360e6e388fdf4e6282fd917d0430711bb70595e978ee0dcf1259dc168b2e6133cf7da929445c69802c73af83374b00830", 0xd2}, {&(0x7f00000001c0)="5ff568cfa2c303437a4027b1caf18f6ef9edfc6b6573111c6d5129d322dfd0312e431037dfd2af59a9de1ef4eb4e05157e45da0a3d9331553f8c8b24e57dbf687b0d6137f6213494e0702990af7aecab2646f20488bd2c168ce7dcbd2f2233af6785475b9068ebb563dfed5afde7fc38e52b72ffdc3135262c68a4fd0754b9948df6a67ac2792638f4d83c6b31759560f26828dcf2b0c4774271f25e904e253d32f3", 0xa2}, {&(0x7f0000000280)="349644b1b9b883e5cdd313480f39c45522346cfc6640ee9c3fc239f2fb80f43b0c2e14782ad3854925a59a8229c6d5ca800531943db7a6f30751ff921ed656a285d5901552d6491efb67921c5f9f2eb993c9dec164b5467b7c6ea8951a60e30c106161dd15e6", 0x66}], 0x3, &(0x7f0000000340)=[@cred={{0x1c, 0x1, 0x2, {r6, r13, 0xee00}}}], 0x20}}], 0x1, 0x44808) (async) getuid() 00:28:49 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 4: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x1c}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:49 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0xf, &(0x7f0000002580)=@ringbuf={{}, {{0x18, 0x1, 0x1, 0x0, 0x1}}, {}, [], {}, {0x95, 0x1c}}, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:49 executing program 0: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:49 executing program 2: bpf$PROG_LOAD(0x5, 0x0, 0x0) 00:28:49 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 0: bpf$PROG_LOAD(0x5, 0x0, 0x0) 00:28:49 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:49 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 4: bpf$PROG_LOAD(0x5, 0x0, 0x0) 00:28:49 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:49 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:49 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) 00:28:49 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:49 executing program 4: openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) 00:28:49 executing program 0: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:49 executing program 2: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:49 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() 00:28:49 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) 00:28:49 executing program 3: openat$binderfs_ctrl(0xffffffffffffff9c, 0x0, 0x0, 0x0) 00:28:49 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) 00:28:49 executing program 0: bpf$PROG_LOAD(0x5, &(0x7f0000002840)={0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, '\x00', 0x0, 0x0, 0xffffffffffffffff, 0x8, 0x0, 0x0, 0x10, 0x0, 0x0, 0xffffffffffffffff}, 0x90) 00:28:49 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:49 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) 00:28:49 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:49 executing program 1: openat$binderfs_ctrl(0xffffffffffffff9c, 0x0, 0x0, 0x0) 00:28:49 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) getpid() 00:28:49 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) 00:28:49 executing program 4: socket$unix(0x1, 0x5, 0x0) 00:28:49 executing program 2: socketpair(0x0, 0x0, 0x0, &(0x7f0000000000)) 00:28:49 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) getpid() 00:28:49 executing program 1: openat$binderfs_ctrl(0xffffffffffffff9c, 0x0, 0x0, 0x0) 00:28:49 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) getpid() 00:28:49 executing program 3: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:49 executing program 4: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:49 executing program 0: socket$unix(0x1, 0x0, 0x0) 00:28:49 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) 00:28:49 executing program 2: getpid() 00:28:49 executing program 2: socketpair(0x0, 0x0, 0x0, &(0x7f0000000000)) 00:28:49 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:49 executing program 4: socket$unix(0x1, 0x0, 0x0) 00:28:49 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:49 executing program 3: socketpair(0x0, 0x0, 0x0, &(0x7f0000000000)) 00:28:49 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:49 executing program 1: socket$unix(0x1, 0x0, 0x0) 00:28:49 executing program 0: openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) 00:28:49 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:49 executing program 1: openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:49 executing program 3: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 4: socketpair(0x21, 0x0, 0x0, 0x0) 00:28:50 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 4: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x4000000) 00:28:50 executing program 2: socketpair(0x21, 0x0, 0x0, 0x0) 00:28:50 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 1: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x4000000) 00:28:50 executing program 2: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x2200000000000000) 00:28:50 executing program 3: socketpair(0x21, 0x0, 0x0, 0x0) 00:28:50 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 0: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x4000000) 00:28:50 executing program 2: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:50 executing program 0: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x2200000000000000) 00:28:50 executing program 3: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 2: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:50 executing program 1: add_key$fscrypt_v1(0x0, &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:50 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 2: openat$nci(0xffffffffffffff9c, 0x0, 0x2, 0x2200000000000000) 00:28:50 executing program 3: request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:50 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 4: add_key$fscrypt_v1(0x0, &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:50 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)) 00:28:50 executing program 3: add_key$fscrypt_v1(0x0, &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:50 executing program 0: request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:50 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:50 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:50 executing program 3: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)) 00:28:50 executing program 0: add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:50 executing program 3: request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:50 executing program 0: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)) 00:28:50 executing program 1: add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:50 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:50 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:50 executing program 3: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:50 executing program 1: request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:50 executing program 0: add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:50 executing program 1: getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)) 00:28:50 executing program 2: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:50 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:50 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:50 executing program 1: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, 0x0, 0x0, 0xfffffffffffffffb) 00:28:50 executing program 3: getresuid(0x0, &(0x7f00000002c0), &(0x7f0000000300)) 00:28:50 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:50 executing program 4: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:50 executing program 3: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:50 executing program 1: getresuid(0x0, &(0x7f00000002c0), &(0x7f0000000300)) 00:28:50 executing program 2: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, 0x0, 0x0, 0xfffffffffffffffb) 00:28:50 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:50 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:50 executing program 2: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:50 executing program 3: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, 0x0, 0x0, 0xfffffffffffffffb) 00:28:50 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:50 executing program 0: getresuid(0x0, &(0x7f00000002c0), &(0x7f0000000300)) 00:28:50 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:50 executing program 2: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) 00:28:50 executing program 4: request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) 00:28:50 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:50 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:50 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:50 executing program 1: getresuid(&(0x7f0000000280), 0x0, &(0x7f0000000300)) 00:28:50 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:50 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:50 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:50 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:50 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:50 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:50 executing program 0: getresuid(&(0x7f0000000280), 0x0, &(0x7f0000000300)) 00:28:50 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:50 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:50 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:50 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:50 executing program 0: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) 00:28:50 executing program 4: getresuid(&(0x7f0000000280), 0x0, &(0x7f0000000300)) 00:28:50 executing program 2: keyctl$restrict_keyring(0x1d, 0x0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:50 executing program 1: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:50 executing program 0: openat$binderfs_ctrl(0xffffffffffffff9c, 0x0, 0x0, 0x0) 00:28:50 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) 00:28:50 executing program 4: openat$binderfs_ctrl(0xffffffffffffff9c, 0x0, 0x0, 0x0) 00:28:50 executing program 2: getresuid(&(0x7f0000000280), &(0x7f00000002c0), 0x0) 00:28:50 executing program 0: keyctl$restrict_keyring(0x1d, 0x0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:50 executing program 0: openat$binderfs_ctrl(0xffffffffffffff9c, 0x0, 0x0, 0x0) 00:28:50 executing program 3: keyctl$restrict_keyring(0x1d, 0x0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:50 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:50 executing program 2: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:50 executing program 4: getresuid(&(0x7f0000000280), &(0x7f00000002c0), 0x0) 00:28:50 executing program 2: socketpair(0x0, 0x0, 0x0, &(0x7f0000000000)) 00:28:50 executing program 4: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 1: getresuid(&(0x7f0000000280), &(0x7f00000002c0), 0x0) 00:28:51 executing program 0: socket$unix(0x1, 0x0, 0x0) 00:28:51 executing program 3: openat$binderfs_ctrl(0xffffffffffffff9c, 0x0, 0x0, 0x0) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(0x0, &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 4: socketpair(0x0, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) keyctl$link(0x8, r0, 0x0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:51 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 1: socket$unix(0x1, 0x0, 0x0) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) keyctl$link(0x8, r0, 0x0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 1: r0 = add_key$fscrypt_v1(0x0, &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 4: socket$unix(0x1, 0x0, 0x0) 00:28:51 executing program 3: socketpair(0x0, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) keyctl$link(0x8, r0, 0x0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 4: r0 = add_key$fscrypt_v1(0x0, &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 3: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 4: socketpair(0x0, 0x0, 0x0, 0x0) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 4: r0 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, 0x0, r0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 2: socketpair(0x0, 0x0, 0x0, 0x0) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 4: socketpair(0x0, 0x0, 0x0, 0x0) 00:28:51 executing program 3: r0 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, 0x0, r0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 2: socket$unix(0x1, 0x0, 0x0) 00:28:51 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 3: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:51 executing program 4: r0 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, 0x0, r0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) 00:28:51 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 1: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, 0x0, 0x0, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 3: r0 = add_key$fscrypt_v1(0x0, &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 0: add_key$fscrypt_v1(0x0, &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:51 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 4: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, 0x0, 0x0, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 3: request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, 0x0, 0x0, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 3: add_key$fscrypt_v1(0x0, &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:51 executing program 0: request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 0: r0 = add_key$fscrypt_v1(0x0, 0x0, 0x0, 0x0, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 4: request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 1: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 3: socketpair(0x21, 0x0, 0x0, 0x0) 00:28:51 executing program 2: add_key$fscrypt_v1(0x0, &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:51 executing program 4: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, 0x0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 1: request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 3: r0 = add_key$fscrypt_v1(0x0, 0x0, 0x0, 0x0, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 2: request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 1: add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:51 executing program 4: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 3: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, 0x0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(0x0, 0x0, 0x0, 0x0, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 0: request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 1: socketpair(0x21, 0x0, 0x0, 0x0) 00:28:51 executing program 4: add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:51 executing program 3: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 0: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, 0x0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 1: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 3: request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 4: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 1: add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:51 executing program 0: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) keyctl$restrict_keyring(0x1d, r0, 0x0, 0x0) 00:28:51 executing program 1: socketpair(0x21, 0x0, 0x0, 0x0) 00:28:51 executing program 3: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, 0x0, 0x0, 0xfffffffffffffffb) 00:28:51 executing program 4: request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 0: request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) 00:28:51 executing program 1: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 2: socketpair(0x0, 0x0, 0x0, 0x0) 00:28:51 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 3: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 0: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, 0x0, 0x0, 0xfffffffffffffffb) 00:28:51 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r1 = request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 3: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 0: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 2: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, 0x0, 0x0, 0xfffffffffffffffb) 00:28:51 executing program 4: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r1 = request_key(0x0, 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:51 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 3: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:51 executing program 4: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:51 executing program 2: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r1 = request_key(0x0, 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 3: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) 00:28:52 executing program 1: request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) 00:28:52 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:52 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:52 executing program 3: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r1 = request_key(0x0, 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 4: add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:52 executing program 2: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:52 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:52 executing program 3: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:52 executing program 4: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r1 = request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 0: add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:52 executing program 1: getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)) 00:28:52 executing program 2: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:52 executing program 0: getresuid(0x0, &(0x7f00000002c0), &(0x7f0000000300)) 00:28:52 executing program 3: add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:52 executing program 4: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:52 executing program 1: add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r0 = request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) keyctl$link(0x8, 0x0, r0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 2: getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)) 00:28:52 executing program 0: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:52 executing program 3: getresuid(0x0, &(0x7f00000002c0), &(0x7f0000000300)) 00:28:52 executing program 2: add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:52 executing program 0: getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)) 00:28:52 executing program 3: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:52 executing program 1: getresuid(0x0, &(0x7f00000002c0), &(0x7f0000000300)) 00:28:52 executing program 4: add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r0 = request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) keyctl$link(0x8, 0x0, r0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 2: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:52 executing program 3: add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:52 executing program 1: request_key(0x0, 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:52 executing program 4: getresuid(0x0, 0x0, &(0x7f0000000300)) 00:28:52 executing program 0: add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r0 = request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) keyctl$link(0x8, 0x0, r0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 1: getresuid(0x0, &(0x7f00000002c0), &(0x7f0000000300)) 00:28:52 executing program 3: request_key(0x0, 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:52 executing program 4: add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) 00:28:52 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:52 executing program 0: getresuid(0x0, &(0x7f00000002c0), &(0x7f0000000300)) 00:28:52 executing program 2: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) keyctl$link(0x8, r0, 0x0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 4: getresuid(0x0, 0x0, 0x0) 00:28:52 executing program 2: add_key$fscrypt_v1(&(0x7f00000000c0), 0x0, 0x0, 0x0, 0xfffffffffffffffb) 00:28:52 executing program 3: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) keyctl$link(0x8, r0, 0x0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 0: request_key(0x0, 0x0, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) 00:28:52 executing program 2: getresuid(0x0, &(0x7f00000002c0), &(0x7f0000000300)) 00:28:52 executing program 1: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) 00:28:52 executing program 4: getresuid(&(0x7f0000000280), 0x0, &(0x7f0000000300)) 00:28:52 executing program 1: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) getpid() getpid() getuid() getuid() 00:28:52 executing program 3: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:52 executing program 0: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) keyctl$link(0x8, r0, 0x0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 2: sendmsg$unix(0xffffffffffffffff, 0x0, 0x0) openat$binderfs_ctrl(0xffffffffffffff9c, &(0x7f0000000840)='./binderfs/binder-control\x00', 0x0, 0x0) socket$unix(0x1, 0x5, 0x0) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x4000000) 00:28:52 executing program 4: getresuid(&(0x7f0000000280), 0x0, &(0x7f0000000300)) 00:28:52 executing program 0: request_key(&(0x7f00000001c0)='rxrpc\x00', 0x0, 0x0, 0x0) 00:28:52 executing program 4: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r1 = request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x0, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 1: getresuid(&(0x7f0000000280), 0x0, &(0x7f0000000300)) 00:28:52 executing program 0: getresuid(&(0x7f0000000280), 0x0, &(0x7f0000000300)) 00:28:52 executing program 2: getresuid(&(0x7f0000000280), &(0x7f00000002c0), 0x0) 00:28:52 executing program 0: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 0: getresuid(&(0x7f0000000280), 0x0, &(0x7f0000000300)) 00:28:52 executing program 3: add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 1: socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r0, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x2200000000000000) 00:28:52 executing program 4: getresuid(&(0x7f0000000280), 0x0, &(0x7f0000000300)) 00:28:52 executing program 2: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r1 = request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x0, 0x0, 0x0, &(0x7f0000000000)) 00:28:52 executing program 3: getresuid(&(0x7f0000000280), &(0x7f00000002c0), 0x0) 00:28:52 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:52 executing program 4: getresuid(0x0, 0x0, &(0x7f0000000300)) 00:28:52 executing program 2: getresuid(&(0x7f0000000280), &(0x7f00000002c0), 0x0) 00:28:52 executing program 1: r0 = add_key$fscrypt_v1(0x0, 0x0, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999"}, 0x48, 0xfffffffffffffffb) r1 = request_key(0x0, &(0x7f0000000200)={'syz', 0x1}, 0x0, 0x0) keyctl$link(0x8, r0, r1) keyctl$get_persistent(0x16, 0x0, 0x0) socketpair(0x0, 0x0, 0x0, 0x0) 00:28:52 executing program 0: r0 = add_key$fscrypt_v1(&(0x7f00000000c0), &(0x7f0000000100)={'fscrypt:', @desc3}, &(0x7f0000000140)={0x0, "f09d707e72262e8e07f170b21ac2e4d1421a85dd4a5c695f718294d916ecb781cd7253af46111967b3589246a45b8fff78c9b9f4512e00cb074b31ab24205999", 0x27}, 0x48, 0xfffffffffffffffb) r1 = request_key(&(0x7f00000001c0)='rxrpc\x00', &(0x7f0000000200)={'syz', 0x1}, &(0x7f0000000240)='/dev/virtual_nci\x00', 0x0) keyctl$link(0x8, r0, r1) getresuid(&(0x7f0000000280), &(0x7f00000002c0), &(0x7f0000000300)=0x0) keyctl$get_persistent(0x16, r2, 0x0) socketpair(0x21, 0x0, 0x0, &(0x7f0000000000)={0xffffffffffffffff, 0xffffffffffffffff}) sendmsg$unix(r3, &(0x7f0000000580)={0x0, 0x0, &(0x7f0000000500)=[{0x0}], 0x1}, 0x4008804) keyctl$restrict_keyring(0x1d, r0, &(0x7f0000000340)='encrypted\x00', 0x0) bind(r3, &(0x7f0000000040)=@pptp={0x18, 0x2, {0x3, @empty}}, 0x80) openat$nci(0xffffffffffffff9c, &(0x7f0000000880), 0x2, 0x0) 00:28:52 executing program 3: VM DIAGNOSIS: Warning: Permanently added '10.128.0.41' (ED25519) to the list of known hosts. lock-classes: 8180 [max: 8192] direct dependencies: 32626 [max: 131072] indirect dependencies: 171742 all direct dependencies: 2562211 dependency chains: 56465 [max: 65536] dependency chain hlocks used: 224551 [max: 327680] dependency chain hlocks lost: 0 in-hardirq chains: 130 in-softirq chains: 1670 in-process chains: 54664 stack-trace entries: 310889 [max: 1048576] number of stack traces: 14063 number of stack hash chains: 9387 combined max dependencies:hardirq-safe locks: 59 hardirq-unsafe locks: 7551 softirq-safe locks: 268 softirq-unsafe locks: 7236 irq-safe locks: 276 irq-unsafe locks: 7551 hardirq-read-safe locks: 4 hardirq-read-unsafe locks: 203 softirq-read-safe locks: 19 softirq-read-unsafe locks: 190 irq-read-safe locks: 19 irq-read-unsafe locks: 203 uncategorized locks: 346 unused locks: 2 max locking depth: 20 max bfs queue depth: 1439 max lock class index: 8191 debug_locks: 0 zapped classes: 20940 zapped lock chains: 192526 large chain blocks: 1 all lock classes: FD: 26 BD: 188 -...: (console_sem).lock ->&p->pi_lock FD: 173 BD: 6 +.+.: console_lock ->pool_lock#2 ->&obj_hash[i].lock ->&____s->seqcount ->&c->lock ->kbd_event_lock ->(console_sem).lock ->console_owner_lock ->fs_reclaim ->&x->wait#9 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#11 ->&fb_info->lock ->vt_event_lock ->&base->lock ->subsys mutex#6 ->&helper->lock ->&helper->damage_lock ->&rq->__lock ->&lock->wait_lock ->&p->pi_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->req_lock ->&x->wait#11 ->subsys mutex#24 ->&n->list_lock ->&cfs_rq->removed.lock ->&sem->wait_lock ->&____s->seqcount#2 FD: 1 BD: 1 ....: console_srcu FD: 29 BD: 1 +.+.: fill_pool_map-wait-type-override ->pool_lock#2 ->pool_lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 2 BD: 3257 -.-.: &obj_hash[i].lock ->pool_lock FD: 1 BD: 3258 -.-.: pool_lock FD: 732 BD: 17 +.+.: cgroup_mutex ->pcpu_alloc_mutex ->&c->lock ->&____s->seqcount ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&obj_hash[i].lock ->cgroup_file_kn_lock ->css_set_lock ->blkcg_pol_mutex ->percpu_counters_lock ->shrinker_mutex ->&base->lock ->devcgroup_mutex ->cpu_hotplug_lock ->fs_reclaim ->&n->list_lock ->&x->wait#2 ->&rq->__lock ->cgroup_mutex.wait_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->cgroup_rstat_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->cpuset_mutex ->&dom->lock ->batched_entropy_u32.lock ->cgroup_idr_lock ->task_group_lock ->(wq_completion)cpuset_migrate_mm ->&wq->mutex ->&____s->seqcount#2 ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->stock_lock ->key ->pcpu_lock ->&cfs_rq->removed.lock FD: 50 BD: 1 +.+.: fixmap_lock ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 FD: 227 BD: 90 ++++: cpu_hotplug_lock ->jump_label_mutex ->cpuhp_state_mutex ->wq_pool_mutex ->freezer_mutex ->rcu_tasks_trace__percpu.cbs_pcpu_lock ->&ACCESS_PRIVATE(rtpcp, lock) ->smpboot_threads_lock ->&obj_hash[i].lock ->&pool->lock ->&rq->__lock ->&x->wait#5 ->mem_hotplug_lock ->mem_hotplug_lock.waiters.lock ->mem_hotplug_lock.rss.gp_wait.lock ->cpu_hotplug_lock.rss.gp_wait.lock ->rcu_node_0 ->&swhash->hlist_mutex ->pmus_lock ->pcp_batch_high_lock ->&xa->xa_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->kthread_create_lock ->&p->pi_lock ->&x->wait ->wq_pool_attach_mutex ->pcpu_alloc_mutex ->sparse_irq_lock ->cpu_hotplug_lock.waiters.lock ->&x->wait#6 ->cpuhp_state-up ->stop_cpus_mutex ->&wq->mutex ->hrtimer_bases.lock ->flush_lock ->xps_map_mutex ->css_set_lock ->cpuset_mutex ->cgroup_threadgroup_rwsem ->cgroup_threadgroup_rwsem.waiters.lock ->cgroup_threadgroup_rwsem.rss.gp_wait.lock ->&list->lock#12 ->wq_pool_mutex.wait_lock ->(work_completion)(flush) ->&x->wait#10 ->&cfs_rq->removed.lock FD: 29 BD: 95 +.+.: jump_label_mutex ->patch_lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 97 BD: 13 +.+.: console_mutex ->&port_lock_key ->syslog_lock ->(console_sem).lock ->&rq->__lock ->&root->kernfs_rwsem ->kernfs_notify_lock FD: 1 BD: 113 ..-.: input_pool.lock FD: 1 BD: 3224 ..-.: base_crng.lock FD: 1 BD: 96 ....: patch_lock FD: 1 BD: 1 ....: rcu_read_lock FD: 1 BD: 1 ....: crng_init_wait.lock FD: 1 BD: 1 ....: early_pfn_lock FD: 1 BD: 8 ....: devtree_lock FD: 1 BD: 1 ....: rcu_read_lock_sched FD: 10 BD: 95 ++++: resource_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount FD: 1 BD: 1 ....: restart_handler_list.lock FD: 1 BD: 1 +.+.: system_transition_mutex FD: 3 BD: 604 ..-.: pcpu_lock ->stock_lock FD: 1 BD: 1 ....: debug_hook_lock FD: 2 BD: 1 ....: zonelist_update_seq ->zonelist_update_seq.seqcount FD: 1 BD: 2 ....: zonelist_update_seq.seqcount FD: 165 BD: 91 +.+.: cpuhp_state_mutex ->cpuhp_state-down ->cpuhp_state-up ->resource_lock ->pool_lock#2 ->(console_sem).lock ->clockevents_lock ->&irq_desc_lock_class ->&p->pi_lock ->&x->wait#6 ->&rq->__lock ->fs_reclaim ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->crypto_alg_sem ->scomp_lock FD: 2 BD: 899 ..-.: &zone->lock ->&____s->seqcount FD: 1 BD: 3216 .-.-: &____s->seqcount FD: 27 BD: 99 +.+.: &pcp->lock ->&zone->lock FD: 1 BD: 3281 -.-.: pool_lock#2 FD: 57 BD: 170 +.+.: pcpu_alloc_mutex ->pcpu_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->&obj_hash[i].lock ->&c->lock ->&rq->__lock ->&cfs_rq->removed.lock ->pcpu_alloc_mutex.wait_lock ->remove_cache_srcu ->rcu_node_0 ->&rcu_state.expedited_wq ->stock_lock ->key ->percpu_counters_lock ->purge_vmap_area_lock FD: 6 BD: 3208 -.-.: &n->list_lock ->&c->lock FD: 5 BD: 3228 -.-.: &c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 105 BD: 63 +.+.: slab_mutex ->pool_lock#2 ->pcpu_alloc_mutex ->&c->lock ->&____s->seqcount ->fs_reclaim ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rq->__lock ->lock ->&root->kernfs_rwsem ->&k->list_lock ->&obj_hash[i].lock FD: 66 BD: 3 +.+.: trace_types_lock ->fs_reclaim ->pool_lock#2 ->pin_fs_lock ->&sb->s_type->i_mutex_key#5 FD: 1 BD: 2 ....: panic_notifier_list.lock FD: 1 BD: 1 ....: die_chain.lock FD: 52 BD: 4 +.+.: trace_event_sem ->trace_event_ida.xa_lock ->&rq->__lock ->fs_reclaim ->batched_entropy_u8.lock ->kfence_freelist_lock ->pool_lock#2 ->eventfs_mutex ->&c->lock ->&____s->seqcount FD: 3 BD: 230 ..-.: batched_entropy_u32.lock ->crngs.lock FD: 2 BD: 3223 ..-.: crngs.lock ->base_crng.lock FD: 17 BD: 420 +.+.: sysctl_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 24 BD: 3106 -.-.: &rq->__lock ->&per_cpu_ptr(group->pcpu, cpu)->seq ->&obj_hash[i].lock ->&base->lock ->&cfs_rq->removed.lock ->&rt_b->rt_runtime_lock ->&cp->lock ->&rt_rq->rt_runtime_lock ->pool_lock#2 ->cpu_asid_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->cid_lock FD: 1 BD: 3107 ....: &cfs_b->lock FD: 25 BD: 1 ....: init_task.pi_lock ->&rq->__lock FD: 1 BD: 1 ....: init_task.vtime_seqcount FD: 61 BD: 94 +.+.: wq_pool_mutex ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&wq->mutex ->&obj_hash[i].lock ->fs_reclaim ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->wq_pool_attach_mutex ->&pool->lock ->&xa->xa_lock ->&n->list_lock ->&____s->seqcount#2 ->rcu_node_0 ->remove_cache_srcu ->wq_pool_mutex.wait_lock ->quarantine_lock ->&base->lock ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&rcu_state.expedited_wq FD: 32 BD: 110 +.+.: &wq->mutex ->&pool->lock ->&x->wait#10 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 29 BD: 564 -.-.: &pool->lock ->&obj_hash[i].lock ->&p->pi_lock ->pool_lock#2 ->(worker)->lock ->&base->lock ->&x->wait#10 FD: 50 BD: 73 +.+.: shrinker_mutex ->pool_lock#2 ->fs_reclaim ->&rq->__lock FD: 1 BD: 680 -.-.: rcu_node_0 FD: 10 BD: 52 -.-.: rcu_state.barrier_lock ->rcu_node_0 ->&obj_hash[i].lock FD: 30 BD: 3 ....: &rnp->exp_poll_lock FD: 9 BD: 5 ....: trace_event_ida.xa_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 FD: 1 BD: 1 ....: trigger_cmd_mutex FD: 30 BD: 239 +.+.: free_vmap_area_lock ->&obj_hash[i].lock ->pool_lock#2 ->init_mm.page_table_lock ->quarantine_lock ->&____s->seqcount ->&meta->lock ->kfence_freelist_lock ->&base->lock FD: 1 BD: 245 +.+.: vmap_area_lock FD: 244 BD: 1 ....: acpi_probe_mutex ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&zone->lock ->&____s->seqcount ->init_mm.page_table_lock ->resource_lock ->&c->lock ->cpu_hotplug_lock ->(console_sem).lock ->irq_domain_mutex ->pcpu_alloc_mutex ->&domain->mutex ->&desc->request_mutex ->&irq_desc_lock_class ->cpu_pm_notifier.lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->iort_msi_chip_lock ->its_lock ->efi_mem_reserve_persistent_lock ->lpi_range_lock ->syscore_ops_lock ->clocksource_mutex FD: 27 BD: 240 +.+.: init_mm.page_table_lock ->&obj_hash[i].lock ->&base->lock FD: 50 BD: 5 +.+.: irq_domain_mutex ->pool_lock#2 ->fs_reclaim FD: 121 BD: 7 +.+.: &domain->mutex ->sparse_irq_lock ->pool_lock#2 ->&irq_desc_lock_class ->fs_reclaim ->&obj_hash[i].lock ->&its->dev_alloc_lock FD: 118 BD: 97 +.+.: sparse_irq_lock ->&____s->seqcount ->pool_lock#2 ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&c->lock ->(cpu_running).wait.lock ->&base->lock ->&rq->__lock ->(&timer.timer) ->&x->wait#6 ->&p->pi_lock ->&irq_desc_lock_class ->fs_reclaim ->lock ->&root->kernfs_rwsem ->proc_subdir_lock ->&ent->pde_unload_lock ->proc_inum_ida.xa_lock ->sysfs_symlink_target_lock ->kernfs_idr_lock FD: 7 BD: 110 -.-.: &irq_desc_lock_class ->irq_controller_lock ->mask_lock ->&its->lock ->irq_resend_lock ->tmp_mask_lock FD: 35 BD: 17 +.+.: &desc->request_mutex ->&irq_desc_lock_class ->proc_subdir_lock ->&ent->pde_unload_lock ->proc_inum_ida.xa_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock FD: 1 BD: 2 ....: cpu_pm_notifier.lock FD: 15 BD: 186 +.+.: purge_vmap_area_lock ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->&base->lock ->&____s->seqcount FD: 1 BD: 3 +.+.: iort_msi_chip_lock FD: 2 BD: 2 ....: its_lock ->&its->lock FD: 1 BD: 2 ....: efi_mem_reserve_persistent_lock FD: 4 BD: 9 +.+.: lpi_range_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 2 +.+.: syscore_ops_lock FD: 1 BD: 115 ....: &its->lock FD: 1 BD: 92 +.+.: cpuhp_state-down FD: 155 BD: 92 +.+.: cpuhp_state-up ->smpboot_threads_lock ->sparse_irq_lock ->&swhash->hlist_mutex ->pmus_lock ->&x->wait#4 ->&obj_hash[i].lock ->hrtimer_bases.lock ->wq_pool_mutex ->rcu_node_0 ->resource_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&rq->__lock ->fs_reclaim ->lock ->&root->kernfs_rwsem ->&x->wait#9 ->&k->list_lock ->bus_type_sem ->&k->k_lock ->&pcp->lock ->swap_slots_cache_mutex FD: 4 BD: 1 -.-.: timekeeper_lock ->tk_core.seq.seqcount FD: 3 BD: 3149 ----: tk_core.seq.seqcount ->&obj_hash[i].lock FD: 1 BD: 111 ....: irq_controller_lock FD: 7 BD: 92 ....: clockevents_lock ->tk_core.seq.seqcount ->tick_broadcast_lock ->jiffies_seq.seqcount FD: 3 BD: 93 -...: tick_broadcast_lock ->jiffies_lock FD: 1 BD: 95 -.-.: jiffies_seq.seqcount FD: 233 BD: 2 +.+.: clocksource_mutex ->cpu_hotplug_lock ->(console_sem).lock FD: 11 BD: 3164 -.-.: &base->lock ->&obj_hash[i].lock FD: 3 BD: 5 ....: batched_entropy_u64.lock ->crngs.lock FD: 125 BD: 93 +.+.: pmus_lock ->pcpu_alloc_mutex ->pool_lock#2 ->&obj_hash[i].lock ->&cpuctx_mutex ->fs_reclaim ->&k->list_lock ->lock ->&root->kernfs_rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->&x->wait#9 ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->subsys mutex#29 FD: 1 BD: 93 +.+.: &swhash->hlist_mutex FD: 1 BD: 94 +.+.: &cpuctx_mutex FD: 1 BD: 7 ....: tty_ldiscs_lock FD: 2 BD: 7 ....: kbd_event_lock ->led_lock FD: 1 BD: 8 ..-.: led_lock FD: 1 BD: 160 ....: console_owner_lock FD: 40 BD: 3 +.+.: init_task.alloc_lock ->init_fs.lock FD: 53 BD: 2 +.+.: acpi_ioremap_lock ->pool_lock#2 ->fs_reclaim ->&____s->seqcount ->&c->lock ->free_vmap_area_lock ->vmap_area_lock FD: 1 BD: 18 ....: semaphore->lock FD: 1 BD: 17 +.+.: *(&acpi_gbl_reference_count_lock) FD: 13 BD: 3133 -.-.: hrtimer_bases.lock ->tk_core.seq.seqcount ->&obj_hash[i].lock FD: 1 BD: 578 ..-.: percpu_counters_lock FD: 35 BD: 2 +.+.: tomoyo_policy_lock ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&c->lock ->&rq->__lock ->&n->list_lock ->tomoyo_policy_lock.wait_lock FD: 908 BD: 4 ++++: pernet_ops_rwsem ->stack_depot_init_mutex ->crngs.lock ->net_rwsem ->proc_inum_ida.xa_lock ->pool_lock#2 ->proc_subdir_lock ->fs_reclaim ->&____s->seqcount ->&c->lock ->sysctl_lock ->pcpu_alloc_mutex ->net_generic_ids.xa_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&dir->lock ->&obj_hash[i].lock ->k-sk_lock-AF_NETLINK ->k-slock-AF_NETLINK ->nl_table_lock ->nl_table_wait.lock ->rtnl_mutex ->uevent_sock_mutex ->&net->rules_mod_lock ->slab_mutex ->batched_entropy_u32.lock ->percpu_counters_lock ->k-slock-AF_INET/1 ->&zone->lock ->cache_list_lock ->rcu_node_0 ->&rq->__lock ->tk_core.seq.seqcount ->&k->list_lock ->lock ->&root->kernfs_rwsem ->running_helpers_waitq.lock ->&sn->pipefs_sb_lock ->krc.lock ->&s->s_inode_list_lock ->pool_lock ->nf_hook_mutex ->cpu_hotplug_lock ->rtnl_mutex.wait_lock ->&p->pi_lock ->hwsim_netgroup_ida.xa_lock ->&cfs_rq->removed.lock ->nf_ct_ecache_mutex ->nf_log_mutex ->ipvs->est_mutex ->&base->lock ->__ip_vs_app_mutex ->&hashinfo->lock#2 ->&net->ipv6.ip6addrlbl_table.lock ->(console_sem).lock ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->k-clock-AF_INET6 ->wq_pool_mutex ->pcpu_lock ->&list->lock#4 ->&dir->lock#2 ->ptype_lock ->k-clock-AF_TIPC ->k-sk_lock-AF_TIPC ->k-slock-AF_TIPC ->&this->receive_lock ->once_lock ->nf_ct_proto_mutex ->k-sk_lock-AF_RXRPC ->k-slock-AF_RXRPC ->&rxnet->conn_lock ->&call->waitq ->&rx->call_lock ->&rxnet->call_lock ->&n->list_lock ->uevent_sock_mutex.wait_lock ->rdma_nets.xa_lock ->devices_rwsem ->&____s->seqcount#2 ->&sem->wait_lock ->remove_cache_srcu ->&net->nsid_lock ->ebt_mutex ->nf_nat_proto_mutex ->&xt[i].mutex ->&nft_net->commit_mutex ->netns_bpf_mutex ->&x->wait#2 ->(&net->fs_probe_timer) ->&net->cells_lock ->(&net->cells_timer) ->bit_wait_table + i ->(&net->fs_timer) ->(wq_completion)kafsd ->&wq->mutex ->k-clock-AF_RXRPC ->&local->services_lock ->(wq_completion)krxrpcd ->rlock-AF_RXRPC ->&x->wait ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&ent->pde_unload_lock ->ovs_mutex ->(work_completion)(&(&ovs_net->masks_rebalance)->work) ->(work_completion)(&ovs_net->dp_notify_work) ->&srv->idr_lock ->&rnp->exp_wq[1] ->(work_completion)(&tn->work) ->&rnp->exp_wq[2] ->&tn->nametbl_lock ->&rnp->exp_wq[3] ->&rnp->exp_wq[0] ->(work_completion)(&ht->run_work) ->&ht->mutex ->(work_completion)(&(&c->work)->work) ->(wq_completion)krdsd ->(work_completion)(&rtn->rds_tcp_accept_w) ->rds_tcp_conn_lock ->loop_conns_lock ->(wq_completion)l2tp ->rcu_state.barrier_mutex ->(&rxnet->peer_keepalive_timer) ->(work_completion)(&rxnet->peer_keepalive_work) ->(&rxnet->service_conn_reap_timer) ->&x->wait#10 ->dev_base_lock ->lweventlist_lock ->napi_hash_lock ->netdev_unregistering_wq.lock ->&fn->fou_lock ->ipvs->sync_mutex ->hwsim_radio_lock ->pin_fs_lock ->&dentry->d_lock ->&sb->s_type->i_mutex_key#3 ->&sb->s_type->i_lock_key#7 ->mount_lock ->(inetaddr_chain).rwsem ->inet6addr_chain.lock ->rcu_state.barrier_mutex.wait_lock ->(work_completion)(&local->restart_work) ->&list->lock#16 ->&rdev->wiphy.mtx ->(work_completion)(&rfkill->uevent_work) ->(work_completion)(&rfkill->sync_work) ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->&k->k_lock ->sysfs_symlink_target_lock ->subsys mutex#40 ->&x->wait#9 ->dpm_list_mtx ->&dev->power.lock ->deferred_probe_mutex ->device_links_lock ->&rfkill->lock ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->triggers_list_lock ->leds_list_lock ->(work_completion)(&rdev->wiphy_work) ->(work_completion)(&rdev->conn_work) ->(work_completion)(&rdev->event_work) ->(work_completion)(&(&rdev->dfs_update_channels_wk)->work) ->(work_completion)(&(&rdev->background_cac_done_wk)->work) ->(work_completion)(&rdev->destroy_work) ->(work_completion)(&rdev->propagate_radar_detect_wk) ->(work_completion)(&rdev->propagate_cac_done_wk) ->(work_completion)(&rdev->mgmt_registrations_update_wk) ->(work_completion)(&rdev->background_cac_abort_wk) ->subsys mutex#53 ->gdp_mutex ->(&local->sta_cleanup) ->rdma_nets_rwsem ->&rcu_state.gp_wq ->&rcu_state.expedited_wq ->k-clock-AF_NETLINK ->&nlk->wait ->quarantine_lock ->&wg->device_update_lock ->&bat_priv->forw_bcast_list_lock ->&bat_priv->forw_bat_list_lock ->&bat_priv->gw.list_lock ->(work_completion)(&(&bat_priv->bat_v.ogm_wq)->work) ->&bat_priv->bat_v.ogm_buff_mutex ->&bat_priv->tvlv.container_list_lock ->&bat_priv->tvlv.handler_list_lock ->(work_completion)(&(&bat_priv->nc.work)->work) ->key#17 ->key#18 ->(work_completion)(&(&bat_priv->dat.work)->work) ->&hash->list_locks[i] ->(work_completion)(&(&bat_priv->bla.work)->work) ->key#20 ->(work_completion)(&(&bat_priv->mcast.work)->work) ->(work_completion)(&(&bat_priv->tt.work)->work) ->key#16 ->key#21 ->&bat_priv->tt.req_list_lock ->&bat_priv->tt.changes_list_lock ->&bat_priv->tt.roam_list_lock ->(work_completion)(&(&bat_priv->orig_work)->work) ->key#19 ->wq_mayday_lock ->&hn->hn_lock ->&pnettable->lock ->&pnetids_ndev->lock ->k-sk_lock-AF_INET6/1 ->&net->sctp.addr_wq_lock ->k-sk_lock-AF_INET ->k-slock-AF_INET#2 ->&sn->gssp_lock ->&cd->hash_lock ->(&net->can.stattimer) ->vmap_area_lock ->purge_vmap_area_lock ->stock_lock ->xfrm_state_gc_work ->&net->xfrm.xfrm_state_lock ->&sb->s_type->i_lock_key#23 ->rename_lock.seqcount ->(work_completion)(&(&net->ipv6.addr_chk_work)->work) ->ip6_fl_lock ->(&net->ipv6.ip6_fib_timer) ->__ip_vs_mutex ->(&ipvs->dest_trash_timer) ->(work_completion)(&(&ipvs->expire_nodest_conn_work)->work) ->(work_completion)(&(&ipvs->defense_work)->work) ->(work_completion)(&(&ipvs->est_reload_work)->work) ->nfnl_subsys_ipset ->recent_lock ->hashlimit_mutex ->trans_gc_work ->nf_conntrack_mutex ->(work_completion)(&(&cnet->ecache.dwork)->work) ->tcp_metrics_lock ->k-clock-AF_INET ->(work_completion)(&net->xfrm.policy_hash_work) ->&net->xfrm.xfrm_policy_lock ->(work_completion)(&net->xfrm.state_hash_work) ->&xa->xa_lock#4 ->genl_sk_destructing_waitq.lock ->&lock->wait_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->rcu_state.exp_mutex.wait_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->pcpu_alloc_mutex.wait_lock ->kn->active#4 ->(wq_completion)tipc_rcv#6 ->(wq_completion)tipc_send#6 ->(wq_completion)tipc_crypto#6 ->(wq_completion)phy228 ->(wq_completion)phy227 ->(wq_completion)bond0#113 ->(wq_completion)tipc_rcv#7 ->(wq_completion)tipc_send#7 ->(wq_completion)tipc_crypto#7 ->(wq_completion)phy226 ->(wq_completion)phy225 ->(wq_completion)bond0#112 ->(wq_completion)tipc_rcv#8 ->(wq_completion)tipc_send#8 ->(wq_completion)tipc_crypto#8 ->(wq_completion)phy230 ->(wq_completion)phy229 ->(wq_completion)bond0#114 ->(wq_completion)tipc_rcv#9 ->(wq_completion)tipc_send#9 ->(wq_completion)tipc_rcv#10 ->(wq_completion)tipc_send#10 ->(wq_completion)tipc_crypto#9 ->(wq_completion)tipc_crypto#10 ->(wq_completion)phy214 ->(wq_completion)phy213 ->(wq_completion)phy222 ->(wq_completion)phy221 ->(wq_completion)bond0#106 ->(wq_completion)bond0#110 FD: 25 BD: 53 +.+.: stack_depot_init_mutex ->&rq->__lock FD: 38 BD: 73 ++++: net_rwsem ->&list->lock#2 ->&rq->__lock ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&cfs_rq->removed.lock ->quarantine_lock ->&n->list_lock ->&____s->seqcount#2 ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 2 BD: 114 ..-.: proc_inum_ida.xa_lock ->pool_lock#2 FD: 723 BD: 48 +.+.: rtnl_mutex ->&c->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->fs_reclaim ->pcpu_alloc_mutex ->&xa->xa_lock#4 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#20 ->&dir->lock#2 ->dev_hotplug_mutex ->dev_base_lock ->input_pool.lock ->nl_table_lock ->nl_table_wait.lock ->net_rwsem ->batched_entropy_u32.lock ->&tbl->lock ->sysctl_lock ->krc.lock ->&rq->__lock ->stack_depot_init_mutex ->cpu_hotplug_lock ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&cfs_rq->removed.lock ->wq_pool_mutex ->crngs.lock ->lweventlist_lock ->rtnl_mutex.wait_lock ->proc_subdir_lock ->proc_inum_ida.xa_lock ->&k->k_lock ->param_lock ->(console_sem).lock ->&rdev->wiphy.mtx ->&base->lock ->subsys mutex#55 ->&sdata->sec_mtx ->&local->iflist_mtx#2 ->lock#7 ->failover_lock ->&tn->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&idev->mc_lock ->&ndev->lock ->&pnettable->lock ->smc_ib_devices.mutex ->&(&net->nexthop.notifier_chain)->rwsem ->reg_requests_lock ->reg_pending_beacons_lock ->devnet_rename_sem ->&x->wait#2 ->&nft_net->commit_mutex ->&ent->pde_unload_lock ->target_list_lock ->rlock-AF_NETLINK ->(inetaddr_validator_chain).rwsem ->(inetaddr_chain).rwsem ->_xmit_LOOPBACK ->netpoll_srcu ->&n->list_lock ->&in_dev->mc_tomb_lock ->&im->lock ->fib_info_lock ->cbs_list_lock ->(inet6addr_validator_chain).rwsem ->&net->ipv6.addrconf_hash_lock ->&ifa->lock ->&tb->tb6_lock ->&dev_addr_list_lock_key ->napi_hash_lock ->lapb_list_lock ->x25_neigh_list_lock ->console_owner_lock ->console_owner ->_xmit_ETHER ->_xmit_SLIP ->quarantine_lock ->remove_cache_srcu ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->&cma->lock ->cma_mutex ->rcu_node_0 ->&sem->wait_lock ->&rfkill->lock ->_xmit_VOID ->_xmit_X25 ->&lapbeth->up_lock ->&lapb->lock ->class ->(&tbl->proxy_timer) ->&dev->tx_global_lock ->&rnp->exp_wq[2] ->&sch->q.lock ->&rnp->exp_wq[3] ->&dir->lock ->&ul->lock#2 ->&n->lock ->&wpan_dev->association_lock ->dev_addr_sem ->_xmit_IEEE802154 ->&nr_netdev_addr_lock_key ->listen_lock ->uevent_sock_mutex.wait_lock ->pool_lock ->&r->consumer_lock ->&mm->mmap_lock ->pcpu_lock ->(switchdev_blocking_notif_chain).rwsem ->&br->hash_lock ->nf_hook_mutex ->j1939_netdev_lock ->key ->percpu_counters_lock ->&bat_priv->tvlv.handler_list_lock ->&bat_priv->tvlv.container_list_lock ->&bat_priv->softif_vlan_list_lock ->key#16 ->&bat_priv->tt.changes_list_lock ->kernfs_idr_lock ->noop_qdisc.q.lock ->tk_core.seq.seqcount ->&wq->mutex ->init_lock ->&rnp->exp_wq[0] ->&rnp->exp_wq[1] ->deferred_lock ->&br->lock ->&pn->hash_lock ->&rcu_state.expedited_wq ->hrtimer_bases.lock ->&hard_iface->bat_iv.ogm_buff_mutex ->ptype_lock ->_xmit_NONE ->lock#9 ->&hsr->list_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->mount_lock ->&meta->lock ->&xa->xa_lock#18 ->&dev_addr_list_lock_key#3/1 ->req_lock ->&x->wait#11 ->subsys mutex#75 ->bpf_devs_lock ->(work_completion)(&(&devlink_port->type_warn_dw)->work) ->&devlink_port->type_lock ->&vn->sock_lock ->&wg->device_update_lock ->_xmit_SIT ->&bridge_netdev_addr_lock_key/1 ->_xmit_TUNNEL ->_xmit_IPGRE ->_xmit_TUNNEL6 ->&dev_addr_list_lock_key/1 ->&dev_addr_list_lock_key#2/1 ->_xmit_ETHER/1 ->&nn->netlink_tap_lock ->&batadv_netdev_addr_lock_key/1 ->&vlan_netdev_addr_lock_key/1 ->&macvlan_netdev_addr_lock_key/1 ->&ipvlan->addrs_lock ->&macsec_netdev_addr_lock_key/1 ->key#20 ->&bat_priv->tt.commit_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->k-slock-AF_INET/1 ->k-sk_lock-AF_INET ->k-slock-AF_INET#2 ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->&ul->lock ->&____s->seqcount#2 ->&tun->lock ->dev->qdisc_tx_busylock ?: &qdisc_tx_busylock ->__ip_vs_mutex ->flowtable_lock ->&idev->mc_query_lock ->(work_completion)(&(&idev->mc_report_work)->work) ->&hwstats->hwsdev_list_lock ->&net->xdp.lock ->mirred_list_lock ->&idev->mc_report_lock ->&sb->s_type->i_lock_key#23 ->&dentry->d_lock ->rename_lock.seqcount ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&pnn->pndevs.lock ->&pnn->routes.lock ->dev_pm_qos_sysfs_mtx ->deferred_probe_mutex ->device_links_lock ->&net->xfrm.xfrm_state_lock ->&net->xfrm.xfrm_policy_lock ->&sb->s_type->i_lock_key#7 ->(work_completion)(&wdev->disconnect_wk) ->(work_completion)(&wdev->pmsr_free_wk) ->&rdev->dev_wait ->&fq->lock ->netlbl_unlhsh_lock ->nr_list_lock ->nr_neigh_list_lock ->&bpq_netdev_addr_lock_key ->&app->lock#2 ->(&app->join_timer)#2 ->(&app->periodic_timer) ->&list->lock#11 ->(&app->join_timer) ->&app->lock ->&list->lock#10 ->(work_completion)(&(&priv->scan_result)->work) ->(work_completion)(&(&priv->connect)->work) ->(&hsr->prune_timer) ->(&hsr->announce_timer) ->rcu_state.exp_mutex.wait_lock ->key#19 ->&bat_priv->forw_bcast_list_lock ->&bat_priv->forw_bat_list_lock ->(work_completion)(&(&forw_packet_aggr->delayed_work)->work) ->(&pmctx->ip6_mc_router_timer) ->(&pmctx->ip4_mc_router_timer) ->(work_completion)(&ht->run_work) ->&ht->mutex ->&br->multicast_lock ->(work_completion)(&(&br->gc_work)->work) ->dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 ->&table->hash[i].lock ->k-clock-AF_INET6 ->k-clock-AF_INET ->&r->consumer_lock#2 ->&wg->socket_update_lock ->(work_completion)(&(&bond->mii_work)->work) ->(work_completion)(&(&bond->arp_work)->work) ->(work_completion)(&(&bond->alb_work)->work) ->(work_completion)(&(&bond->ad_work)->work) ->(work_completion)(&(&bond->mcast_work)->work) ->(work_completion)(&(&bond->slave_arr_work)->work) ->(&br->hello_timer) ->(&br->topology_change_timer) ->(&br->tcn_timer) ->(&brmctx->ip4_mc_router_timer) ->(&brmctx->ip4_other_query.timer) ->(&brmctx->ip4_other_query.delay_timer) ->(&brmctx->ip4_own_query.timer) ->(&brmctx->ip6_mc_router_timer) ->(&brmctx->ip6_other_query.timer) ->(&brmctx->ip6_other_query.delay_timer) ->(&brmctx->ip6_own_query.timer) ->raw_notifier_lock ->bcm_notifier_lock ->isotp_notifier_lock ->(work_completion)(&port->bc_work) ->(work_completion)(&port->wq) ->(work_completion)(&(&slave->notify_work)->work) ->_xmit_NETROM#2 ->&pmc->lock ->(&mp->timer) ->(work_completion)(&br->mcast_gc_work) ->rcu_state.barrier_mutex ->&caifn->caifdevs.lock ->&net->rules_mod_lock ->(&mrt->ipmr_expire_timer) ->stock_lock ->key#24 ->reg_indoor_lock ->pcpu_alloc_mutex.wait_lock ->&rnp->exp_lock ->dev_pm_qos_sysfs_mtx.wait_lock ->rcu_state.exp_mutex ->gdp_mutex.wait_lock ->&lock->wait_lock ->&x->wait#10 ->(inetaddr_chain).rwsem.wait_lock ->sk_lock-AF_CAN ->slock-AF_CAN ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&bridge_netdev_addr_lock_key ->team->team_lock_key#101 ->wq_pool_mutex.wait_lock ->team->team_lock_key#115 ->team->team_lock_key#116 ->team->team_lock_key#117 ->team->team_lock_key#118 FD: 42 BD: 196 +.+.: lock ->kernfs_idr_lock ->cgroup_idr_lock ->pidmap_lock ->drm_minor_lock ->&file_private->table_lock ->&q->queue_lock ->&group->inotify_data.idr_lock ->map_idr_lock ->prog_idr_lock FD: 13 BD: 207 +.+.: kernfs_idr_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 FD: 82 BD: 198 ++++: &root->kernfs_rwsem ->&root->kernfs_iattr_rwsem ->kernfs_idr_lock ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount ->&rq->__lock ->&cfs_rq->removed.lock ->&sem->wait_lock ->&p->pi_lock ->rcu_node_0 ->&meta->lock ->kfence_freelist_lock ->inode_hash_lock ->fs_reclaim ->&c->lock ->mmu_notifier_invalidate_range_start ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#24 ->remove_cache_srcu ->kernfs_rename_lock ->&n->list_lock ->&sb->s_type->i_lock_key#30 ->&sb->s_type->i_lock_key#31 ->&xa->xa_lock#15 ->&____s->seqcount#2 ->stock_lock ->quarantine_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq ->&base->lock ->batched_entropy_u8.lock FD: 1 BD: 4 ++++: file_systems_lock FD: 51 BD: 202 ++++: &root->kernfs_iattr_rwsem ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->iattr_mutex ->&sem->wait_lock ->tk_core.seq.seqcount ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&____s->seqcount ->quarantine_lock FD: 1 BD: 1 +.+.: dq_list_lock FD: 12 BD: 56 +.+.: sb_lock ->unnamed_dev_ida.xa_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 86 BD: 2 +.+.: &type->s_umount_key/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->percpu_counters_lock ->crngs.lock ->&sbinfo->stat_lock ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&dentry->d_lock ->fs_reclaim ->mmu_notifier_invalidate_range_start ->&rq->__lock ->&xa->xa_lock#15 ->stock_lock ->&n->list_lock FD: 25 BD: 44 +.+.: list_lrus_mutex ->&rq->__lock FD: 1 BD: 57 ....: unnamed_dev_ida.xa_lock FD: 1 BD: 22 +.+.: &sbinfo->stat_lock FD: 47 BD: 132 +.+.: &sb->s_type->i_lock_key ->&dentry->d_lock ->&xa->xa_lock#8 FD: 1 BD: 332 +.+.: &s->s_inode_list_lock FD: 37 BD: 419 +.+.: &dentry->d_lock ->&wq ->&dentry->d_lock/1 ->&obj_hash[i].lock ->pool_lock#2 ->&wq#2 ->&lru->node[i].lock ->sysctl_lock ->&dentry->d_lock/2 ->&p->pi_lock FD: 2 BD: 28 ....: mnt_id_ida.xa_lock ->pool_lock#2 FD: 41 BD: 125 +.+.: mount_lock ->mount_lock.seqcount ->&dentry->d_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 39 BD: 125 +.+.: mount_lock.seqcount ->&new_ns->poll ->&dentry->d_lock ->&obj_hash[i].lock ->pool_lock#2 ->&p->pi_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#2/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->&c->lock ->&____s->seqcount ->list_lrus_mutex ->sb_lock ->&sb->s_type->i_lock_key#2 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 38 BD: 117 +.+.: &sb->s_type->i_lock_key#2 ->&dentry->d_lock FD: 1 BD: 5 ..-.: ucounts_lock FD: 39 BD: 136 +.+.: init_fs.lock ->init_fs.seq.seqcount ->&dentry->d_lock FD: 1 BD: 130 +.+.: init_fs.seq.seqcount FD: 2 BD: 94 -.-.: jiffies_lock ->jiffies_seq.seqcount FD: 26 BD: 1 -.-.: log_wait.lock ->&p->pi_lock FD: 84 BD: 2 +.+.: &type->s_umount_key#3/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&____s->seqcount ->&c->lock ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock ->fs_reclaim ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->mmu_notifier_invalidate_range_start ->&rq->__lock ->&n->list_lock ->rcu_node_0 FD: 47 BD: 266 +.+.: &sb->s_type->i_lock_key#3 ->&dentry->d_lock ->&xa->xa_lock#8 FD: 1 BD: 116 ++++: proc_subdir_lock FD: 77 BD: 1 +.+.: &type->s_umount_key#4/1 ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&sb->s_type->i_lock_key#4 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 39 BD: 2 +.+.: &sb->s_type->i_lock_key#4 ->&dentry->d_lock ->&p->pi_lock ->bit_wait_table + i FD: 31 BD: 95 ....: cgroup_file_kn_lock ->kernfs_notify_lock FD: 33 BD: 94 ..-.: css_set_lock ->cgroup_file_kn_lock ->&p->pi_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 2 BD: 197 +...: cgroup_idr_lock ->pool_lock#2 FD: 53 BD: 93 +.+.: cpuset_mutex ->callback_lock ->jump_label_mutex ->&p->pi_lock ->&p->alloc_lock ->cpuset_attach_wq.lock ->&rq->__lock FD: 1 BD: 94 ....: callback_lock FD: 58 BD: 18 +.+.: blkcg_pol_mutex ->pcpu_alloc_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 18 +.+.: devcgroup_mutex FD: 35 BD: 93 +.+.: freezer_mutex ->freezer_lock ->rcu_node_0 ->&rq->__lock ->freezer_mutex.wait_lock ->&rcu_state.expedited_wq ->stock_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock#2 ->&cfs_rq->removed.lock FD: 44 BD: 55 +.+.: rcu_state.exp_mutex ->rcu_node_0 ->rcu_state.exp_wake_mutex ->&obj_hash[i].lock ->&rnp->exp_wq[2] ->&rq->__lock ->&rnp->exp_wq[0] ->&rnp->exp_wq[1] ->&rnp->exp_wq[3] ->pool_lock#2 ->&cfs_rq->removed.lock ->rcu_state.exp_mutex.wait_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq FD: 38 BD: 97 +.+.: rcu_state.exp_wake_mutex ->rcu_node_0 ->&rnp->exp_lock ->&rnp->exp_wq[0] ->&rnp->exp_wq[1] ->&rnp->exp_wq[2] ->&rnp->exp_wq[3] ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_state.exp_wake_mutex.wait_lock ->&rcu_state.expedited_wq FD: 1 BD: 98 +.+.: &rnp->exp_lock FD: 26 BD: 100 ....: &rnp->exp_wq[0] ->&p->pi_lock FD: 26 BD: 100 ....: &rnp->exp_wq[1] ->&p->pi_lock FD: 1 BD: 98 ....: init_sighand.siglock FD: 1 BD: 3 +.+.: init_files.file_lock FD: 13 BD: 208 ....: pidmap_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 120 BD: 92 ++++: cgroup_threadgroup_rwsem ->css_set_lock ->&p->pi_lock ->tk_core.seq.seqcount ->tasklist_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->pool_lock ->rcu_node_0 ->key ->pcpu_lock ->percpu_counters_lock ->&sighand->siglock ->cgroup_threadgroup_rwsem.rss.gp_wait.lock ->&x->wait#2 ->inode_hash_lock ->fs_reclaim ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#30 ->&root->kernfs_iattr_rwsem ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&c->lock ->cpuset_mutex ->freezer_mutex ->&p->alloc_lock ->&____s->seqcount#2 ->&____s->seqcount ->cgroup_threadgroup_rwsem.waiters.lock ->freezer_mutex.wait_lock ->&n->list_lock ->&rcu_state.expedited_wq ->stock_lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 25 BD: 1007 -.-.: &p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock FD: 62 BD: 97 .+.+: tasklist_lock ->init_sighand.siglock ->&sighand->siglock ->&pid->wait_pidfd ->&obj_hash[i].lock ->quarantine_lock ->&base->lock ->&p->alloc_lock ->stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 3107 -.-.: &per_cpu_ptr(group->pcpu, cpu)->seq FD: 1 BD: 1 ....: (kthreadd_done).wait.lock FD: 39 BD: 103 ....: &sighand->siglock ->&sig->wait_chldexit ->input_pool.lock ->&(&sig->stats_lock)->lock ->&p->pi_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->hrtimer_bases.lock ->&obj_hash[i].lock ->&sighand->signalfd_wqh ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&tty->ctrl.lock ->&rq->__lock ->stock_lock ->&n->list_lock ->&base->lock ->&____s->seqcount#2 ->quarantine_lock FD: 46 BD: 134 +.+.: &p->alloc_lock ->&____s->seqcount#2 ->init_fs.lock ->&fs->lock ->&x->wait ->cpu_asid_lock ->&x->wait#25 ->&newf->file_lock ->&p->pi_lock FD: 1 BD: 3154 .-.-: &____s->seqcount#2 FD: 49 BD: 588 +.+.: fs_reclaim ->mmu_notifier_invalidate_range_start ->&mapping->i_mmap_rwsem ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->icc_bw_lock ->pool_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->key ->pcpu_lock ->percpu_counters_lock ->stock_lock ->&____s->seqcount FD: 33 BD: 610 +.+.: mmu_notifier_invalidate_range_start ->dma_fence_map ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->stock_lock FD: 1 BD: 108 +.+.: kthread_create_lock FD: 26 BD: 153 ....: &x->wait ->&p->pi_lock FD: 31 BD: 1 +.+.: sched_map-wait-type-override ->&pool->lock ->&rq->__lock ->&cfs_rq->removed.lock ->rcu_node_0 ->&obj_hash[i].lock ->pool_lock#2 FD: 26 BD: 565 ....: (worker)->lock ->&p->pi_lock FD: 32 BD: 96 +.+.: wq_pool_attach_mutex ->&p->pi_lock ->&x->wait#7 ->&pool->lock ->&rq->__lock ->wq_pool_attach_mutex.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 52 ....: wq_mayday_lock FD: 2 BD: 95 ....: &xa->xa_lock ->pool_lock#2 FD: 1 BD: 1 ....: (&pool->mayday_timer) FD: 52 BD: 1 +.+.: (wq_completion)rcu_gp ->(work_completion)(&rnp->exp_poll_wq) ->(work_completion)(&(&ssp->srcu_sup->work)->work) ->(work_completion)(&sdp->work) ->(work_completion)(&rew->rew_work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 31 BD: 2 +.+.: (work_completion)(&rnp->exp_poll_wq) ->&rnp->exp_poll_lock FD: 4 BD: 1 -.-.: (null) ->tk_core.seq.seqcount FD: 12 BD: 1 +.-.: (&wq_watchdog_timer) ->&obj_hash[i].lock ->&base->lock FD: 507 BD: 1 +.+.: (wq_completion)events_unbound ->(work_completion)(&(&kfence_timer)->work) ->(work_completion)(&entry->work) ->(next_reseed).work ->(stats_flush_dwork).work ->(work_completion)(&sub_info->work) ->deferred_probe_work ->connector_reaper_work ->(reaper_work).work ->(work_completion)(&barr->work) ->(work_completion)(&rdev->wiphy_work) ->(work_completion)(&port->bc_work) ->(work_completion)(&map->work) ->&rq->__lock ->(work_completion)(&pool->idle_cull_work) FD: 229 BD: 2 +.+.: (work_completion)(&(&kfence_timer)->work) ->cpu_hotplug_lock ->allocation_wait.lock ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->&cfs_rq->removed.lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 26 BD: 3 -.-.: allocation_wait.lock ->&p->pi_lock FD: 3 BD: 3212 ..-.: batched_entropy_u8.lock ->crngs.lock FD: 1 BD: 3216 ..-.: kfence_freelist_lock FD: 1 BD: 645 ..-.: &meta->lock FD: 48 BD: 1 +.+.: rcu_tasks.tasks_gp_mutex ->&obj_hash[i].lock ->&base->lock ->tasks_rcu_exit_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(sdp, lock) ->tasks_rcu_exit_srcu ->&rq->__lock ->&x->wait#3 ->kernel/rcu/tasks.h:152 ->rcu_tasks__percpu.cbs_pcpu_lock ->&x->wait#2 ->(&timer.timer) ->(console_sem).lock FD: 1 BD: 1 ....: rcu_tasks.cbs_gbl_lock FD: 12 BD: 3 ..-.: rcu_tasks__percpu.cbs_pcpu_lock ->&obj_hash[i].lock ->&base->lock FD: 26 BD: 136 ..-.: &x->wait#2 ->&p->pi_lock FD: 228 BD: 1 +.+.: rcu_tasks_trace.tasks_gp_mutex ->cpu_hotplug_lock ->rcu_tasks_trace__percpu.cbs_pcpu_lock ->&x->wait#2 ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->(&timer.timer) ->(console_sem).lock FD: 1 BD: 1 ....: rcu_tasks_trace.cbs_gbl_lock FD: 26 BD: 100 ....: &rnp->exp_wq[2] ->&p->pi_lock FD: 30 BD: 6 ....: tasks_rcu_exit_srcu_srcu_usage.lock ->&obj_hash[i].lock FD: 1 BD: 20 ....: &ACCESS_PRIVATE(sdp, lock) FD: 1 BD: 2 ....: tasks_rcu_exit_srcu FD: 37 BD: 2 +.+.: (work_completion)(&(&ssp->srcu_sup->work)->work) ->&ssp->srcu_sup->srcu_gp_mutex ->tasks_rcu_exit_srcu_srcu_usage.lock ->&ssp->srcu_sup->srcu_cb_mutex ->remove_cache_srcu_srcu_usage.lock ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->&cfs_rq->removed.lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 35 BD: 3 +.+.: &ssp->srcu_sup->srcu_gp_mutex ->tasks_rcu_exit_srcu_srcu_usage.lock ->&rq->__lock ->&ssp->srcu_sup->srcu_cb_mutex ->remove_cache_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->&cfs_rq->removed.lock ->rcu_node_0 ->&obj_hash[i].lock ->pool_lock#2 FD: 26 BD: 20 ....: &x->wait#3 ->&p->pi_lock FD: 33 BD: 4 +.+.: &ssp->srcu_sup->srcu_cb_mutex ->tasks_rcu_exit_srcu_srcu_usage.lock ->remove_cache_srcu_srcu_usage.lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->&cfs_rq->removed.lock ->pool_lock#2 FD: 30 BD: 2 +.+.: (work_completion)(&sdp->work) ->&ACCESS_PRIVATE(sdp, lock) ->&obj_hash[i].lock ->&x->wait#3 ->&rq->__lock ->&cfs_rq->removed.lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 2 ....: kernel/rcu/tasks.h:152 FD: 26 BD: 100 ....: &rnp->exp_wq[3] ->&p->pi_lock FD: 12 BD: 92 ..-.: rcu_tasks_trace__percpu.cbs_pcpu_lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 91 ....: &ACCESS_PRIVATE(rtpcp, lock) FD: 1 BD: 1 +.+.: (memory_chain).rwsem FD: 52 BD: 93 +.+.: smpboot_threads_lock ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock ->hrtimer_bases.lock FD: 26 BD: 494 ..-.: &rcu_state.gp_wq ->&p->pi_lock FD: 25 BD: 93 -.-.: &stop_pi_lock ->&rq->__lock FD: 1 BD: 93 -.-.: &stopper->lock FD: 1 BD: 2 +.+.: (module_notify_list).rwsem FD: 1 BD: 1 +.+.: ddebug_lock FD: 2 BD: 1 +.+.: cci_probing ->devtree_lock FD: 1 BD: 1 +.+.: ptlock_ptr(ptdesc) FD: 228 BD: 1 +.+.: watchdog_mutex ->cpu_hotplug_lock FD: 885 BD: 1 +.+.: (wq_completion)events ->(work_completion)(&sscs.work) ->rdist_memreserve_cpuhp_cleanup_work ->(shepherd).work ->(work_completion)(&(&group->avgs_work)->work) ->(work_completion)(&rfkill_global_led_trigger_work) ->timer_update_work ->pcpu_balance_work ->(work_completion)(&p->wq) ->(debug_obj_work).work ->(work_completion)(&helper->damage_work) ->(work_completion)(&rfkill->sync_work) ->(linkwatch_work).work ->(work_completion)(&w->work) ->(work_completion)(&gadget->work) ->kernfs_notify_work ->async_lookup_work ->autoload_work ->(work_completion)(&barr->work) ->drain_vmap_work ->netstamp_work ->reg_work ->(work_completion)(&fw_work->work) ->(work_completion)(&s->destroy_work) ->(work_completion)(&(&krcp->monitor_work)->work) ->(work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) ->(work_completion)(&(&ovs_net->masks_rebalance)->work) ->(work_completion)(&ht->run_work) ->(work_completion)(&aux->work) ->(work_completion)(&w->work)#2 ->(deferred_probe_timeout_work).work ->(work_completion)(&cgrp->bpf.release_work) ->(regulator_init_complete_work).work ->(work_completion)(&w->w) ->(work_completion)(&sbi->s_sb_upd_work) ->deferred_process_work ->(work_completion)(&data->fib_event_work) ->(work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) ->(work_completion)(&(&hwstats->traffic_dw)->work) ->wireless_nlevent_work ->(work_completion)(&rdev->wiphy_work) ->(work_completion)(&(&conn->info_timer)->work) ->free_ipc_work ->(work_completion)(&(&devlink->rwork)->work) ->fqdir_free_work ->&rq->__lock ->(work_completion)(&tty->hangup_work)#2 ->(work_completion)(&aux->work)#2 ->(ima_keys_delayed_work).work ->key_gc_work ->(work_completion)(&blkg->free_work) ->&cfs_rq->removed.lock FD: 16 BD: 2 +.+.: (work_completion)(&sscs.work) ->&x->wait#4 ->&obj_hash[i].lock ->hrtimer_bases.lock ->&x->wait#5 FD: 1 BD: 94 -.-.: &x->wait#4 FD: 1 BD: 92 ....: &x->wait#5 FD: 2 BD: 135 +.+.: &newf->file_lock ->&newf->resize_wait FD: 1 BD: 1 ....: &p->vtime.seqcount FD: 40 BD: 91 +.+.: mem_hotplug_lock ->mem_hotplug_lock.rss.gp_wait.lock FD: 3 BD: 92 ..-.: mem_hotplug_lock.rss.gp_wait.lock ->&obj_hash[i].lock FD: 1 BD: 91 ....: mem_hotplug_lock.waiters.lock FD: 229 BD: 3 +.+.: cpu_add_remove_lock ->cpu_hotplug_lock ->cpu_hotplug_lock.waiters.lock ->cpu_hotplug_lock.rss.gp_wait.lock ->cpuset_hotplug_work FD: 3 BD: 91 ..-.: cpu_hotplug_lock.rss.gp_wait.lock ->&obj_hash[i].lock FD: 26 BD: 91 ....: cpu_hotplug_lock.waiters.lock ->&p->pi_lock FD: 1 BD: 4 +.+.: cpuset_hotplug_work FD: 1 BD: 92 +.+.: pcp_batch_high_lock FD: 26 BD: 98 ....: (cpu_running).wait.lock ->&p->pi_lock FD: 26 BD: 184 +.-.: (&timer.timer) ->&p->pi_lock FD: 1 BD: 1 ....: rcu_callback FD: 26 BD: 98 ....: &x->wait#6 ->&p->pi_lock FD: 1 BD: 3109 -.-.: &cfs_rq->removed.lock FD: 1 BD: 97 ....: &x->wait#7 FD: 15 BD: 3106 -...: &rt_b->rt_runtime_lock ->&rt_rq->rt_runtime_lock ->tk_core.seq.seqcount ->hrtimer_bases.lock FD: 1 BD: 3107 -...: &rt_rq->rt_runtime_lock FD: 29 BD: 91 +.+.: stop_cpus_mutex ->&stopper->lock ->&stop_pi_lock ->&rq->__lock ->&x->wait#8 FD: 26 BD: 93 ....: &x->wait#8 ->&p->pi_lock FD: 228 BD: 2 +.+.: rdist_memreserve_cpuhp_cleanup_work ->cpu_hotplug_lock FD: 59 BD: 1 +.+.: sched_domains_mutex ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->pcpu_alloc_mutex ->&c->lock ->&____s->seqcount ->pcpu_lock FD: 1 BD: 3106 ....: &cp->lock FD: 78 BD: 1 +.+.: &type->s_umount_key#5/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&obj_hash[i].lock ->percpu_counters_lock ->crngs.lock ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#5 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&dentry->d_lock FD: 38 BD: 119 +.+.: &sb->s_type->i_lock_key#5 ->&dentry->d_lock FD: 26 BD: 1 ....: (setup_done).wait.lock ->&p->pi_lock FD: 74 BD: 25 ++++: namespace_sem ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->mnt_id_ida.xa_lock ->pcpu_alloc_mutex ->&dentry->d_lock ->mount_lock ->rename_lock ->&obj_hash[i].lock ->&____s->seqcount#2 ->&n->list_lock ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->remove_cache_srcu FD: 1 BD: 129 +.+.: &____s->seqcount#3 FD: 62 BD: 2 ++++: &type->s_umount_key#6 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&____s->seqcount ->&c->lock ->&lru->node[i].lock ->&sbinfo->stat_lock ->&obj_hash[i].lock ->stock_lock ->&rq->__lock ->&____s->seqcount#2 FD: 26 BD: 421 +.+.: &lru->node[i].lock FD: 92 BD: 7 ++++: &sb->s_type->i_mutex_key ->namespace_sem ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->tomoyo_ss ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#2 ->&wb->list_lock ->&c->lock ->&____s->seqcount ->&rq->__lock FD: 39 BD: 24 +.+.: rename_lock ->rename_lock.seqcount FD: 38 BD: 158 +.+.: rename_lock.seqcount ->&dentry->d_lock ->&dentry->d_lock/2 FD: 1 BD: 126 ....: &new_ns->poll FD: 39 BD: 135 +.+.: &fs->lock ->&____s->seqcount#3 ->&dentry->d_lock FD: 1 BD: 89 +.+.: req_lock FD: 95 BD: 1 +.+.: of_mutex ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem FD: 1 BD: 161 ....: &x->wait#9 FD: 1 BD: 191 +.+.: &k->list_lock FD: 25 BD: 166 ++++: bus_type_sem ->&rq->__lock FD: 32 BD: 225 -...: &dev->power.lock ->&dev->power.wait_queue ->hrtimer_bases.lock ->&dev->power.lock/1 FD: 26 BD: 164 +.+.: dpm_list_mtx ->&rq->__lock ->rcu_node_0 FD: 61 BD: 171 +.+.: uevent_sock_mutex ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->&rq->__lock ->&cfs_rq->removed.lock ->rlock-AF_NETLINK ->&n->list_lock ->quarantine_lock ->remove_cache_srcu ->rcu_node_0 ->uevent_sock_mutex.wait_lock ->&rcu_state.expedited_wq ->&____s->seqcount#2 ->mmu_notifier_invalidate_range_start ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&base->lock ->pool_lock FD: 1 BD: 141 ....: running_helpers_waitq.lock FD: 1 BD: 178 +.+.: sysfs_symlink_target_lock FD: 2 BD: 233 +.+.: &k->k_lock ->klist_remove_lock FD: 1 BD: 1 ....: &dev->mutex FD: 1 BD: 1 +.+.: subsys mutex FD: 2 BD: 1 +.+.: memory_blocks.xa_lock ->pool_lock#2 FD: 1 BD: 1 +.+.: subsys mutex#2 FD: 1 BD: 1 +.+.: subsys mutex#3 FD: 57 BD: 98 +.+.: dev_pm_qos_mtx ->fs_reclaim ->pool_lock#2 ->&dev->power.lock ->pm_qos_lock ->&c->lock ->&____s->seqcount ->&rq->__lock FD: 1 BD: 99 ....: pm_qos_lock FD: 104 BD: 96 +.+.: dev_pm_qos_sysfs_mtx ->dev_pm_qos_mtx ->&root->kernfs_rwsem ->fs_reclaim ->pool_lock#2 ->lock ->&c->lock ->&____s->seqcount ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock ->dev_pm_qos_sysfs_mtx.wait_lock FD: 52 BD: 16 +.+.: register_lock ->proc_subdir_lock ->fs_reclaim ->pool_lock#2 ->proc_inum_ida.xa_lock ->&c->lock ->&____s->seqcount ->&rq->__lock FD: 1 BD: 1 +.+.: (cpufreq_policy_notifier_list).rwsem FD: 1 BD: 2 +.+.: (pm_chain_head).rwsem FD: 1 BD: 1 +.+.: cpufreq_governor_mutex FD: 40 BD: 2 +.+.: (work_completion)(&rew->rew_work) ->rcu_node_0 ->rcu_state.exp_wake_mutex ->&rcu_state.expedited_wq ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->(&timer.timer) ->pool_lock#2 ->pool_lock ->&cfs_rq->removed.lock ->rcu_state.exp_wake_mutex.wait_lock ->&p->pi_lock FD: 30 BD: 1 ..-.: &(&kfence_timer)->timer FD: 26 BD: 679 -.-.: &rcu_state.expedited_wq ->&p->pi_lock FD: 1 BD: 2 ++++: binfmt_lock FD: 1 BD: 74 +.+.: pin_fs_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#7/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->&____s->seqcount ->&c->lock ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#6 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 38 BD: 3 +.+.: &sb->s_type->i_lock_key#6 ->&dentry->d_lock FD: 64 BD: 1 +.+.: &sb->s_type->i_mutex_key#2 ->&sb->s_type->i_lock_key#6 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount FD: 26 BD: 422 ....: &wq ->&p->pi_lock FD: 77 BD: 1 +.+.: &type->s_umount_key#8/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#7 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 39 BD: 72 +.+.: &sb->s_type->i_lock_key#7 ->&dentry->d_lock ->bit_wait_table + i FD: 89 BD: 70 +.+.: &sb->s_type->i_mutex_key#3 ->&sb->s_type->i_lock_key#7 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->rcu_node_0 ->&rq->__lock ->(console_sem).lock ->&n->list_lock ->remove_cache_srcu ->pin_fs_lock ->mount_lock ->&fsnotify_mark_srcu ->&xa->xa_lock#8 ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rcu_state.gp_wq ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&xa->xa_lock#15 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&sem->wait_lock FD: 1 BD: 23 +.+.: chrdevs_lock FD: 755 BD: 2 ++++: cb_lock ->genl_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->rlock-AF_NETLINK ->rtnl_mutex ->&obj_hash[i].lock ->&____s->seqcount ->&n->list_lock ->&rdev->wiphy.mtx ->nlk_cb_mutex-GENERIC ->&xa->xa_lock#16 ->rcu_node_0 ->&rq->__lock ->genl_mutex.wait_lock ->&p->pi_lock ->rtnl_mutex.wait_lock ->&lock->wait_lock ->&____s->seqcount#2 ->remove_cache_srcu ->quarantine_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&dir->lock#2 ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 742 BD: 3 +.+.: genl_mutex ->fs_reclaim ->pool_lock#2 ->nl_table_lock ->nl_table_wait.lock ->&c->lock ->&____s->seqcount ->rlock-AF_NETLINK ->&obj_hash[i].lock ->&n->list_lock ->&zone->lock ->rtnl_mutex ->rcu_node_0 ->&rq->__lock ->rtnl_mutex.wait_lock ->&p->pi_lock ->hwsim_radio_lock ->&x->wait#9 ->batched_entropy_u32.lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->subsys mutex#53 ->device_links_lock ->&k->k_lock ->deferred_probe_mutex ->genl_mutex.wait_lock ->pcpu_alloc_mutex ->cpu_hotplug_lock ->wq_pool_mutex ->crngs.lock ->triggers_list_lock ->leds_list_lock ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->(inetaddr_chain).rwsem ->inet6addr_chain.lock ->&rcu_state.expedited_wq ->&____s->seqcount#2 ->remove_cache_srcu ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&sem->wait_lock ->&cfs_rq->removed.lock ->&base->lock ->&pernet->lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->nbd_index_mutex ->pcpu_alloc_mutex.wait_lock ->triggers_list_lock.wait_lock ->&pnettable->lock ->&x->wait#2 ->uevent_sock_mutex.wait_lock ->&nbd->config_lock ->&lock->wait_lock FD: 1 BD: 4 +.+.: subsys mutex#4 FD: 4 BD: 5 ....: async_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 3 +.+.: regulator_list_mutex FD: 357 BD: 2 +.+.: (work_completion)(&entry->work) ->tk_core.seq.seqcount ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->async_lock ->async_done.lock ->&dentry->d_lock ->&sb->s_type->i_mutex_key ->sb_writers#2 ->&sb->s_type->i_lock_key#2 ->quarantine_lock FD: 1 BD: 17 .+.+: device_links_srcu FD: 3 BD: 16 +.+.: fwnode_link_lock ->&k->k_lock FD: 28 BD: 98 +.+.: device_links_lock ->&k->list_lock ->&k->k_lock ->&rq->__lock FD: 1 BD: 18 ....: &dev->devres_lock FD: 4 BD: 16 +.+.: pinctrl_list_mutex ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 16 +.+.: pinctrl_maps_mutex FD: 1 BD: 4 +.+.: regulator_nesting_mutex FD: 2 BD: 1 +.+.: regulator_ww_class_mutex ->regulator_nesting_mutex FD: 99 BD: 94 +.+.: gdp_mutex ->&k->list_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->kobj_ns_type_lock ->&rq->__lock ->&obj_hash[i].lock ->&____s->seqcount#2 ->sysfs_symlink_target_lock ->kernfs_idr_lock ->&sem->wait_lock ->&p->pi_lock ->gdp_mutex.wait_lock ->&n->list_lock ->&cfs_rq->removed.lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 1 +.+.: (reboot_notifier_list).rwsem FD: 3 BD: 3 +.+.: subsys mutex#5 ->&k->k_lock FD: 25 BD: 99 +.+.: deferred_probe_mutex ->&rq->__lock FD: 1 BD: 16 ....: probe_waitqueue.lock FD: 26 BD: 3 ....: async_done.lock ->&p->pi_lock FD: 77 BD: 1 +.+.: &type->s_umount_key#9/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 39 BD: 55 +.+.: &sb->s_type->i_lock_key#8 ->&dentry->d_lock ->bit_wait_table + i ->&p->pi_lock FD: 35 BD: 7 +.+.: vmap_purge_lock ->purge_vmap_area_lock ->free_vmap_area_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->vmap_purge_lock.wait_lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 4 +.+.: &fp->aux->used_maps_mutex FD: 1 BD: 1 +.+.: proto_list_mutex FD: 1 BD: 1 +.+.: targets_mutex FD: 27 BD: 297 ...-: nl_table_lock ->pool_lock#2 ->nl_table_wait.lock ->&obj_hash[i].lock ->&c->lock FD: 26 BD: 298 ..-.: nl_table_wait.lock ->&p->pi_lock FD: 1 BD: 1 +.+.: net_family_lock FD: 8 BD: 5 ....: net_generic_ids.xa_lock ->&c->lock ->&____s->seqcount ->pool_lock#2 FD: 6 BD: 61 ..-.: &dir->lock ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock FD: 34 BD: 5 +.+.: k-sk_lock-AF_NETLINK ->k-slock-AF_NETLINK ->&rq->__lock FD: 1 BD: 6 +...: k-slock-AF_NETLINK FD: 2 BD: 103 ..-.: rhashtable_bucket ->rhashtable_bucket/1 FD: 1 BD: 51 ....: &cma->lock FD: 43 BD: 51 +.+.: cma_mutex ->&zone->lock ->&obj_hash[i].lock ->lock#2 FD: 28 BD: 1 +.+.: pcpu_drain_mutex ->&pcp->lock FD: 38 BD: 53 +.+.: lock#2 ->&obj_hash[i].lock ->(work_completion)(work) ->&x->wait#10 ->&rq->__lock FD: 1 BD: 1 +.+.: &pool->lock#2 FD: 1 BD: 94 ....: freezer_lock FD: 1 BD: 1 ....: &list->lock FD: 1 BD: 1 ....: kauditd_wait.lock FD: 1 BD: 1 ....: printk_ratelimit_state.lock FD: 1 BD: 1 ....: audit_backlog_wait.lock FD: 3 BD: 2 +.+.: lock#3 ->&zone->lock FD: 54 BD: 1 +.+.: khugepaged_mutex ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock ->lock#3 ->pcp_batch_high_lock FD: 128 BD: 7 ++++: &(&priv->bus_notifier)->rwsem ->&device->physical_node_lock ->iommu_probe_device_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->i2c_dev_list_lock ->&x->wait#9 ->&obj_hash[i].lock ->chrdevs_lock ->&k->list_lock ->gdp_mutex ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#61 FD: 3 BD: 7 +.+.: subsys mutex#6 ->&k->k_lock FD: 4 BD: 1 +.+.: subsys mutex#7 ->&k->list_lock ->&k->k_lock FD: 1 BD: 1 +.+.: regmap_debugfs_early_lock FD: 1 BD: 1 +.+.: (acpi_reconfig_chain).rwsem FD: 1 BD: 1 +.+.: __i2c_board_lock FD: 1 BD: 632 -.-.: quarantine_lock FD: 53 BD: 1 +.+.: core_lock ->&k->list_lock ->&k->k_lock ->fs_reclaim ->pool_lock#2 FD: 36 BD: 472 .+.+: remove_cache_srcu ->quarantine_lock ->&c->lock ->&n->list_lock ->&obj_hash[i].lock ->&rq->__lock ->pool_lock#2 ->&____s->seqcount ->pool_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&meta->lock ->kfence_freelist_lock ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&base->lock FD: 2 BD: 1 +.+.: thermal_governor_lock ->thermal_list_lock FD: 1 BD: 2 +.+.: thermal_list_lock FD: 33 BD: 1 +.+.: cpuidle_lock ->&obj_hash[i].lock ->(console_sem).lock FD: 53 BD: 1 +.+.: k-sk_lock-AF_QIPCRTR ->k-slock-AF_QIPCRTR ->fs_reclaim ->qrtr_ports.xa_lock ->pool_lock#2 ->qrtr_node_lock ->&obj_hash[i].lock FD: 1 BD: 2 +...: k-slock-AF_QIPCRTR FD: 1 BD: 2 +.+.: qrtr_ports.xa_lock FD: 1 BD: 2 +.+.: qrtr_node_lock FD: 51 BD: 98 ++++: (crypto_chain).rwsem ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->quarantine_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->remove_cache_srcu FD: 302 BD: 2 +.+.: tty_mutex ->(console_sem).lock ->console_lock ->fs_reclaim ->pool_lock#2 ->tty_ldiscs_lock ->&obj_hash[i].lock ->&k->list_lock ->&k->k_lock ->&tty->legacy_mutex ->stock_lock ->&rq->__lock ->&c->lock ->&n->list_lock ->tty_mutex.wait_lock ->(work_completion)(&buf->work) ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->remove_cache_srcu ->quarantine_lock ->&____s->seqcount#2 ->&____s->seqcount ->rcu_node_0 ->&rcu_state.expedited_wq ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 229 BD: 1 +.+.: iova_cache_mutex ->cpu_hotplug_lock ->slab_mutex FD: 3 BD: 1 +.+.: subsys mutex#8 ->&k->k_lock FD: 1 BD: 1 ..-.: uidhash_lock FD: 7 BD: 55 +.+.: (work_completion)(work) ->lock#4 ->lock#5 FD: 53 BD: 1 +.+.: (work_completion)(&eval_map_work) ->trace_event_sem FD: 1 BD: 1 ....: oom_reaper_wait.lock FD: 1 BD: 1 +.+.: subsys mutex#9 FD: 1 BD: 1 ....: &pgdat->kcompactd_wait FD: 1 BD: 1 ....: hugetlb_lock FD: 118 BD: 1 +.+.: memory_tier_lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->&c->lock ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#10 FD: 1 BD: 2 +.+.: subsys mutex#10 FD: 1 BD: 1 +.+.: ksm_thread_mutex FD: 1 BD: 1 ....: ksm_thread_wait.lock FD: 1 BD: 2 +.+.: damon_ops_lock FD: 52 BD: 97 ++++: crypto_alg_sem ->(crypto_chain).rwsem ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: khugepaged_mm_lock FD: 1 BD: 1 ....: khugepaged_wait.lock FD: 107 BD: 8 +.+.: bio_slab_lock ->fs_reclaim ->pool_lock#2 ->slab_mutex ->bio_slabs.xa_lock ->&rq->__lock FD: 8 BD: 9 +.+.: bio_slabs.xa_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount FD: 52 BD: 1 +.+.: major_names_lock ->fs_reclaim ->pool_lock#2 ->major_names_spinlock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&rq->__lock ->&obj_hash[i].lock ->loop_ctl_mutex FD: 1 BD: 2 +.+.: major_names_spinlock FD: 14 BD: 1 +.-.: (&rtpcp->lazy_timer) ->rcu_tasks_trace__percpu.cbs_pcpu_lock ->rcu_tasks__percpu.cbs_pcpu_lock FD: 52 BD: 1 +.+.: &pgdat->kswapd_lock ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock FD: 30 BD: 1 ..-.: drivers/char/random.c:251 FD: 13 BD: 2 +.+.: (next_reseed).work ->&obj_hash[i].lock ->&base->lock ->input_pool.lock ->base_crng.lock FD: 30 BD: 1 ..-.: mm/vmstat.c:2022 FD: 228 BD: 2 +.+.: (shepherd).work ->cpu_hotplug_lock ->&obj_hash[i].lock ->&base->lock FD: 35 BD: 54 +.+.: (wq_completion)mm_percpu_wq ->(work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) ->(work_completion)(work) ->(work_completion)(&barr->work) FD: 28 BD: 55 +.+.: (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->&pcp->lock FD: 3 BD: 7 +.+.: subsys mutex#11 ->&k->k_lock FD: 1 BD: 1 -...: shrink_qlist.lock FD: 30 BD: 5 ....: remove_cache_srcu_srcu_usage.lock ->&obj_hash[i].lock FD: 33 BD: 91 +.+.: flush_lock ->&obj_hash[i].lock ->(work_completion)(&sfw->work) ->&x->wait#10 ->&rq->__lock FD: 9 BD: 93 +.+.: (work_completion)(&sfw->work) ->&c->lock ->&n->list_lock ->&obj_hash[i].lock FD: 29 BD: 92 +.+.: (wq_completion)slub_flushwq ->(work_completion)(&sfw->work) ->(work_completion)(&barr->work) FD: 26 BD: 570 ....: &x->wait#10 ->&p->pi_lock FD: 27 BD: 100 +.+.: (work_completion)(&barr->work) ->&x->wait#10 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 30 BD: 1 ..-.: &(&ssp->srcu_sup->work)->timer FD: 1 BD: 1 +.+.: prepare_lock FD: 90 BD: 1 +.+.: clk_debug_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 FD: 1 BD: 1 +.+.: clocks_mutex FD: 355 BD: 1 +.+.: acpi_scan_lock ->semaphore->lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&x->wait#9 ->acpi_device_lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->subsys mutex#12 ->uevent_sock_mutex ->running_helpers_waitq.lock ->*(&acpi_gbl_reference_count_lock) ->&n->list_lock ->quarantine_lock ->acpi_ioremap_lock ->&device->physical_node_lock ->irq_domain_mutex ->&domain->mutex ->resource_lock ->&(&priv->bus_notifier)->rwsem ->fwnode_link_lock ->device_links_srcu ->acpi_pm_notifier_install_lock ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->subsys mutex#4 ->(console_sem).lock ->io_range_mutex ->pci_bus_sem ->gdp_mutex ->subsys mutex#17 ->acpi_hp_context_lock ->bridge_mutex ->pci_lock ->pci_acpi_companion_lookup_sem ->pci_slot_mutex ->resource_alignment_lock ->iort_msi_chip_lock ->subsys mutex#18 ->devtree_lock ->pci_rescan_remove_lock ->acpi_link_lock ->acpi_dep_list_lock ->power_resource_list_lock FD: 52 BD: 2 +.+.: acpi_device_lock ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#2 ->semaphore->lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount FD: 1 BD: 3 ....: &xa->xa_lock#2 FD: 1 BD: 2 +.+.: subsys mutex#12 FD: 96 BD: 34 +.+.: &device->physical_node_lock ->sysfs_symlink_target_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock FD: 122 BD: 2 +.+.: acpi_pm_notifier_install_lock ->semaphore->lock ->fs_reclaim ->pool_lock#2 ->*(&acpi_gbl_reference_count_lock) ->acpi_pm_notifier_lock ->&c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 119 BD: 3 +.+.: acpi_pm_notifier_lock ->fs_reclaim ->pool_lock#2 ->wakeup_ida.xa_lock ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#13 ->events_lock FD: 1 BD: 5 ....: wakeup_ida.xa_lock FD: 3 BD: 5 +.+.: subsys mutex#13 ->&k->k_lock FD: 1 BD: 5 ....: events_lock FD: 1 BD: 2 +.+.: acpi_wakeup_lock FD: 166 BD: 2 +.+.: port_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#14 ->&xa->xa_lock#3 ->&port->mutex FD: 1 BD: 3 +.+.: subsys mutex#14 FD: 1 BD: 3 ....: &xa->xa_lock#3 FD: 1 BD: 226 ....: &dev->power.wait_queue FD: 73 BD: 1 +.+.: (wq_completion)pm ->(work_completion)(&dev->power.work) FD: 72 BD: 2 +.+.: (work_completion)(&dev->power.work) ->&dev->power.lock ->&hub->irq_urb_lock ->(&hub->irq_urb_retry) ->&obj_hash[i].lock ->&base->lock ->hcd_urb_unlink_lock ->hcd_root_hub_lock ->usb_kill_urb_queue.lock ->&rq->__lock ->(work_completion)(&hub->tt.clear_work) ->&cfs_rq->removed.lock ->pool_lock#2 ->&dum_hcd->dum->lock ->device_state_lock ->hcd_urb_list_lock ->&vhci_hcd->vhci->lock ->mmu_notifier_invalidate_range_start ->fs_reclaim ->&x->wait#19 ->(&timer.timer) ->&c->lock ->&____s->seqcount ->&port_lock_key FD: 157 BD: 12 +.+.: &port->mutex ->fs_reclaim ->pool_lock#2 ->(console_sem).lock ->&port_lock_key ->console_mutex ->batched_entropy_u8.lock ->kfence_freelist_lock ->ctrl_ida.xa_lock ->&x->wait#9 ->&obj_hash[i].lock ->&dev->power.lock ->&____s->seqcount ->&k->list_lock ->&c->lock ->lock ->&root->kernfs_rwsem ->&device->physical_node_lock ->semaphore->lock ->sysfs_symlink_target_lock ->&k->k_lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#15 ->*(&acpi_gbl_reference_count_lock) ->&n->list_lock ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->gdp_mutex ->bus_type_sem ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->subsys mutex#16 ->chrdevs_lock ->&cfs_rq->removed.lock ->&desc->request_mutex ->register_lock ->&irq_desc_lock_class ->proc_subdir_lock ->proc_inum_ida.xa_lock ->&tpk_port.spinlock FD: 35 BD: 163 ....: &port_lock_key ->&dev->power.lock ->&port->lock ->&tty->write_wait FD: 25 BD: 14 +.+.: syslog_lock ->&rq->__lock FD: 37 BD: 159 ....: console_owner ->&port_lock_key ->console_owner_lock FD: 30 BD: 1 ..-.: &(&group->avgs_work)->timer FD: 30 BD: 1 ..-.: &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer FD: 30 BD: 1 ..-.: mm/memcontrol.c:679 FD: 26 BD: 2 +.+.: (work_completion)(&(&group->avgs_work)->work) ->&group->avgs_lock FD: 25 BD: 3 +.+.: &group->avgs_lock ->&per_cpu_ptr(group->pcpu, cpu)->seq ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 26 BD: 2 +.+.: (stats_flush_dwork).work ->cgroup_rstat_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 2 BD: 20 ....: cgroup_rstat_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 3123 -.-.: per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 13 ....: ctrl_ida.xa_lock FD: 1 BD: 13 +.+.: subsys mutex#15 FD: 1 BD: 234 +.+.: klist_remove_lock FD: 26 BD: 89 ....: &x->wait#11 ->&p->pi_lock FD: 133 BD: 1 .+.+: sb_writers ->mount_lock ->&type->i_mutex_dir_key/1 ->&sb->s_type->i_mutex_key#4 ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#5 ->&wb->list_lock ->&type->i_mutex_dir_key#2 ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock ->&s->s_inode_list_lock ->&obj_hash[i].lock ->&sbinfo->stat_lock ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&cfs_rq->removed.lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock#2 FD: 114 BD: 2 +.+.: &type->i_mutex_dir_key/1 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&obj_hash[i].lock ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#5 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&c->lock ->&____s->seqcount ->&xattrs->lock ->&simple_offset_xa_lock ->smack_known_lock ->&sb->s_type->i_mutex_key#4 ->&rq->__lock ->&cfs_rq->removed.lock ->rcu_node_0 ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->tomoyo_ss ->&u->bindlock ->&n->list_lock ->remove_cache_srcu ->&sem->wait_lock ->&fsnotify_mark_srcu ->&xa->xa_lock#8 ->&conn->lock ->&group->mark_mutex ->&mark->lock ->&group->notification_lock ->&group->notification_waitq ->&group->inotify_data.idr_lock ->destroy_lock ->&base->lock ->&____s->seqcount#2 FD: 1 BD: 11 ++++: &xattrs->lock FD: 13 BD: 12 +.+.: &simple_offset_xa_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 37 BD: 102 +.+.: smack_known_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->smack_known_lock.wait_lock ->&cfs_rq->removed.lock ->&base->lock ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->mmu_notifier_invalidate_range_start ->&c->lock FD: 30 BD: 3 +.+.: &sb->s_type->i_mutex_key#4 ->tk_core.seq.seqcount ->tomoyo_ss ->&xattrs->lock ->&rq->__lock ->&simple_offset_xa_lock ->&dentry->d_lock ->&sem->wait_lock FD: 3 BD: 13 +.+.: subsys mutex#16 ->&k->k_lock FD: 1 BD: 2 +.+.: io_range_mutex FD: 1 BD: 2 ++++: pci_bus_sem FD: 3 BD: 2 +.+.: subsys mutex#17 ->&k->k_lock FD: 50 BD: 2 +.+.: acpi_hp_context_lock ->fs_reclaim ->pool_lock#2 FD: 1 BD: 2 +.+.: bridge_mutex FD: 1 BD: 13 ....: pci_lock FD: 1 BD: 2 .+.+: pci_acpi_companion_lookup_sem FD: 1 BD: 2 +.+.: pci_slot_mutex FD: 1 BD: 2 +.+.: resource_alignment_lock FD: 1 BD: 226 ....: &dev->power.lock/1 FD: 1 BD: 2 +.+.: subsys mutex#18 FD: 36 BD: 2 +.+.: pci_rescan_remove_lock FD: 64 BD: 4 +.+.: acpi_link_lock ->fs_reclaim ->pool_lock#2 ->semaphore->lock ->&obj_hash[i].lock ->*(&acpi_gbl_reference_count_lock) ->(console_sem).lock ->&c->lock ->&____s->seqcount FD: 1 BD: 2 +.+.: acpi_dep_list_lock FD: 1 BD: 2 +.+.: power_resource_list_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#10/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#9 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 38 BD: 2 +.+.: &sb->s_type->i_lock_key#9 ->&dentry->d_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#11/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#10 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount ->&dentry->d_lock FD: 38 BD: 2 +.+.: &sb->s_type->i_lock_key#10 ->&dentry->d_lock FD: 158 BD: 83 ++++: &mm->mmap_lock ->reservation_ww_class_acquire ->fs_reclaim ->pool_lock#2 ->&vma->vm_lock->lock ->&____s->seqcount ->&mm->page_table_lock ->ptlock_ptr(ptdesc)#2 ->&c->lock ->&anon_vma->rwsem ->mmu_notifier_invalidate_range_start ->lock#4 ->lock#5 ->&obj_hash[i].lock ->&mapping->i_mmap_rwsem ->&p->alloc_lock ->&lruvec->lru_lock ->rcu_node_0 ->&rq->__lock ->tk_core.seq.seqcount ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&mm->mmap_lock/1 ->quarantine_lock ->&sem->wait_lock ->&p->pi_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&n->list_lock ->remove_cache_srcu ->&folio_wait_table[i] ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->&sb->s_type->i_lock_key ->&kcov->lock ->stock_lock ->&____s->seqcount#2 ->&rcu_state.expedited_wq ->&base->lock ->&xa->xa_lock#8 ->&info->lock ->key#22 FD: 75 BD: 94 +.+.: reservation_ww_class_acquire ->reservation_ww_class_mutex FD: 74 BD: 95 +.+.: reservation_ww_class_mutex ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&xa->xa_lock#8 ->&sb->s_type->i_lock_key ->&info->lock ->lock#4 ->&rq->__lock ->&c->lock ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock FD: 47 BD: 590 ++++: &mapping->i_mmap_rwsem ->&obj_hash[i].lock ->pool_lock#2 ->&anon_vma->rwsem ->&rq->__lock ->&____s->seqcount ->&sem->wait_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->stock_lock ->quarantine_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&base->lock ->&p->pi_lock FD: 1 BD: 611 +.+.: dma_fence_map FD: 1 BD: 569 ....: key FD: 1 BD: 1 +.+.: attribute_container_mutex FD: 26 BD: 14 +.+.: triggers_list_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->triggers_list_lock.wait_lock FD: 25 BD: 14 .+.+: leds_list_lock ->&rq->__lock FD: 138 BD: 2 ++++: (usb_notifier_list).rwsem ->fs_reclaim ->pool_lock#2 ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#57 ->mon_lock FD: 1 BD: 1 +.+.: rc_map_lock FD: 1 BD: 1 +.+.: subsys mutex#19 FD: 2 BD: 8 +.+.: iommu_probe_device_lock ->iommu_device_lock FD: 1 BD: 9 +.+.: iommu_device_lock FD: 1 BD: 7 ....: (efi_runtime_lock).lock FD: 26 BD: 9 ....: &x->wait#12 ->&p->pi_lock FD: 29 BD: 1 +.+.: (wq_completion)efi_rts_wq ->(work_completion)(&efi_rts_work.work) FD: 28 BD: 2 +.+.: (work_completion)(&efi_rts_work.work) ->cpu_asid_lock ->efi_rt_lock ->&x->wait#12 FD: 1 BD: 3106 ....: cpu_asid_lock FD: 1 BD: 3 +.+.: efi_rt_lock FD: 1 BD: 6 ....: (efivars_lock).lock FD: 1 BD: 1 +.+.: devfreq_list_lock FD: 1 BD: 2 +.+.: &entry->access FD: 52 BD: 2 +.+.: info_mutex ->proc_subdir_lock ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->proc_inum_ida.xa_lock ->&rq->__lock FD: 1 BD: 95 +.+.: kobj_ns_type_lock FD: 12 BD: 59 +.+.: &xa->xa_lock#4 ->pool_lock#2 ->&c->lock ->&n->list_lock ->&obj_hash[i].lock ->&____s->seqcount FD: 27 BD: 53 +.+.: subsys mutex#20 ->&k->k_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 14 BD: 173 ..-.: &dir->lock#2 ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&base->lock FD: 35 BD: 59 +.+.: dev_hotplug_mutex ->&dev->power.lock ->&k->k_lock ->&rq->__lock FD: 13 BD: 58 ++++: dev_base_lock ->&xa->xa_lock#4 FD: 1 BD: 1 +.+.: qdisc_mod_lock FD: 21 BD: 1 ++++: bt_proto_lock ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->chan_list_lock ->l2cap_sk_list.lock ->&sk->sk_peer_lock ->hci_sk_list.lock ->&n->list_lock ->sco_sk_list.lock ->&____s->seqcount#2 FD: 67 BD: 26 +.+.: hci_cb_list_lock ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->&obj_hash[i].lock ->chan_list_lock ->&conn->ident_lock ->&base->lock ->&list->lock#8 ->&conn->chan_lock ->&c->lock ->(work_completion)(&(&conn->id_addr_timer)->work) ->&x->wait#2 ->&rq->__lock ->(work_completion)(&(&conn->info_timer)->work) ->remove_cache_srcu ->hci_cb_list_lock.wait_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&cfs_rq->removed.lock ->&meta->lock ->&____s->seqcount#2 ->&conn->lock#2 ->(work_completion)(&(&conn->timeout_work)->work) ->pool_lock ->sk_lock-AF_BLUETOOTH-BTPROTO_SCO ->slock-AF_BLUETOOTH-BTPROTO_SCO FD: 1 BD: 1 +.+.: mgmt_chan_list_lock FD: 1 BD: 74 ....: &list->lock#2 FD: 50 BD: 51 +.+.: rate_ctrl_mutex ->fs_reclaim ->pool_lock#2 FD: 2 BD: 1 +.+.: netlbl_domhsh_lock ->pool_lock#2 FD: 1 BD: 49 +.+.: netlbl_unlhsh_lock FD: 374 BD: 1 +.+.: misc_mtx ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&____s->seqcount ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#21 ->misc_minors_ida.xa_lock ->&cfs_rq->removed.lock ->&n->list_lock ->&base->lock ->pool_lock ->&dir->lock ->rfkill_global_mutex ->nfc_index_ida.xa_lock ->pcpu_alloc_mutex ->cpu_hotplug_lock ->kthread_create_lock ->&x->wait ->wq_pool_mutex ->nfc_devlist_mutex ->&k->k_lock ->llcp_devices_lock ->nl_table_lock ->nl_table_wait.lock ->&____s->seqcount#2 ->misc_mtx.wait_lock ->nfc_devlist_mutex.wait_lock ->hrtimer_bases.lock ->quarantine_lock ->remove_cache_srcu ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->batched_entropy_u8.lock ->kfence_freelist_lock ->pcpu_alloc_mutex.wait_lock ->&meta->lock ->(console_sem).lock ->&wq->mutex ->wq_mayday_lock ->wq_pool_mutex.wait_lock ->&x->wait#2 ->(wq_completion)nfc4_nci_rx_wq#47 ->(wq_completion)nfc4_nci_cmd_wq#48 ->(wq_completion)nfc4_nci_cmd_wq#49 ->(wq_completion)nfc5_nci_cmd_wq#27 ->(wq_completion)nfc4_nci_tx_wq#74 ->(wq_completion)nfc4_nci_rx_wq#75 ->(wq_completion)nfc4_nci_cmd_wq#77 ->(wq_completion)nfc4_nci_rx_wq#86 ->(wq_completion)nfc4_nci_cmd_wq#88 ->(wq_completion)nfc35_nci_cmd_wq#8 ->(wq_completion)nfc3_nci_cmd_wq#262 ->(wq_completion)nfc48_nci_tx_wq#3 ->(wq_completion)nfc48_nci_rx_wq#3 ->(wq_completion)nfc48_nci_cmd_wq#3 ->(wq_completion)nfc3_nci_rx_wq#289 ->(wq_completion)nfc3_nci_cmd_wq#294 ->(wq_completion)nfc3_nci_cmd_wq#295 ->(wq_completion)nfc8_nci_cmd_wq#15 ->(wq_completion)nfc5_nci_cmd_wq#44 ->(wq_completion)nfc2_nci_cmd_wq#713 ->(wq_completion)nfc3_nci_cmd_wq#331 ->(wq_completion)nfc3_nci_cmd_wq#334 ->(wq_completion)nfc2_nci_rx_wq#745 ->(wq_completion)nfc2_nci_cmd_wq#749 ->(wq_completion)nfc2_nci_cmd_wq#762 ->(wq_completion)nfc43_nci_tx_wq#4 ->(wq_completion)nfc43_nci_rx_wq#4 ->(wq_completion)nfc43_nci_cmd_wq#4 ->(wq_completion)nfc2_nci_tx_wq#826 ->(wq_completion)nfc2_nci_rx_wq#829 ->(wq_completion)nfc2_nci_cmd_wq#834 ->(wq_completion)nfc36_nci_cmd_wq#7 ->(wq_completion)nfc2_nci_rx_wq#856 ->(wq_completion)nfc2_nci_cmd_wq#861 ->(wq_completion)nfc2_nci_cmd_wq#872 ->(wq_completion)nfc2_nci_rx_wq#867 ->(wq_completion)nfc2_nci_cmd_wq#873 ->(wq_completion)nfc2_nci_cmd_wq#874 ->(wq_completion)nfc2_nci_rx_wq#886 ->(wq_completion)nfc2_nci_cmd_wq#893 ->(wq_completion)nfc2_nci_tx_wq#885 ->(wq_completion)nfc2_nci_rx_wq#891 ->(wq_completion)nfc2_nci_cmd_wq#898 ->(wq_completion)nfc4_nci_rx_wq#160 ->(wq_completion)nfc4_nci_cmd_wq#162 ->(wq_completion)nfc3_nci_rx_wq#416 ->(wq_completion)nfc3_nci_cmd_wq#424 ->(wq_completion)nfc2_nci_tx_wq#902 ->(wq_completion)nfc2_nci_rx_wq#908 ->(wq_completion)nfc2_nci_cmd_wq#915 ->(wq_completion)nfc6_nci_cmd_wq#24 ->(wq_completion)nfc6_nci_cmd_wq#26 ->(wq_completion)nfc2_nci_tx_wq#925 ->(wq_completion)nfc2_nci_rx_wq#931 ->(wq_completion)nfc2_nci_cmd_wq#938 ->(wq_completion)nfc2_nci_cmd_wq#1003 ->(wq_completion)nfc3_nci_rx_wq#456 ->(wq_completion)nfc3_nci_cmd_wq#464 ->(wq_completion)nfc2_nci_cmd_wq#1016 ->(wq_completion)nfc2_nci_cmd_wq#1025 ->(wq_completion)nfc2_nci_rx_wq#1027 ->(wq_completion)nfc2_nci_cmd_wq#1037 ->(wq_completion)nfc6_nci_cmd_wq#30 ->(wq_completion)nfc8_nci_rx_wq#21 ->(wq_completion)nfc8_nci_cmd_wq#22 ->(wq_completion)nfc3_nci_rx_wq#502 ->(wq_completion)nfc3_nci_cmd_wq#510 ->(wq_completion)nfc2_nci_rx_wq#1071 ->(wq_completion)nfc2_nci_cmd_wq#1081 ->(wq_completion)nfc2_nci_rx_wq#1076 ->(wq_completion)nfc2_nci_cmd_wq#1086 ->(wq_completion)nfc2_nci_tx_wq#1085 ->(wq_completion)nfc2_nci_rx_wq#1094 ->(wq_completion)nfc2_nci_cmd_wq#1104 ->(wq_completion)nfc3_nci_cmd_wq#540 ->(wq_completion)nfc2_nci_rx_wq#1115 ->(wq_completion)nfc2_nci_cmd_wq#1125 ->(wq_completion)nfc5_nci_tx_wq#78 ->(wq_completion)nfc5_nci_rx_wq#79 ->(wq_completion)nfc5_nci_cmd_wq#81 ->(wq_completion)nfc2_nci_cmd_wq#1142 ->(wq_completion)nfc4_nci_cmd_wq#232 ->(wq_completion)nfc5_nci_cmd_wq#85 ->(wq_completion)nfc3_nci_cmd_wq#564 ->(wq_completion)nfc2_nci_rx_wq#1148 ->(wq_completion)nfc2_nci_cmd_wq#1159 ->(wq_completion)nfc3_nci_cmd_wq#572 ->(wq_completion)nfc3_nci_tx_wq#559 ->(wq_completion)nfc3_nci_rx_wq#564 ->(wq_completion)nfc3_nci_cmd_wq#575 ->(wq_completion)nfc2_nci_cmd_wq#1169 ->(wq_completion)nfc3_nci_cmd_wq#576 ->(wq_completion)nfc3_nci_cmd_wq#598 ->(wq_completion)nfc9_nci_rx_wq#23 ->(wq_completion)nfc9_nci_cmd_wq#23 ->(wq_completion)nfc2_nci_cmd_wq#1211 ->(wq_completion)nfc4_nci_cmd_wq#259 ->(wq_completion)nfc3_nci_cmd_wq#612 ->(wq_completion)nfc4_nci_rx_wq#256 ->(wq_completion)nfc4_nci_cmd_wq#260 ->(wq_completion)nfc4_nci_cmd_wq#265 ->(wq_completion)nfc4_nci_cmd_wq#266 FD: 3 BD: 2 +.+.: subsys mutex#21 ->&k->k_lock FD: 125 BD: 2 +.+.: (work_completion)(&rfkill_global_led_trigger_work) ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->&p->pi_lock ->rcu_node_0 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 124 BD: 13 +.+.: rfkill_global_mutex ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->&rfkill->lock ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#40 ->triggers_list_lock ->leds_list_lock ->&rq->__lock ->rfkill_global_mutex.wait_lock ->&n->list_lock ->&data->mtx ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->&sem->wait_lock ->&p->pi_lock ->remove_cache_srcu ->quarantine_lock ->&data->read_wait#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->uevent_sock_mutex.wait_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 125 BD: 1 +.+.: input_mutex ->input_devices_poll_wait.lock ->fs_reclaim ->pool_lock#2 ->&dev->mutex#2 ->input_ida.xa_lock ->&x->wait#9 ->&obj_hash[i].lock ->chrdevs_lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#31 FD: 1 BD: 2 ....: input_devices_poll_wait.lock FD: 46 BD: 2 ++++: (netlink_chain).rwsem ->hwsim_radio_lock ->pool_lock#2 ->&obj_hash[i].lock ->reg_indoor_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rq->__lock ->rcu_node_0 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&cfs_rq->removed.lock ->&n->list_lock ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 29 BD: 1 ++++: proto_tab_lock ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 3 BD: 1 ....: random_ready_notifier.lock ->crngs.lock FD: 1 BD: 2 ....: misc_minors_ida.xa_lock FD: 1 BD: 1 +.+.: wtd_deferred_reg_mutex FD: 76 BD: 1 +.+.: &type->s_umount_key#12/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#11 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 38 BD: 2 +.+.: &sb->s_type->i_lock_key#11 ->&dentry->d_lock FD: 229 BD: 2 +.+.: timer_update_work ->timer_keys_mutex FD: 228 BD: 3 +.+.: timer_keys_mutex ->cpu_hotplug_lock FD: 77 BD: 1 +.+.: &type->s_umount_key#13/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#12 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&____s->seqcount ->&dentry->d_lock ->bit_wait_table + i FD: 90 BD: 1 +.+.: (work_completion)(&tracerfs_init_work) ->pin_fs_lock ->fs_reclaim ->pool_lock#2 ->sb_lock ->bit_wait_table + i ->&rq->__lock ->&type->s_umount_key#14 ->mnt_id_ida.xa_lock ->pcpu_alloc_mutex ->&dentry->d_lock ->mount_lock ->&obj_hash[i].lock ->&fsnotify_mark_srcu ->&sb->s_type->i_mutex_key#5 ->event_mutex ->(module_notify_list).rwsem ->trace_types_lock FD: 38 BD: 6 +.+.: &sb->s_type->i_lock_key#12 ->&dentry->d_lock FD: 26 BD: 289 -.-.: bit_wait_table + i ->&p->pi_lock FD: 63 BD: 3 ++++: &type->s_umount_key#14 ->sb_lock ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&lru->node[i].lock ->&obj_hash[i].lock ->stock_lock ->&rq->__lock ->&c->lock FD: 64 BD: 4 +.+.: &sb->s_type->i_mutex_key#5 ->&sb->s_type->i_lock_key#12 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount FD: 53 BD: 136 .+.+: &fsnotify_mark_srcu ->&conn->lock ->fs_reclaim ->pool_lock#2 ->&group->notification_lock ->&group->notification_waitq ->&rq->__lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->remove_cache_srcu ->&obj_hash[i].lock ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 76 BD: 1 +.+.: &type->s_umount_key#15/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#13 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 38 BD: 2 +.+.: &sb->s_type->i_lock_key#13 ->&dentry->d_lock FD: 70 BD: 2 +.+.: event_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#5 ->trace_event_sem ->trace_types_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#16/1 ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#14 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 1 BD: 5 +.+.: eventfs_mutex FD: 38 BD: 2 +.+.: &sb->s_type->i_lock_key#14 ->&dentry->d_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#17/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->&c->lock ->&____s->seqcount ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#15 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 38 BD: 2 +.+.: &sb->s_type->i_lock_key#15 ->&dentry->d_lock FD: 51 BD: 1 +.+.: kclist_lock ->resource_lock ->fs_reclaim ->pool_lock#2 FD: 77 BD: 2 +.+.: &type->s_umount_key#18/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->&c->lock ->&____s->seqcount ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#16 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->&n->list_lock FD: 38 BD: 4 +.+.: &sb->s_type->i_lock_key#16 ->&dentry->d_lock FD: 174 BD: 32 .+.+: tomoyo_ss ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->tomoyo_policy_lock ->(console_sem).lock ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->tomoyo_log_lock ->tomoyo_log_wait.lock ->&rq->__lock ->file_systems_lock ->fs_reclaim ->&mm->mmap_lock ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->remove_cache_srcu ->&n->list_lock ->rename_lock ->&base->lock ->&____s->seqcount#2 ->stock_lock ->&rcu_state.expedited_wq ->mount_lock ->pool_lock FD: 50 BD: 1 +.+.: pnp_lock ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: subsys mutex#22 FD: 3 BD: 1 +.+.: subsys mutex#23 ->&k->k_lock FD: 3 BD: 7 +.+.: subsys mutex#24 ->&k->k_lock FD: 4 BD: 1 +.+.: subsys mutex#25 ->&k->list_lock ->&k->k_lock FD: 1 BD: 1 ....: netevent_notif_chain.lock FD: 51 BD: 1 +.+.: clients_rwsem ->fs_reclaim ->clients.xa_lock FD: 2 BD: 2 +.+.: clients.xa_lock ->pool_lock#2 FD: 29 BD: 8 .+.+: devices_rwsem ->rcu_node_0 ->&rq->__lock ->&rcu_state.expedited_wq FD: 1 BD: 1 +.+.: (blocking_lsm_notifier_chain).rwsem FD: 117 BD: 49 ++++: (inetaddr_chain).rwsem ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->&c->lock ->fib_info_lock ->&dir->lock#2 ->&____s->seqcount ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->&net->sctp.local_addr_lock ->rlock-AF_NETLINK ->&rq->__lock ->&n->list_lock ->rcu_node_0 ->&ipvlan->addrs_lock ->&____s->seqcount#2 ->&tbl->lock ->class ->(&tbl->proxy_timer) ->&base->lock ->krc.lock ->mmu_notifier_invalidate_range_start ->remove_cache_srcu ->&cfs_rq->removed.lock ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock ->quarantine_lock ->(inetaddr_chain).rwsem.wait_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&x->wait#2 FD: 1 BD: 8 ....: inet6addr_chain.lock FD: 1 BD: 1 +.+.: buses_mutex FD: 1 BD: 1 +.+.: offload_lock FD: 1 BD: 1 +...: inetsw_lock FD: 58 BD: 2 +.+.: pcpu_balance_work ->pcpu_alloc_mutex ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock FD: 1 BD: 51 +.+.: ptype_lock FD: 744 BD: 1 +.+.: (wq_completion)events_power_efficient ->(work_completion)(&(&tbl->managed_work)->work) ->(check_lifetime_work).work ->(work_completion)(&(&cache_cleaner)->work) ->(work_completion)(&(&ops->cursor_work)->work) ->(work_completion)(&(&hub->init_work)->work) ->(work_completion)(&(&gc_work->dwork)->work) ->(work_completion)(&(&tbl->gc_work)->work) ->(gc_work).work ->(crda_timeout).work FD: 45 BD: 2 +.+.: (work_completion)(&(&tbl->managed_work)->work) ->&tbl->lock ->&rq->__lock FD: 44 BD: 128 +.-.: &tbl->lock ->&obj_hash[i].lock ->&base->lock ->&n->lock ->pool_lock#2 ->nl_table_lock ->rlock-AF_NETLINK ->nl_table_wait.lock ->&dir->lock#2 ->krc.lock ->&c->lock ->batched_entropy_u32.lock ->&____s->seqcount ->quarantine_lock ->&____s->seqcount#2 ->tk_core.seq.seqcount ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->init_task.mems_allowed_seq.seqcount FD: 29 BD: 2 +.+.: (check_lifetime_work).work ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->rcu_node_0 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 49 +.+.: &net->rules_mod_lock FD: 1 BD: 1 +.+.: tcp_ulp_list_lock FD: 1 BD: 1 +...: xfrm_state_afinfo_lock FD: 1 BD: 1 +.+.: xfrm_policy_afinfo_lock FD: 1 BD: 1 +...: xfrm_input_afinfo_lock FD: 15 BD: 587 ..-.: krc.lock ->&obj_hash[i].lock ->hrtimer_bases.lock ->&base->lock FD: 1 BD: 50 +...: k-slock-AF_INET/1 FD: 56 BD: 1 +.+.: (wq_completion)events_highpri ->(work_completion)(&(&krcp->page_cache_work)->work) ->(work_completion)(flush) ->(work_completion)(&barr->work) FD: 51 BD: 2 +.+.: (work_completion)(&(&krcp->page_cache_work)->work) ->fs_reclaim ->&____s->seqcount ->krc.lock FD: 1 BD: 2 +.+.: &hashinfo->lock FD: 1 BD: 1 +.+.: tcp_cong_list_lock FD: 1 BD: 1 +.+.: mptcp_sched_list_lock FD: 2 BD: 7 +.+.: cache_list_lock ->&cd->hash_lock FD: 27 BD: 2 +.+.: (work_completion)(&(&cache_cleaner)->work) ->cache_list_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 79 BD: 9 ++++: (rpc_pipefs_notifier_list).rwsem ->&sn->rpc_client_lock ->rename_lock.seqcount ->&sb->s_type->i_mutex_key#20/1 FD: 1 BD: 1 +.+.: svc_xprt_class_lock FD: 39 BD: 1 +.+.: xprt_list_lock ->(console_sem).lock FD: 131 BD: 3 ++++: umhelper_sem ->usermodehelper_disabled_waitq.lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->&c->lock ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->&k->k_lock ->subsys mutex#73 ->fw_lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&x->wait#22 ->&base->lock ->&rq->__lock ->(&timer.timer) ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start FD: 1 BD: 4 ....: usermodehelper_disabled_waitq.lock FD: 27 BD: 420 +.+.: &dentry->d_lock/1 ->&lru->node[i].lock FD: 93 BD: 3 .+.+: sb_writers#2 ->mount_lock ->&sb->s_type->i_mutex_key/1 ->&sb->s_type->i_mutex_key FD: 89 BD: 4 +.+.: &sb->s_type->i_mutex_key/1 ->rename_lock.seqcount ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&dentry->d_lock ->&obj_hash[i].lock ->tomoyo_ss ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#2 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&sb->s_type->i_mutex_key FD: 1 BD: 2 +.+.: tomoyo_log_lock FD: 1 BD: 2 ....: tomoyo_log_wait.lock FD: 55 BD: 115 +.+.: &wb->list_lock ->&sb->s_type->i_lock_key#2 ->&sb->s_type->i_lock_key#23 ->&sb->s_type->i_lock_key#22 ->&sb->s_type->i_lock_key ->&sb->s_type->i_lock_key#5 ->&sb->s_type->i_lock_key#24 ->&sb->s_type->i_lock_key#3 FD: 140 BD: 2 +.+.: (work_completion)(&sub_info->work) ->&sighand->siglock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->batched_entropy_u64.lock ->&obj_hash[i].lock ->init_files.file_lock ->init_fs.lock ->&c->lock ->&p->alloc_lock ->lock ->pidmap_lock ->cgroup_threadgroup_rwsem ->input_pool.lock ->&p->pi_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rq->__lock ->&cfs_rq->removed.lock ->rcu_node_0 ->&n->list_lock ->&sig->wait_chldexit ->tasklist_lock ->&prev->lock ->&(&sig->stats_lock)->lock ->css_set_lock ->&x->wait#17 ->&____s->seqcount#2 ->remove_cache_srcu ->&rcu_state.expedited_wq ->&meta->lock FD: 1 BD: 1 +.+.: umh_sysctl_lock FD: 1 BD: 3 +.+.: &drv->dynids.lock FD: 159 BD: 2 +.+.: &tsk->futex_exit_mutex ->&p->pi_lock ->&mm->mmap_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 28 BD: 1 +.+.: &child->perf_event_mutex ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 98 ....: &pid->wait_pidfd FD: 26 BD: 104 ....: &sig->wait_chldexit ->&p->pi_lock FD: 15 BD: 104 ....: &(&sig->stats_lock)->lock ->&____s->seqcount#4 FD: 14 BD: 106 ....: &____s->seqcount#4 ->pidmap_lock ->pool_lock#2 ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: subsys mutex#26 FD: 143 BD: 1 +.+.: subsys mutex#27 ->&k->list_lock ->&k->k_lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->platform_devid_ida.xa_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->&(&priv->bus_notifier)->rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#4 ->&c->lock ->&____s->seqcount ->wakeup_ida.xa_lock ->gdp_mutex ->subsys mutex#13 ->events_lock ->rtcdev_lock FD: 1 BD: 1 +.+.: subsys mutex#28 FD: 1 BD: 94 +.+.: subsys mutex#29 FD: 1 BD: 8 +.+.: key_user_lock FD: 1 BD: 8 +.+.: key_serial_lock FD: 57 BD: 8 +.+.: key_construction_mutex ->&obj_hash[i].lock ->pool_lock#2 ->keyring_name_lock ->root_key_user.lock ->&rq->__lock ->fs_reclaim ->&c->lock ->&base->lock ->&type->lock_class#3/1 ->&n->list_lock ->key_construction_mutex.wait_lock FD: 61 BD: 6 +.+.: &type->lock_class ->keyring_serialise_link_lock ->fs_reclaim ->pool_lock#2 ->key_user_lock ->crngs.lock ->key_serial_lock ->key_construction_mutex ->ima_keys_lock ->root_key_user.lock ->&c->lock ->&rq->__lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->&obj_hash[i].lock ->&type->lock_class/1 ->&base->lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 58 BD: 7 +.+.: keyring_serialise_link_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->root_key_user.lock ->key_construction_mutex ->key_user_lock ->crngs.lock ->key_serial_lock ->&rq->__lock ->&____s->seqcount#2 ->&n->list_lock ->remove_cache_srcu FD: 1 BD: 1 ....: &pgdat->kswapd_wait FD: 1 BD: 1 +.+.: drivers_lock FD: 92 BD: 1 +.+.: damon_dbgfs_lock ->fs_reclaim ->pool_lock#2 ->damon_ops_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 FD: 76 BD: 1 +.+.: &type->s_umount_key#19/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&c->lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#17 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 38 BD: 2 +.+.: &sb->s_type->i_lock_key#17 ->&dentry->d_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#20/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->&c->lock ->&____s->seqcount ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 38 BD: 8 +.+.: &sb->s_type->i_lock_key#18 ->&dentry->d_lock FD: 1 BD: 1 +.+.: configfs_subsystem_mutex FD: 72 BD: 1 +.+.: &sb->s_type->i_mutex_key#6/1 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&____s->seqcount ->&c->lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]/2 ->&sb->s_type->i_mutex_key#7/2 ->&default_group_class[depth - 1]#3 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 9 +.+.: configfs_dirent_lock FD: 69 BD: 2 +.+.: &default_group_class[depth - 1]/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#4/2 FD: 50 BD: 1 +.+.: ecryptfs_daemon_hash_mux ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 FD: 1 BD: 1 ....: &ecryptfs_kthread_ctl.wait FD: 2 BD: 1 +.+.: ecryptfs_msg_ctx_lists_mux ->&ecryptfs_msg_ctx_arr[i].mux FD: 1 BD: 2 +.+.: &ecryptfs_msg_ctx_arr[i].mux FD: 1 BD: 1 +.+.: nfs_version_lock FD: 86 BD: 1 ++++: key_types_sem ->(console_sem).lock ->asymmetric_key_parsers_sem ->&type->lock_class ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->keyring_serialise_link_lock ->fs_reclaim ->&c->lock ->rcu_node_0 ->&n->list_lock ->&type->lock_class#2 ->key_user_lock ->root_key_user.cons_lock ->crngs.lock ->key_serial_lock ->key_construction_mutex ->init_user_ns.keyring_sem ->root_key_user.lock ->&x->wait#17 ->running_helpers_waitq.lock ->bit_wait_table + i ->&cfs_rq->removed.lock ->root_key_user.cons_lock.wait_lock ->&p->pi_lock ->key_construction_mutex.wait_lock ->&____s->seqcount#2 ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock ->remove_cache_srcu FD: 38 BD: 1 +.+.: put_task_map-wait-type-override ->&obj_hash[i].lock ->pool_lock#2 ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&base->lock ->task_group_lock ->stock_lock ->css_set_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 1 BD: 1 +.+.: pnfs_spinlock FD: 80 BD: 8 +.+.: &sn->pipefs_sb_lock ->&rq->__lock ->(rpc_pipefs_notifier_list).rwsem FD: 1 BD: 1 +.+.: nls_lock FD: 30 BD: 1 ..-.: &(&cache_cleaner)->timer FD: 1 BD: 1 +.+.: jffs2_compressor_list_lock FD: 1 BD: 1 +.+.: next_tag_value_lock FD: 12 BD: 1 +.-.: (&tcp_orphan_timer) ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 1 ....: log_redrive_lock FD: 2 BD: 1 ....: &TxAnchor.LazyLock ->jfs_commit_thread_wait.lock FD: 1 BD: 2 ....: jfs_commit_thread_wait.lock FD: 1 BD: 1 +.+.: jfsTxnLock FD: 39 BD: 1 +.+.: ocfs2_stack_lock ->(console_sem).lock FD: 1 BD: 1 +.+.: o2hb_callback_sem FD: 1 BD: 1 +.+.: o2net_handler_lock FD: 3 BD: 1 +.+.: subsys mutex#30 ->&k->k_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#21/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#19 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock ->&n->list_lock ->&rq->__lock ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock FD: 38 BD: 5 +.+.: &sb->s_type->i_lock_key#19 ->&dentry->d_lock FD: 76 BD: 1 +.+.: &type->s_umount_key#22/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&c->lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#20 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 38 BD: 2 +.+.: &sb->s_type->i_lock_key#20 ->&dentry->d_lock FD: 1 BD: 1 +.+.: cipso_v4_doi_list_lock FD: 223 BD: 55 +.+.: nf_hook_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->remove_cache_srcu ->stock_lock ->&____s->seqcount#2 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&rq->__lock ->cpu_hotplug_lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 50 BD: 1 ++++: alg_types_sem ->fs_reclaim ->pool_lock#2 ->&rq->__lock FD: 1 BD: 1 +.+.: dma_list_mutex FD: 67 BD: 2 ++++: asymmetric_key_parsers_sem ->(console_sem).lock ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->crypto_alg_sem ->&obj_hash[i].lock ->(crypto_chain).rwsem ->&x->wait#20 ->&base->lock ->&rq->__lock ->(&timer.timer) ->&n->list_lock FD: 733 BD: 1 +.+.: blkcg_pol_register_mutex ->blkcg_pol_mutex ->cgroup_mutex FD: 1 BD: 2 +.+.: elv_list_lock FD: 53 BD: 1 +.+.: crc_t10dif_mutex ->crypto_alg_sem ->fs_reclaim ->pool_lock#2 FD: 53 BD: 1 +.+.: crc64_rocksoft_mutex ->crypto_alg_sem ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: ts_mod_lock FD: 36 BD: 2 +.+.: (work_completion)(&p->wq) ->vmap_area_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->pool_lock#2 ->&rq->__lock ->&cfs_rq->removed.lock ->quarantine_lock ->&base->lock ->&meta->lock ->kfence_freelist_lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 1 +.+.: pci_ep_cfs_subsys.su_mutex FD: 65 BD: 1 +.+.: &default_group_class[depth - 1]#2/1 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&sb->s_type->i_mutex_key#7/2 FD: 1 BD: 3 +.+.: &sb->s_type->i_mutex_key#7/2 FD: 1 BD: 1 +.+.: pci_epf_mutex FD: 229 BD: 1 +.+.: ipmi_interfaces_mutex ->&k->list_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&k->k_lock ->&c->lock ->&____s->seqcount ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->pcpu_alloc_mutex ->cpu_hotplug_lock ->kthread_create_lock ->&p->pi_lock ->&rq->__lock ->&x->wait ->wq_pool_mutex ->&base->lock ->panic_notifier_list.lock FD: 2 BD: 1 +.+.: smi_watchers_mutex ->&ipmi_interfaces_srcu FD: 1 BD: 3 .+.?: &ipmi_interfaces_srcu FD: 1 BD: 1 +.+.: smi_infos_lock FD: 4 BD: 111 ....: mask_lock ->tmp_mask_lock FD: 3 BD: 112 ....: tmp_mask_lock ->tmpmask_lock ->&its->lock FD: 26 BD: 1 ....: &desc->wait_for_threads ->&p->pi_lock FD: 3 BD: 2 +.+.: subsys mutex#31 ->&k->k_lock FD: 27 BD: 3 +.+.: &dev->mutex#2 ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock FD: 110 BD: 1 +.+.: register_count_mutex ->&k->list_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&k->k_lock ->&c->lock ->&____s->seqcount ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock FD: 1 BD: 1 ....: thermal_cdev_ida.xa_lock FD: 1 BD: 1 ....: cpufreq_driver_lock FD: 3 BD: 1 +.+.: subsys mutex#32 ->&k->k_lock FD: 50 BD: 1 +.+.: scmi_requested_devices_mtx ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 ....: virtio_index_ida.xa_lock FD: 1 BD: 1 +.+.: subsys mutex#33 FD: 1 BD: 1 +.+.: vdpa_dev_lock FD: 86 BD: 2 ++++: &type->i_mutex_dir_key#2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->&c->lock ->&____s->seqcount ->namespace_sem ->&sem->wait_lock ->&rq->__lock ->&n->list_lock ->remove_cache_srcu ->rcu_node_0 ->&rcu_state.gp_wq ->&rcu_state.expedited_wq ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->tomoyo_ss ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#5 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&xattrs->lock ->&simple_offset_xa_lock ->smack_known_lock ->key ->pcpu_lock ->percpu_counters_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 3 BD: 1 +.+.: subsys mutex#34 ->&k->k_lock FD: 30 BD: 1 ..-.: lib/debugobjects.c:101 FD: 30 BD: 2 +.+.: (debug_obj_work).work ->&____s->seqcount ->pool_lock#2 ->&rq->__lock ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 25 BD: 9 +.+.: (work_completion)(&buf->work) ->&rq->__lock FD: 1 BD: 1 ....: rng_index_ida.xa_lock FD: 124 BD: 4 +.+.: &md->mutex ->pci_lock ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#5 ->&its->dev_alloc_lock ->&domain->mutex ->&irq_desc_lock_class ->tmpmask_lock ->&its->lock ->&root->kernfs_rwsem ->lock ->&obj_hash[i].lock ->sparse_irq_lock ->vmap_area_lock ->purge_vmap_area_lock ->&c->lock ->&____s->seqcount FD: 2 BD: 5 +.+.: &xa->xa_lock#5 ->pool_lock#2 FD: 52 BD: 8 +.+.: &its->dev_alloc_lock ->&its->lock ->fs_reclaim ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->lpi_range_lock ->&obj_hash[i].lock FD: 1 BD: 113 ....: tmpmask_lock FD: 1 BD: 1 +.+.: &dev->vqs_list_lock FD: 1 BD: 1 ....: &vp_dev->lock FD: 53 BD: 1 +.+.: rng_mutex ->&x->wait#13 ->fs_reclaim ->pool_lock#2 ->kthread_create_lock ->&p->pi_lock ->&rq->__lock ->&x->wait ->&obj_hash[i].lock FD: 1 BD: 2 ....: &x->wait#13 FD: 26 BD: 2 -.-.: &x->wait#14 ->&p->pi_lock FD: 27 BD: 1 +.+.: reading_mutex ->&x->wait#14 ->&rq->__lock FD: 1 BD: 1 ....: &dev->config_lock FD: 2 BD: 1 +.-.: drivers/char/random.c:1010 ->input_pool.lock FD: 1 BD: 1 ....: &dev->managed.lock FD: 76 BD: 1 +.+.: &type->s_umount_key#23/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#21 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 38 BD: 2 +.+.: &sb->s_type->i_lock_key#21 ->&dentry->d_lock FD: 8 BD: 197 ....: drm_minor_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount FD: 3 BD: 3 +.+.: subsys mutex#35 ->&k->k_lock FD: 50 BD: 17 +.+.: &dev->mode_config.idr_mutex ->fs_reclaim ->pool_lock#2 FD: 92 BD: 13 +.+.: crtc_ww_class_acquire ->crtc_ww_class_mutex ->fs_reclaim ->pool_lock#2 FD: 91 BD: 14 +.+.: crtc_ww_class_mutex ->reservation_ww_class_acquire ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&dev->mode_config.idr_mutex ->&dev->mode_config.blob_lock ->&crtc->commit_lock ->reservation_ww_class_mutex ->tk_core.seq.seqcount ->&vkms_out->lock ->&dev->vbl_lock ->&x->wait#15 ->(work_completion)(&vkms_state->composer_work) ->&base->lock ->&rq->__lock ->(&timer.timer) ->(work_completion)(&vkms_state->composer_work)#2 FD: 1 BD: 15 +.+.: &dev->mode_config.blob_lock FD: 1 BD: 1 ....: &xa->xa_lock#6 FD: 1 BD: 1 ....: &xa->xa_lock#7 FD: 1 BD: 16 ....: &dev->mode_config.connector_list_lock FD: 17 BD: 18 ..-.: &dev->vbl_lock ->&dev->vblank_time_lock FD: 136 BD: 1 .+.+: drm_connector_list_iter ->&dev->mode_config.connector_list_lock ->fs_reclaim ->pool_lock#2 ->&connector->mutex FD: 134 BD: 2 +.+.: &connector->mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#35 ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&dev->mode_config.idr_mutex ->connector_list_lock FD: 1 BD: 3 +.+.: connector_list_lock FD: 1 BD: 1 +.+.: &dev->filelist_mutex FD: 177 BD: 1 +.+.: &dev->clientlist_mutex ->&helper->lock ->registration_lock ->(console_sem).lock ->kernel_fb_helper_lock FD: 122 BD: 9 +.+.: &helper->lock ->fs_reclaim ->pool_lock#2 ->&client->modeset_mutex ->&obj_hash[i].lock ->&____s->seqcount ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&mgr->vm_lock ->&dev->object_name_lock ->&node->vm_lock ->&file_private->table_lock ->&dev->mode_config.idr_mutex ->&dev->mode_config.fb_lock ->&file->fbs_lock ->&prime_fpriv->lock ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->&dev->master_mutex ->reservation_ww_class_mutex ->&lock->wait_lock ->&rq->__lock FD: 94 BD: 11 +.+.: &client->modeset_mutex ->&dev->mode_config.mutex ->fs_reclaim ->pool_lock#2 ->crtc_ww_class_acquire FD: 93 BD: 12 +.+.: &dev->mode_config.mutex ->crtc_ww_class_acquire ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock FD: 2 BD: 10 +.+.: &mgr->vm_lock ->pool_lock#2 FD: 43 BD: 10 +.+.: &dev->object_name_lock ->lock FD: 4 BD: 197 +.+.: &file_private->table_lock ->pool_lock#2 ->&obj_hash[i].lock FD: 4 BD: 10 +.+.: &node->vm_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 10 +.+.: &dev->mode_config.fb_lock FD: 1 BD: 10 +.+.: &file->fbs_lock FD: 1 BD: 10 +.+.: &prime_fpriv->lock FD: 175 BD: 2 +.+.: registration_lock ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&rq->__lock ->&x->wait#11 ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#11 ->vt_switch_mutex ->(console_sem).lock ->console_lock FD: 50 BD: 3 +.+.: vt_switch_mutex ->fs_reclaim ->pool_lock#2 FD: 1 BD: 7 +.+.: &fb_info->lock FD: 1 BD: 7 ....: vt_event_lock FD: 95 BD: 10 +.+.: &dev->master_mutex ->&client->modeset_mutex FD: 1 BD: 15 +.+.: &crtc->commit_lock FD: 21 BD: 303 -.-.: &xa->xa_lock#8 ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->key#11 ->&s->s_inode_wblist_lock ->&base->lock ->key#12 ->&wb->work_lock ->stock_lock ->key#14 ->&xa->xa_lock#15 ->&pl->lock FD: 1 BD: 100 +.+.: &info->lock FD: 5 BD: 599 +.+.: lock#4 ->&lruvec->lru_lock ->&obj_hash[i].lock FD: 2 BD: 600 ....: &lruvec->lru_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 33 BD: 15 -.-.: &vkms_out->lock ->&dev->event_lock FD: 32 BD: 16 -.-.: &dev->event_lock ->&dev->vbl_lock ->&____s->seqcount#5 ->&x->wait#15 ->&obj_hash[i].lock ->pool_lock#2 ->&dev->vblank_time_lock ->&vblank->queue ->&base->lock FD: 1 BD: 21 ----: &____s->seqcount#5 FD: 26 BD: 17 -...: &x->wait#15 ->&p->pi_lock FD: 16 BD: 19 -.-.: &dev->vblank_time_lock ->tk_core.seq.seqcount ->&(&vblank->seqlock)->lock ->&obj_hash[i].lock ->hrtimer_bases.lock FD: 2 BD: 20 -.-.: &(&vblank->seqlock)->lock ->&____s->seqcount#5 FD: 1 BD: 15 +.+.: (work_completion)(&vkms_state->composer_work) FD: 1 BD: 11 ....: &helper->damage_lock FD: 124 BD: 2 +.+.: (work_completion)(&helper->damage_work) ->&helper->damage_lock ->&helper->lock FD: 1 BD: 17 -.-.: &vblank->queue FD: 1 BD: 15 +.+.: (work_completion)(&vkms_state->composer_work)#2 FD: 1 BD: 205 +.+.: &lock->wait_lock FD: 1 BD: 2 +.+.: kernel_fb_helper_lock FD: 1 BD: 1 +.+.: drivers_lock#2 FD: 1 BD: 1 +.+.: devices_lock FD: 1 BD: 10 ....: blk_queue_ida.xa_lock FD: 2 BD: 12 +.+.: &xa->xa_lock#9 ->pool_lock#2 FD: 33 BD: 202 ....: &q->queue_lock ->&blkcg->lock ->pool_lock#2 ->pcpu_lock ->&obj_hash[i].lock ->percpu_counters_lock ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 28 BD: 203 ....: &blkcg->lock ->pool_lock#2 ->&c->lock ->percpu_ref_switch_lock FD: 1 BD: 17 +.+.: &bdev->bd_size_lock FD: 3 BD: 12 +.+.: subsys mutex#36 ->&k->k_lock FD: 249 BD: 8 +.+.: &q->sysfs_dir_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&c->lock ->&____s->seqcount ->&q->sysfs_lock ->&obj_hash[i].lock ->&rq->__lock ->sysfs_symlink_target_lock ->kernfs_idr_lock ->&sem->wait_lock ->&p->pi_lock ->&____s->seqcount#2 ->&n->list_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->remove_cache_srcu ->&cfs_rq->removed.lock FD: 248 BD: 9 +.+.: &q->sysfs_lock ->&q->debugfs_mutex ->&q->unused_hctx_lock ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->cpu_hotplug_lock ->fs_reclaim ->&xa->xa_lock#10 ->pcpu_alloc_mutex ->&q->rq_qos_mutex ->&stats->lock ->lock ->&root->kernfs_rwsem ->&rq->__lock ->&cfs_rq->removed.lock ->&n->list_lock ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock FD: 90 BD: 11 +.+.: &q->debugfs_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&dentry->d_lock ->&fsnotify_mark_srcu ->&sb->s_type->i_lock_key#7 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->&obj_hash[i].lock ->pool_lock#2 ->mount_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&sem->wait_lock ->&p->pi_lock FD: 27 BD: 212 ..-.: percpu_ref_switch_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 27 BD: 8 +.+.: subsys mutex#37 ->&k->k_lock ->&rq->__lock FD: 1 BD: 9 ....: cgwb_lock FD: 1 BD: 8 +...: bdi_lock FD: 54 BD: 247 +.+.: inode_hash_lock ->&sb->s_type->i_lock_key#3 ->&sb->s_type->i_lock_key#22 ->&s->s_inode_list_lock ->&sb->s_type->i_lock_key#24 ->&sb->s_type->i_lock_key#30 ->&sb->s_type->i_lock_key#31 FD: 2 BD: 10 +.+.: bdev_lock ->&bdev->bd_holder_lock FD: 148 BD: 9 +.+.: &disk->open_mutex ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&xa->xa_lock#8 ->lock#4 ->mmu_notifier_invalidate_range_start ->&c->lock ->&mapping->i_private_lock ->tk_core.seq.seqcount ->&ret->b_uptodate_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->&sb->s_type->i_lock_key#3 ->lock#5 ->&lruvec->lru_lock ->&rq->__lock ->&folio_wait_table[i] ->(console_sem).lock ->console_owner_lock ->console_owner ->&cfs_rq->removed.lock ->&s->s_inode_list_lock ->pcpu_alloc_mutex ->&bdev->bd_size_lock ->&x->wait#9 ->ext_devt_ida.xa_lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&k->k_lock ->subsys mutex#36 ->&xa->xa_lock#9 ->inode_hash_lock ->bdev_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&lo->lo_mutex ->nbd_index_mutex ->&nbd->config_lock ->&new->lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&lock->wait_lock FD: 1 BD: 109 +.+.: &mapping->i_private_lock FD: 27 BD: 13 -.-.: &ret->b_uptodate_lock ->bit_wait_table + i FD: 1 BD: 595 +.+.: lock#5 FD: 50 BD: 2 +.+.: loop_ctl_mutex ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&rq->__lock ->&c->lock FD: 1 BD: 10 +.+.: &q->unused_hctx_lock FD: 2 BD: 12 +.+.: &xa->xa_lock#10 ->pool_lock#2 FD: 1 BD: 7 +.+.: &set->tag_list_lock FD: 30 BD: 11 +.+.: &q->mq_freeze_lock ->percpu_ref_switch_lock ->&q->mq_freeze_wq ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 26 BD: 213 ..-.: &q->mq_freeze_wq ->&p->pi_lock FD: 97 BD: 10 +.+.: &q->rq_qos_mutex ->&q->mq_freeze_lock ->percpu_ref_switch_lock ->&q->debugfs_mutex ->set->srcu ->&stats->lock ->(&cb->timer) ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 11 ....: &stats->lock FD: 72 BD: 13 +.+.: nbd_index_mutex ->fs_reclaim ->pool_lock#2 ->&nbd->config_lock ->rlock-AF_NETLINK ->&c->lock ->&rq->__lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 20 .+.+: set->srcu FD: 32 BD: 3 +.+.: (work_completion)(&(&q->requeue_work)->work) ->&q->requeue_lock ->&hctx->lock ->&__ctx->lock FD: 29 BD: 3 +.+.: (work_completion)(&(&hctx->run_work)->work) FD: 30 BD: 1 ..-.: &(&ops->cursor_work)->timer FD: 32 BD: 2 +.+.: (work_completion)(&(&ops->cursor_work)->work) ->(console_sem).lock ->&obj_hash[i].lock ->&base->lock FD: 266 BD: 1 +.+.: zram_index_mutex ->fs_reclaim ->pool_lock#2 ->blk_queue_ida.xa_lock ->&obj_hash[i].lock ->pcpu_alloc_mutex ->&rq->__lock ->bio_slab_lock ->percpu_counters_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->lock ->&q->queue_lock ->&x->wait#9 ->&bdev->bd_size_lock ->&k->list_lock ->gdp_mutex ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->subsys mutex#36 ->dev_hotplug_mutex ->&q->sysfs_dir_lock ->percpu_ref_switch_lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#37 ->cgwb_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->bdi_lock ->inode_hash_lock ->(console_sem).lock FD: 3 BD: 1 +.+.: subsys mutex#38 ->&k->k_lock FD: 51 BD: 2 +.+.: &default_group_class[depth - 1]#3 ->fs_reclaim ->pool_lock#2 ->configfs_dirent_lock FD: 2 BD: 1 +.+.: &lock ->nullb_indexes.xa_lock FD: 1 BD: 2 ....: nullb_indexes.xa_lock FD: 2 BD: 2 ....: nfc_index_ida.xa_lock ->pool_lock#2 FD: 138 BD: 4 +.+.: nfc_devlist_mutex ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->subsys mutex#39 ->&k->k_lock ->&genl_data->genl_data_mutex ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->&x->wait#9 ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->&n->list_lock ->&rq->__lock ->nfc_devlist_mutex.wait_lock ->&sem->wait_lock ->&p->pi_lock ->&____s->seqcount#2 ->uevent_sock_mutex.wait_lock ->quarantine_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->remove_cache_srcu ->kn->active#4 ->dev_pm_qos_sysfs_mtx.wait_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->rcu_node_0 ->&rcu_state.expedited_wq ->gdp_mutex.wait_lock FD: 27 BD: 5 +.+.: subsys mutex#39 ->&k->k_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 4 BD: 4 +.+.: llcp_devices_lock ->&k->list_lock ->&k->k_lock FD: 1 BD: 57 ....: &rfkill->lock FD: 27 BD: 14 +.+.: subsys mutex#40 ->&k->k_lock ->&rq->__lock FD: 125 BD: 6 +.+.: (work_completion)(&rfkill->sync_work) ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 2 BD: 1 +.+.: dma_heap_minors.xa_lock ->pool_lock#2 FD: 3 BD: 1 +.+.: subsys mutex#41 ->&k->k_lock FD: 1 BD: 1 +.+.: heap_list_lock FD: 4 BD: 1 +.+.: subsys mutex#42 ->&k->list_lock ->&k->k_lock FD: 1 BD: 1 +.+.: nvmf_hosts_mutex FD: 3 BD: 1 +.+.: subsys mutex#43 ->&k->k_lock FD: 1 BD: 1 +.+.: nvmf_transports_rwsem FD: 3 BD: 1 +.+.: subsys mutex#44 ->&k->k_lock FD: 1 BD: 3 ....: nvme_instance_ida.xa_lock FD: 3 BD: 3 +.+.: subsys mutex#45 ->&k->k_lock FD: 96 BD: 3 +.+.: pools_reg_lock ->pools_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem FD: 1 BD: 4 +.+.: pools_lock FD: 68 BD: 3 +.+.: &default_group_class[depth - 1]#4/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#5/2 FD: 1 BD: 1 +.+.: nvmet_config_sem FD: 3 BD: 1 +.+.: subsys mutex#46 ->&k->k_lock FD: 2 BD: 3 ....: &ctrl->lock ->&ctrl->state_wq FD: 1 BD: 4 ....: &ctrl->state_wq FD: 1 BD: 9 +.+.: &hctx->lock FD: 1 BD: 113 +.+.: &nvmeq->sq_lock FD: 26 BD: 6 ..-.: &x->wait#16 ->&p->pi_lock FD: 120 BD: 6 +.+.: nvme_subsystems_lock ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->subsys mutex#47 FD: 3 BD: 7 +.+.: subsys mutex#47 ->&k->k_lock FD: 1 BD: 7 +.+.: &xa->xa_lock#11 FD: 127 BD: 1 +.+.: &dev->shutdown_lock ->&md->mutex ->&desc->request_mutex ->&obj_hash[i].lock ->pool_lock#2 ->pci_lock ->&rq->__lock ->&cfs_rq->removed.lock ->fs_reclaim ->&c->lock ->&____s->seqcount ->free_vmap_area_lock ->vmap_area_lock ->register_lock ->&irq_desc_lock_class ->proc_subdir_lock ->proc_inum_ida.xa_lock FD: 1 BD: 111 ....: irq_resend_lock FD: 1 BD: 110 +.+.: &ent->pde_unload_lock FD: 67 BD: 4 +.+.: &default_group_class[depth - 1]#5/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#6/2 FD: 66 BD: 5 +.+.: &default_group_class[depth - 1]#6/2 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->configfs_dirent_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#18 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&default_group_class[depth - 1]#7 ->&default_group_class[depth - 1]#7/2 FD: 51 BD: 6 +.+.: &default_group_class[depth - 1]#7 ->fs_reclaim ->pool_lock#2 ->configfs_dirent_lock FD: 1 BD: 6 +.+.: &default_group_class[depth - 1]#7/2 FD: 1 BD: 1 +.+.: backend_mutex FD: 1 BD: 1 +.+.: scsi_mib_index_lock FD: 1 BD: 1 +.+.: hba_lock FD: 1 BD: 6 ++++: &ctrl->namespaces_rwsem FD: 50 BD: 1 +.+.: device_mutex ->fs_reclaim ->pool_lock#2 FD: 295 BD: 3 +.+.: (wq_completion)nvme-wq ->(work_completion)(&ctrl->async_event_work) ->(work_completion)(&ctrl->scan_work) ->(work_completion)(&barr->work) FD: 2 BD: 4 +.+.: (work_completion)(&ctrl->async_event_work) ->&nvmeq->sq_lock FD: 293 BD: 4 +.+.: (work_completion)(&ctrl->scan_work) ->&ctrl->scan_lock FD: 292 BD: 5 +.+.: &ctrl->scan_lock ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&hctx->lock ->&x->wait#16 ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->(&timer.timer) ->&ctrl->namespaces_rwsem ->blk_queue_ida.xa_lock ->pcpu_alloc_mutex ->&q->sysfs_lock ->&set->tag_list_lock ->bio_slab_lock ->percpu_counters_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->lock ->&q->mq_freeze_lock ->percpu_ref_switch_lock ->&q->queue_lock ->&x->wait#9 ->nvme_subsystems_lock ->&subsys->lock ->&bdev->bd_size_lock ->ext_devt_ida.xa_lock ->&k->list_lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&k->k_lock ->subsys mutex#36 ->dev_hotplug_mutex ->&q->sysfs_dir_lock ->gdp_mutex ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#37 ->cgwb_lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->bdi_lock ->inode_hash_lock ->bdev_lock ->&disk->open_mutex ->nvme_ns_chr_minor_ida.xa_lock ->chrdevs_lock ->subsys mutex#50 ->&dentry->d_lock ->&cfs_rq->removed.lock ->quarantine_lock FD: 60 BD: 6 +.+.: &subsys->lock ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#11 ->&obj_hash[i].lock ->pcpu_alloc_mutex FD: 1 BD: 1 +.+.: &hba->device_lock FD: 1 BD: 10 ....: ext_devt_ida.xa_lock FD: 1 BD: 1 +.+.: part_parser_lock FD: 282 BD: 1 +.+.: mtd_table_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->&cfs_rq->removed.lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#48 ->devtree_lock ->nvmem_ida.xa_lock ->nvmem_cell_mutex ->&k->k_lock ->subsys mutex#49 ->nvmem_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->(console_sem).lock ->pcpu_alloc_mutex ->batched_entropy_u32.lock ->mmu_notifier_invalidate_range_start ->blk_queue_ida.xa_lock ->&q->sysfs_lock ->&set->tag_list_lock ->bio_slab_lock ->percpu_counters_lock ->&sb->s_type->i_lock_key#3 ->&s->s_inode_list_lock ->&xa->xa_lock#9 ->&q->mq_freeze_lock ->set->srcu ->percpu_ref_switch_lock ->&q->queue_lock ->&bdev->bd_size_lock ->elv_list_lock ->(work_completion)(&(&q->requeue_work)->work) ->(work_completion)(&(&hctx->run_work)->work) ->&q->debugfs_mutex ->subsys mutex#36 ->dev_hotplug_mutex ->&q->sysfs_dir_lock ->subsys mutex#37 ->cgwb_lock ->bdi_lock ->inode_hash_lock FD: 1 BD: 1 +.+.: chip_drvs_lock FD: 1 BD: 8 ....: (kmod_concurrent_max).lock FD: 26 BD: 11 ....: &x->wait#17 ->&p->pi_lock FD: 1 BD: 3 ....: &prev->lock FD: 3 BD: 2 +.+.: subsys mutex#48 ->&k->k_lock FD: 1 BD: 2 ....: nvmem_ida.xa_lock FD: 1 BD: 2 +.+.: nvmem_cell_mutex FD: 1 BD: 2 +.+.: subsys mutex#49 FD: 1 BD: 2 +.+.: nvmem_mutex FD: 26 BD: 588 -.-.: &folio_wait_table[i] ->&p->pi_lock FD: 1 BD: 49 +.+.: &bond->stats_lock FD: 16 BD: 63 ....: lweventlist_lock ->pool_lock#2 ->&dir->lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&base->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&____s->seqcount#2 FD: 724 BD: 2 +.+.: (linkwatch_work).work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 64 +.+.: rtnl_mutex.wait_lock FD: 3 BD: 50 ..-.: once_lock ->crngs.lock FD: 25 BD: 49 ++++: (inet6addr_validator_chain).rwsem ->&rq->__lock FD: 25 BD: 49 ++++: (inetaddr_validator_chain).rwsem ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 34 BD: 1 +.+.: (wq_completion)gid-cache-wq ->(work_completion)(&ndev_work->work) ->(work_completion)(&work->work) ->&rq->__lock FD: 32 BD: 2 +.+.: (work_completion)(&ndev_work->work) ->devices_rwsem ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->&base->lock ->&cfs_rq->removed.lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 228 BD: 2 +.+.: (work_completion)(&w->work) ->cpu_hotplug_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 589 ....: &sem->wait_lock FD: 1 BD: 6 ....: nvme_ns_chr_minor_ida.xa_lock FD: 3 BD: 6 +.+.: subsys mutex#50 ->&k->k_lock FD: 3 BD: 1 +.+.: subsys mutex#51 ->&k->k_lock FD: 1 BD: 1 +.+.: gpio_lookup_lock FD: 1 BD: 1 +.+.: mdio_board_lock FD: 1 BD: 1 +.+.: mode_list_lock FD: 1 BD: 1 +.+.: l3mdev_lock FD: 1 BD: 102 -.-.: &retval->lock FD: 44 BD: 1 +.+.: (wq_completion)gve ->(work_completion)(&priv->service_task) FD: 43 BD: 2 +.+.: (work_completion)(&priv->service_task) ->(console_sem).lock ->lweventlist_lock ->&rq->__lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: hnae3_common_lock FD: 3 BD: 1 +.+.: subsys mutex#52 ->&k->k_lock FD: 8 BD: 1 +.+.: compressor_list_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount FD: 1 BD: 5 ....: hwsim_netgroup_ida.xa_lock FD: 34 BD: 78 +.-.: hwsim_radio_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&list->lock#16 ->&____s->seqcount#2 ->&zone->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->init_task.mems_allowed_seq.seqcount FD: 3 BD: 8 +.+.: subsys mutex#53 ->&k->k_lock FD: 52 BD: 50 +.+.: param_lock ->rate_ctrl_mutex ->disk_events_mutex FD: 375 BD: 52 +.+.: &rdev->wiphy.mtx ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#54 ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->nl_table_lock ->nl_table_wait.lock ->reg_requests_lock ->stack_depot_init_mutex ->pcpu_alloc_mutex ->&xa->xa_lock#4 ->net_rwsem ->&x->wait#9 ->subsys mutex#20 ->&dir->lock#2 ->dev_hotplug_mutex ->dev_base_lock ->input_pool.lock ->batched_entropy_u32.lock ->&tbl->lock ->sysctl_lock ->&fq->lock ->&local->iflist_mtx ->&rq->__lock ->rlock-AF_NETLINK ->&rdev->bss_lock ->lweventlist_lock ->&base->lock ->&data->mutex ->&rdev->wiphy_work_lock ->&local->filter_lock ->pool_lock ->&tn->lock ->remove_cache_srcu ->&n->list_lock ->failover_lock ->proc_subdir_lock ->proc_inum_ida.xa_lock ->&idev->mc_lock ->&pnettable->lock ->smc_ib_devices.mutex ->&ndev->lock ->&wdev->event_lock ->&rdev->mgmt_registrations_lock ->(&dwork->timer) ->&dentry->d_lock ->&fsnotify_mark_srcu ->&sb->s_type->i_lock_key#7 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->mount_lock ->(&dwork->timer)#2 ->(work_completion)(&(&link->color_collision_detect_work)->work) ->rtnl_mutex.wait_lock ->&p->pi_lock ->&list->lock#15 ->&ifibss->incomplete_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->tk_core.seq.seqcount ->hrtimer_bases.lock ->&list->lock#2 ->quarantine_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&lock->wait_lock ->&sta->lock ->lock#6 ->&____s->seqcount#2 ->&sem->wait_lock ->&local->active_txq_lock[i] ->(work_completion)(&sta->drv_deliver_wk) ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&wdev->pmsr_lock ->krc.lock ->&local->queue_stop_reason_lock ->&x->wait#2 ->(&ifibss->timer) ->_xmit_ETHER ->(&local->dynamic_ps_timer) ->(&dwork->timer)#3 ->&list->lock#16 ->&wq->mutex ->cpu_hotplug_lock ->&rnp->exp_wq[0] ->bpf_devs_lock ->&in_dev->mc_tomb_lock ->class ->(&tbl->proxy_timer) ->&ul->lock ->&net->xdp.lock ->mirred_list_lock ->&nft_net->commit_mutex ->&ent->pde_unload_lock ->&net->ipv6.addrconf_hash_lock ->&idev->mc_query_lock ->(work_completion)(&(&idev->mc_report_work)->work) ->&idev->mc_report_lock ->&pnn->pndevs.lock ->&pnn->routes.lock ->target_list_lock ->kernfs_idr_lock ->dev_pm_qos_sysfs_mtx ->deferred_probe_mutex ->device_links_lock ->&rnp->exp_wq[2] ->(&dwork->timer)#4 ->mmu_notifier_invalidate_range_start ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->rcu_state.exp_mutex.wait_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->&ul->lock#2 ->gdp_mutex.wait_lock ->uevent_sock_mutex.wait_lock ->(wq_completion)phy228 ->(wq_completion)phy227 ->(wq_completion)phy226 ->(wq_completion)phy225 ->(wq_completion)phy230 ->(wq_completion)phy229 ->(wq_completion)phy214 ->(wq_completion)phy213 ->(wq_completion)phy222 ->(wq_completion)phy221 FD: 3 BD: 53 +.+.: subsys mutex#54 ->&k->k_lock FD: 1 BD: 53 +.+.: reg_requests_lock FD: 1 BD: 14 +.+.: rfkill_global_mutex.wait_lock FD: 15 BD: 77 +.-.: &fq->lock ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&zone->lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->&base->lock FD: 1 BD: 53 +.+.: &local->iflist_mtx FD: 3 BD: 49 +.+.: subsys mutex#55 ->&k->k_lock FD: 2 BD: 50 +.+.: &sdata->sec_mtx ->&sec->lock FD: 1 BD: 51 +...: &sec->lock FD: 1 BD: 49 +.+.: &local->iflist_mtx#2 FD: 50 BD: 1 +.+.: hwsim_phys_lock ->fs_reclaim ->pool_lock#2 FD: 50 BD: 1 +.+.: xdomain_lock ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: ioctl_mutex FD: 1 BD: 1 +.+.: address_handler_list_lock FD: 1 BD: 1 +.+.: card_mutex FD: 3 BD: 1 +.+.: subsys mutex#56 ->&k->k_lock FD: 26 BD: 1 ....: &x->wait#18 ->&p->pi_lock FD: 28 BD: 2 ..-.: &txlock ->&list->lock#3 ->&txwq FD: 1 BD: 3 ..-.: &list->lock#3 FD: 26 BD: 3 ..-.: &txwq ->&p->pi_lock FD: 2 BD: 1 ....: &iocq[i].lock ->&ktiowq[i] FD: 1 BD: 2 ....: &ktiowq[i] FD: 1 BD: 1 ....: rcu_read_lock_bh FD: 1 BD: 61 +.-.: noop_qdisc.q.lock FD: 3 BD: 3 +.+.: subsys mutex#57 ->&k->k_lock FD: 176 BD: 1 +.+.: usb_bus_idr_lock ->(usb_notifier_list).rwsem ->fs_reclaim ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->hcd_root_hub_lock ->&obj_hash[i].lock ->&rq->__lock ->&x->wait#19 ->&dev->power.lock ->device_links_srcu ->&c->lock ->(console_sem).lock ->input_pool.lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&(&priv->bus_notifier)->rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#58 ->&x->wait#9 ->&lock->wait_lock ->&vhci_hcd->vhci->lock ->&base->lock ->(&timer.timer) ->batched_entropy_u8.lock ->kfence_freelist_lock ->quarantine_lock FD: 119 BD: 1 +.+.: table_lock ->&k->list_lock ->fs_reclaim ->pool_lock#2 ->lock ->&root->kernfs_rwsem ->&k->k_lock ->uevent_sock_mutex ->&obj_hash[i].lock ->running_helpers_waitq.lock ->(console_sem).lock ->&rq->__lock ->&c->lock ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 2 BD: 1 +.-.: (&ipmi_timer) ->&ipmi_interfaces_srcu FD: 1 BD: 3 +.+.: mon_lock FD: 114 BD: 2 +.+.: usb_port_peer_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->&k->k_lock ->dev_pm_qos_mtx ->component_mutex ->device_links_srcu ->dev_pm_qos_sysfs_mtx ->&rq->__lock ->sysfs_symlink_target_lock FD: 1 BD: 4 ....: device_state_lock FD: 28 BD: 8 ....: hcd_root_hub_lock ->hcd_urb_list_lock ->&bh->lock ->&p->pi_lock FD: 1 BD: 9 ....: hcd_urb_list_lock FD: 1 BD: 9 ..-.: &bh->lock FD: 14 BD: 79 ..-.: lock#6 ->kcov_remote_lock ->&kcov->lock FD: 12 BD: 116 ..-.: kcov_remote_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 26 BD: 8 ..-.: &x->wait#19 ->&p->pi_lock FD: 1 BD: 2 +.+.: set_config_lock FD: 63 BD: 2 +.+.: hcd->bandwidth_mutex ->devtree_lock ->&obj_hash[i].lock ->&x->wait#9 ->&dev->power.lock ->fs_reclaim ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->hcd_root_hub_lock ->&rq->__lock ->&x->wait#19 ->&base->lock ->(&timer.timer) ->&c->lock ->&____s->seqcount FD: 1 BD: 2 +.+.: &new_driver->dynids.lock FD: 1 BD: 7 ....: &dum_hcd->dum->lock FD: 57 BD: 4 +.+.: &hub->status_mutex ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->hcd_root_hub_lock ->fs_reclaim ->&dum_hcd->dum->lock ->&obj_hash[i].lock ->&rq->__lock ->&x->wait#19 ->&base->lock ->(&timer.timer) ->&vhci_hcd->vhci->lock ->&____s->seqcount ->&cfs_rq->removed.lock ->&c->lock FD: 1 BD: 3 +.+.: component_mutex FD: 62 BD: 2 +.+.: (work_completion)(&(&hub->init_work)->work) ->&rq->__lock ->&lock->wait_lock ->&p->pi_lock FD: 1 BD: 2 +.+.: subsys mutex#58 FD: 38 BD: 1 +.+.: (wq_completion)usb_hub_wq ->(work_completion)(&hub->events) FD: 37 BD: 2 +.+.: (work_completion)(&hub->events) ->lock#6 ->&dev->power.lock FD: 1 BD: 3 ....: &hub->irq_urb_lock FD: 1 BD: 3 ....: (&hub->irq_urb_retry) FD: 1 BD: 3 ....: hcd_urb_unlink_lock FD: 26 BD: 3 ..-.: usb_kill_urb_queue.lock ->&p->pi_lock FD: 1 BD: 3 +.+.: (work_completion)(&hub->tt.clear_work) FD: 50 BD: 20 +.+.: udc_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&rq->__lock FD: 3 BD: 1 +.+.: subsys mutex#59 ->&k->k_lock FD: 1 BD: 1 ....: gadget_id_numbers.xa_lock FD: 87 BD: 2 +.+.: (work_completion)(&gadget->work) ->&root->kernfs_rwsem ->kernfs_notify_lock FD: 30 BD: 111 ....: kernfs_notify_lock FD: 60 BD: 2 +.+.: kernfs_notify_work ->kernfs_notify_lock ->&root->kernfs_supers_rwsem ->&rq->__lock FD: 55 BD: 9 ++++: &root->kernfs_supers_rwsem ->inode_hash_lock FD: 1 BD: 1 +.+.: subsys mutex#60 FD: 1 BD: 1 +.+.: func_lock FD: 1 BD: 1 +.+.: g_tf_lock FD: 1 BD: 7 ....: &vhci_hcd->vhci->lock FD: 3 BD: 2 ....: input_ida.xa_lock ->&____s->seqcount ->pool_lock#2 FD: 1 BD: 1 ....: &mousedev->mutex/1 FD: 30 BD: 4 ....: serio_event_lock ->pool_lock#2 FD: 58 BD: 1 +.+.: (wq_completion)events_long ->serio_event_work ->(work_completion)(&(&ipvs->defense_work)->work) ->(work_completion)(&(&br->gc_work)->work) ->(work_completion)(&br->mcast_gc_work) ->&rq->__lock FD: 35 BD: 2 +.+.: serio_event_work ->serio_mutex FD: 34 BD: 3 +.+.: serio_mutex ->serio_event_lock ->&k->list_lock ->&k->k_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 ....: rtc_ida.xa_lock FD: 32 BD: 1 +.+.: &rtc->ops_lock ->(efi_runtime_lock).lock ->&obj_hash[i].lock ->&x->wait#12 ->&rq->__lock FD: 1 BD: 2 ....: platform_devid_ida.xa_lock FD: 1 BD: 2 ....: rtcdev_lock FD: 50 BD: 1 +.+.: g_smscore_deviceslock ->fs_reclaim ->pool_lock#2 FD: 1 BD: 1 +.+.: cx231xx_devlist_mutex FD: 1 BD: 1 +.+.: em28xx_devlist_mutex FD: 1 BD: 1 ....: pvr2_context_sync_data.lock FD: 1 BD: 8 +.+.: i2c_dev_list_lock FD: 3 BD: 8 +.+.: subsys mutex#61 ->&k->k_lock FD: 1 BD: 1 +.+.: subsys mutex#62 FD: 132 BD: 2 +.+.: dvbdev_register_lock ->(console_sem).lock ->fs_reclaim ->pool_lock#2 ->minor_rwsem ->&xa->xa_lock#12 ->&mdev->graph_mutex ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->&c->lock ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->uevent_sock_mutex ->running_helpers_waitq.lock ->&k->k_lock ->subsys mutex#63 FD: 133 BD: 1 +.+.: frontend_mutex ->fs_reclaim ->pool_lock#2 ->(console_sem).lock ->dvbdev_register_lock FD: 1 BD: 3 +.+.: minor_rwsem FD: 2 BD: 3 ....: &xa->xa_lock#12 ->pool_lock#2 FD: 50 BD: 4 +.+.: &mdev->graph_mutex ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount FD: 3 BD: 3 +.+.: subsys mutex#63 ->&k->k_lock FD: 1 BD: 1 ....: &dmxdev->lock FD: 1 BD: 1 +.+.: &dvbdemux->mutex FD: 1 BD: 1 +.+.: media_devnode_lock FD: 1 BD: 1 +.+.: subsys mutex#64 FD: 1 BD: 1 +.+.: videodev_lock FD: 3 BD: 1 +.+.: subsys mutex#65 ->&k->k_lock FD: 1 BD: 1 +.+.: vimc_sensor:396:(&vsensor->hdl)->_lock FD: 1 BD: 1 +.+.: &v4l2_dev->lock FD: 1 BD: 1 +.+.: vimc_debayer:581:(&vdebayer->hdl)->_lock FD: 1 BD: 1 +.+.: vimc_lens:61:(&vlens->hdl)->_lock FD: 60 BD: 1 +.+.: vivid_ctrls:1606:(hdl_user_gen)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->&c->lock ->vivid_ctrls:1622:(hdl_vid_out)->_lock ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock ->vivid_ctrls:1627:(hdl_vbi_out)->_lock ->vivid_ctrls:1630:(hdl_radio_rx)->_lock ->vivid_ctrls:1632:(hdl_radio_tx)->_lock ->vivid_ctrls:1634:(hdl_sdr_cap)->_lock ->vivid_ctrls:1636:(hdl_meta_cap)->_lock ->vivid_ctrls:1638:(hdl_meta_out)->_lock ->vivid_ctrls:1640:(hdl_tch_cap)->_lock ->&zone->lock ->&obj_hash[i].lock FD: 51 BD: 1 +.+.: vivid_ctrls:1608:(hdl_user_vid)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 FD: 54 BD: 1 +.+.: vivid_ctrls:1610:(hdl_user_aud)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->vivid_ctrls:1622:(hdl_vid_out)->_lock ->vivid_ctrls:1630:(hdl_radio_rx)->_lock ->vivid_ctrls:1632:(hdl_radio_tx)->_lock FD: 58 BD: 1 +.+.: vivid_ctrls:1612:(hdl_streaming)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->vivid_ctrls:1622:(hdl_vid_out)->_lock ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock ->vivid_ctrls:1627:(hdl_vbi_out)->_lock ->vivid_ctrls:1634:(hdl_sdr_cap)->_lock ->vivid_ctrls:1636:(hdl_meta_cap)->_lock ->vivid_ctrls:1638:(hdl_meta_out)->_lock ->vivid_ctrls:1640:(hdl_tch_cap)->_lock FD: 52 BD: 1 +.+.: vivid_ctrls:1614:(hdl_sdtv_cap)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock FD: 52 BD: 1 +.+.: vivid_ctrls:1616:(hdl_loop_cap)->_lock ->vivid_ctrls:1620:(hdl_vid_cap)->_lock ->fs_reclaim ->pool_lock#2 ->vivid_ctrls:1625:(hdl_vbi_cap)->_lock FD: 1 BD: 1 +.+.: vivid_ctrls:1618:(hdl_fb)->_lock FD: 1 BD: 7 +.+.: vivid_ctrls:1620:(hdl_vid_cap)->_lock FD: 1 BD: 4 +.+.: vivid_ctrls:1622:(hdl_vid_out)->_lock FD: 1 BD: 5 +.+.: vivid_ctrls:1625:(hdl_vbi_cap)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1627:(hdl_vbi_out)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1630:(hdl_radio_rx)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1632:(hdl_radio_tx)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1634:(hdl_sdr_cap)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1636:(hdl_meta_cap)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1638:(hdl_meta_out)->_lock FD: 1 BD: 3 +.+.: vivid_ctrls:1640:(hdl_tch_cap)->_lock FD: 1 BD: 1 ....: &adap->kthread_waitq FD: 1 BD: 1 +.+.: &dev->cec_xfers_slock FD: 1 BD: 1 ....: &dev->kthread_waitq_cec FD: 1 BD: 1 +.+.: cec_devnode_lock FD: 1 BD: 1 +.+.: subsys mutex#66 FD: 5 BD: 1 +.+.: &adap->lock ->tk_core.seq.seqcount ->&adap->devnode.lock_fhs FD: 1 BD: 2 +.+.: &adap->devnode.lock_fhs FD: 1 BD: 1 ....: ptp_clocks_map.xa_lock FD: 3 BD: 1 +.+.: subsys mutex#67 ->&k->k_lock FD: 1 BD: 1 +.+.: pers_lock FD: 1 BD: 1 +.+.: _lock FD: 1 BD: 3 +.+.: dm_bufio_clients_lock FD: 1 BD: 1 +.+.: _ps_lock FD: 1 BD: 1 +.+.: _lock#2 FD: 1 BD: 1 +.+.: _lock#3 FD: 1 BD: 1 +.+.: register_lock#2 FD: 3 BD: 1 +.+.: subsys mutex#68 ->&k->k_lock FD: 1 BD: 1 .+.+: bp_lock FD: 3 BD: 1 +.+.: subsys mutex#69 ->&k->k_lock FD: 14 BD: 1 +.-.: (&dsp_spl_tl) ->dsp_lock FD: 13 BD: 2 ..-.: dsp_lock ->iclock_lock ->&obj_hash[i].lock ->&base->lock FD: 4 BD: 3 ...-: iclock_lock ->tk_core.seq.seqcount FD: 51 BD: 49 +.+.: lock#7 ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#14 FD: 1 BD: 1 ....: iscsi_transport_lock FD: 3 BD: 1 +.+.: subsys mutex#70 ->&k->k_lock FD: 1 BD: 1 ....: &tx_task->waiting FD: 1 BD: 1 +.+.: link_ops_rwsem FD: 130 BD: 1 +.+.: disable_lock ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&k->k_lock ->&dev->power.lock ->dpm_list_mtx ->&(&priv->bus_notifier)->rwsem ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#4 FD: 2 BD: 1 +.+.: protocol_lock ->pool_lock#2 FD: 72 BD: 1 +.+.: psinfo_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->(console_sem).lock ->&rq->__lock ->pstore_sb_lock ->dump_list_lock FD: 58 BD: 3 +.+.: pstore_sb_lock ->&sb->s_type->i_mutex_key#12 FD: 1 BD: 2 ....: dump_list_lock FD: 1 BD: 1 +.+.: vsock_register_mutex FD: 1 BD: 1 +.+.: comedi_drivers_list_lock FD: 105 BD: 1 +.+.: cscfg_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&dev->power.lock ->dpm_list_mtx FD: 1 BD: 589 +.+.: icc_bw_lock FD: 27 BD: 6 +.+.: subsys mutex#71 ->&rq->__lock ->&k->k_lock FD: 109 BD: 2 ++++: snd_ctl_layer_rwsem ->snd_ctl_led_mutex ->fs_reclaim ->pool_lock#2 ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->&c->lock ->&____s->seqcount ->&dev->power.lock ->dpm_list_mtx ->&k->k_lock ->sysfs_symlink_target_lock FD: 1 BD: 3 +.+.: snd_card_mutex FD: 1 BD: 1 +.+.: snd_ioctl_rwsem FD: 50 BD: 2 +.+.: strings ->fs_reclaim ->pool_lock#2 FD: 1 BD: 2 +.+.: register_mutex FD: 122 BD: 3 +.+.: sound_mutex ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&x->wait#11 ->&rq->__lock ->&obj_hash[i].lock ->uevent_sock_mutex ->&c->lock ->&____s->seqcount ->running_helpers_waitq.lock ->subsys mutex#71 ->&cfs_rq->removed.lock ->&k->k_lock FD: 132 BD: 1 +.+.: register_mutex#2 ->fs_reclaim ->pool_lock#2 ->sound_mutex ->&obj_hash[i].lock ->register_mutex ->&c->lock ->&____s->seqcount ->sound_oss_mutex ->strings ->&entry->access ->info_mutex FD: 124 BD: 1 +.+.: register_mutex#3 ->fs_reclaim ->pool_lock#2 ->sound_mutex ->clients_lock FD: 1 BD: 5 ....: clients_lock FD: 2 BD: 1 +.+.: &client->ports_mutex ->&client->ports_lock FD: 1 BD: 5 .+.+: &client->ports_lock FD: 125 BD: 1 +.+.: register_mutex#4 ->fs_reclaim ->pool_lock#2 ->sound_oss_mutex FD: 124 BD: 3 +.+.: sound_oss_mutex ->fs_reclaim ->pool_lock#2 ->sound_loader_lock ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->gdp_mutex ->&____s->seqcount ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&dev->power.lock ->dpm_list_mtx ->req_lock ->&p->pi_lock ->&rq->__lock ->&x->wait#11 ->uevent_sock_mutex ->running_helpers_waitq.lock ->subsys mutex#71 ->&k->k_lock FD: 1 BD: 4 +.+.: sound_loader_lock FD: 53 BD: 1 .+.+: &grp->list_mutex/1 ->clients_lock ->&client->ports_lock ->register_lock#3 ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount FD: 2 BD: 1 +.+.: &grp->list_mutex#2 ->&grp->list_lock FD: 1 BD: 2 ....: &grp->list_lock FD: 63 BD: 2 +.+.: async_lookup_work ->fs_reclaim ->pool_lock#2 ->clients_lock ->&client->ports_lock ->snd_card_mutex ->(kmod_concurrent_max).lock ->&obj_hash[i].lock ->&x->wait#17 ->&rq->__lock ->&cfs_rq->removed.lock ->running_helpers_waitq.lock ->autoload_work ->&x->wait#10 FD: 1 BD: 2 ....: register_lock#3 FD: 112 BD: 1 ++++: &card->controls_rwsem ->&xa->xa_lock#13 ->fs_reclaim ->&card->ctl_files_rwlock ->snd_ctl_layer_rwsem ->pool_lock#2 FD: 9 BD: 2 +.+.: &xa->xa_lock#13 ->pool_lock#2 ->&c->lock ->&____s->seqcount FD: 1 BD: 2 ....: &card->ctl_files_rwlock FD: 4 BD: 3 +.+.: autoload_work ->&k->list_lock ->&k->k_lock FD: 1 BD: 3 +.+.: snd_ctl_led_mutex FD: 1 BD: 1 +.+.: register_mutex#5 FD: 51 BD: 1 +.+.: client_mutex ->fs_reclaim ->pool_lock#2 ->&dev->devres_lock FD: 1 BD: 53 +.+.: failover_lock FD: 2 BD: 2 +...: llc_sap_list_lock ->pool_lock#2 FD: 50 BD: 1 +.+.: act_id_mutex ->fs_reclaim ->pool_lock#2 ->&____s->seqcount FD: 1 BD: 1 +.+.: act_mod_lock FD: 1 BD: 1 +.+.: ife_mod_lock FD: 1 BD: 1 +.+.: cls_mod_lock FD: 1 BD: 1 +.+.: ematch_mod_lock FD: 1 BD: 1 +.+.: sock_diag_table_mutex FD: 1 BD: 1 +.+.: nfnl_subsys_acct FD: 1 BD: 1 +.+.: nfnl_subsys_queue FD: 1 BD: 1 +.+.: nfnl_subsys_ulog FD: 25 BD: 5 +.+.: nf_log_mutex ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: nfnl_subsys_osf FD: 33 BD: 1 +.+.: nf_sockopt_mutex ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->nf_sockopt_mutex.wait_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 1 +.+.: nfnl_subsys_ctnetlink FD: 1 BD: 1 +.+.: nfnl_subsys_ctnetlink_exp FD: 1 BD: 5 +.+.: nf_ct_ecache_mutex FD: 1 BD: 1 +.+.: nfnl_subsys_cttimeout FD: 1 BD: 1 +.+.: nfnl_subsys_cthelper FD: 1 BD: 1 +.+.: nf_ct_helper_mutex FD: 1 BD: 1 +...: nf_conntrack_expect_lock FD: 1 BD: 1 +.+.: nf_ct_nat_helpers_mutex FD: 1 BD: 1 +.+.: nfnl_subsys_nftables FD: 1 BD: 1 +.+.: nfnl_subsys_nftcompat FD: 909 BD: 1 +.+.: masq_mutex ->pernet_ops_rwsem ->(inetaddr_chain).rwsem ->inet6addr_chain.lock FD: 161 BD: 5 +.+.: &xt[i].mutex ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->fs_reclaim ->&c->lock ->&mm->mmap_lock ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->&per_cpu(xt_recseq, i) ->purge_vmap_area_lock ->&n->list_lock ->&____s->seqcount#2 ->init_mm.page_table_lock ->&lock->wait_lock ->rcu_node_0 ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->quarantine_lock ->&rcu_state.expedited_wq ->&meta->lock ->&base->lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 25 BD: 70 +.+.: &tn->lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 3 BD: 1 +.+.: subsys mutex#72 ->&k->k_lock FD: 25 BD: 5 +.+.: nfnl_subsys_ipset ->&rq->__lock FD: 1 BD: 1 +.+.: ip_set_type_mutex FD: 59 BD: 5 +.+.: ipvs->est_mutex ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->&c->lock ->&n->list_lock ->pcpu_lock ->&obj_hash[i].lock ->pcpu_alloc_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 1 +.+.: ip_vs_sched_mutex FD: 50 BD: 5 +.+.: __ip_vs_app_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 1 +.+.: ip_vs_pe_mutex FD: 1 BD: 1 +.+.: tunnel4_mutex FD: 1 BD: 1 +.+.: xfrm4_protocol_mutex FD: 1 BD: 1 +.+.: inet_diag_table_mutex FD: 1 BD: 1 +...: xfrm_km_lock FD: 1 BD: 1 +.+.: xfrm6_protocol_mutex FD: 1 BD: 1 +.+.: tunnel6_mutex FD: 1 BD: 1 +.+.: xfrm_if_cb_lock FD: 1 BD: 1 +...: inetsw6_lock FD: 1 BD: 5 +.+.: &hashinfo->lock#2 FD: 16 BD: 5 +.+.: &net->ipv6.ip6addrlbl_table.lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 133 BD: 56 +.+.: &idev->mc_lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&____s->seqcount ->&dev_addr_list_lock_key ->_xmit_ETHER ->&c->lock ->batched_entropy_u32.lock ->&base->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->krc.lock ->&n->list_lock ->&rq->__lock ->&bridge_netdev_addr_lock_key ->&dev_addr_list_lock_key#2 ->&batadv_netdev_addr_lock_key ->&vlan_netdev_addr_lock_key ->&macvlan_netdev_addr_lock_key ->&dev_addr_list_lock_key#3 ->remove_cache_srcu ->&bridge_netdev_addr_lock_key/1 ->&dev_addr_list_lock_key/1 ->&dev_addr_list_lock_key#2/1 ->_xmit_ETHER/1 ->rcu_node_0 ->&batadv_netdev_addr_lock_key/1 ->&vlan_netdev_addr_lock_key/1 ->&macvlan_netdev_addr_lock_key/1 ->&cfs_rq->removed.lock ->&dev_addr_list_lock_key#3/1 ->&macsec_netdev_addr_lock_key/1 ->&____s->seqcount#2 ->&rcu_state.expedited_wq FD: 8 BD: 57 +...: &dev_addr_list_lock_key ->pool_lock#2 ->&c->lock ->&n->list_lock FD: 42 BD: 69 +...: _xmit_ETHER ->&local->filter_lock ->&rdev->wiphy_work_lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->(console_sem).lock ->console_owner_lock ->console_owner ->&____s->seqcount#2 ->&obj_hash[i].lock ->krc.lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 726 BD: 1 +.+.: (wq_completion)ipv6_addrconf ->(work_completion)(&(&net->ipv6.addr_chk_work)->work) ->(work_completion)(&(&ifa->dad_work)->work) ->&rq->__lock FD: 724 BD: 6 +.+.: (work_completion)(&(&net->ipv6.addr_chk_work)->work) ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock FD: 26 BD: 3 ....: &x->wait#20 ->&p->pi_lock FD: 47 BD: 83 ++--: &ndev->lock ->&ifa->lock ->pool_lock#2 ->&dir->lock#2 ->pcpu_lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&tb->tb6_lock ->&n->list_lock ->batched_entropy_u32.lock ->&base->lock ->&____s->seqcount#2 FD: 3 BD: 1 +.+.: stp_proto_mutex ->llc_sap_list_lock FD: 1 BD: 1 ....: switchdev_notif_chain.lock FD: 25 BD: 49 ++++: (switchdev_blocking_notif_chain).rwsem ->&rq->__lock FD: 726 BD: 1 +.+.: br_ioctl_mutex ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->br_ioctl_mutex.wait_lock ->&rq->__lock ->rcu_state.barrier_mutex ->dev_base_lock ->lweventlist_lock ->stock_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&dir->lock#2 ->netdev_unregistering_wq.lock FD: 227 BD: 8 +.+.: nf_ct_proto_mutex ->defrag4_mutex ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->pool_lock#2 ->defrag6_mutex FD: 160 BD: 5 +.+.: ebt_mutex ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->stock_lock ->&rq->__lock ->ebt_mutex.wait_lock ->&c->lock ->&n->list_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 1 +.+.: dsa_tag_drivers_lock FD: 36 BD: 2 +.+.: drain_vmap_work ->vmap_purge_lock FD: 1 BD: 1 +...: protocol_list_lock FD: 1 BD: 1 +...: linkfail_lock FD: 1 BD: 1 +...: rose_neigh_list_lock FD: 1 BD: 1 +.+.: proto_tab_lock#2 FD: 1 BD: 29 ++++: chan_list_lock FD: 1 BD: 2 +.+.: l2cap_sk_list.lock FD: 3 BD: 1 +.+.: sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP ->slock-AF_BLUETOOTH-BTPROTO_L2CAP ->chan_list_lock FD: 1 BD: 2 +...: slock-AF_BLUETOOTH-BTPROTO_L2CAP FD: 1 BD: 1 ....: rfcomm_wq.lock FD: 1 BD: 1 +.+.: rfcomm_mutex FD: 1 BD: 1 +.+.: auth_domain_lock FD: 1 BD: 1 +.+.: registered_mechs_lock FD: 1 BD: 1 ....: atm_dev_notify_chain.lock FD: 1 BD: 1 +.+.: proto_tab_lock#3 FD: 725 BD: 1 +.+.: vlan_ioctl_mutex ->&mm->mmap_lock ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->vlan_ioctl_mutex.wait_lock ->&rq->__lock FD: 1 BD: 1 +.+.: rds_info_lock FD: 39 BD: 1 +.+.: rds_trans_sem ->(console_sem).lock FD: 1 BD: 1 ....: &id_priv->lock FD: 2 BD: 50 +.+.: &xa->xa_lock#14 ->pool_lock#2 FD: 47 BD: 53 +.+.: k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->&tcp_hashinfo.bhash[i].lock ->&h->lhash2[i].lock ->&table->hash[i].lock ->k-clock-AF_INET6 ->&icsk->icsk_accept_queue.rskq_lock#2 ->&obj_hash[i].lock ->&rq->__lock FD: 27 BD: 55 +...: k-slock-AF_INET6 ->pool_lock#2 ->&obj_hash[i].lock ->&tcp_hashinfo.bhash[i].lock ->krc.lock FD: 1 BD: 75 ++..: k-clock-AF_INET6 FD: 19 BD: 69 +.-.: &tcp_hashinfo.bhash[i].lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&tcp_hashinfo.bhash2[i].lock ->k-clock-AF_INET6 ->clock-AF_INET ->clock-AF_INET6 ->&obj_hash[i].lock ->stock_lock ->&n->list_lock ->&____s->seqcount#2 FD: 17 BD: 70 +.-.: &tcp_hashinfo.bhash2[i].lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->k-clock-AF_INET6 ->clock-AF_INET ->clock-AF_INET6 ->&obj_hash[i].lock ->batched_entropy_u8.lock ->&hashinfo->ehash_locks[i] ->stock_lock ->&____s->seqcount#2 FD: 1 BD: 55 +.+.: &h->lhash2[i].lock FD: 1 BD: 5 +...: &list->lock#4 FD: 1 BD: 6 +...: k-clock-AF_TIPC FD: 33 BD: 5 +.+.: k-sk_lock-AF_TIPC ->k-slock-AF_TIPC ->&tn->nametbl_lock ->&obj_hash[i].lock ->k-clock-AF_TIPC ->&rq->__lock FD: 1 BD: 6 +...: k-slock-AF_TIPC FD: 18 BD: 6 +...: &tn->nametbl_lock ->pool_lock#2 ->&service->lock ->&nt->cluster_scope_lock ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock FD: 16 BD: 7 +...: &service->lock ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->krc.lock FD: 25 BD: 53 +.+.: &pnettable->lock ->&rq->__lock FD: 25 BD: 53 +.+.: smc_ib_devices.mutex ->&rq->__lock FD: 1 BD: 1 +.+.: smc_wr_rx_hash_lock FD: 1 BD: 2 +.+.: v9fs_trans_lock FD: 1 BD: 5 +...: &this->receive_lock FD: 1 BD: 1 +...: lowpan_nhc_lock FD: 234 BD: 7 +.+.: ovs_mutex ->(work_completion)(&data->gc_work) ->nf_ct_proto_mutex ->&obj_hash[i].lock ->pool_lock#2 ->net_rwsem ->quarantine_lock FD: 225 BD: 9 +.+.: defrag4_mutex ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock FD: 225 BD: 9 +.+.: defrag6_mutex ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock FD: 1 BD: 8 +.+.: ima_keys_lock FD: 53 BD: 92 +.+.: scomp_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->&c->lock FD: 6 BD: 578 +.+.: &mm->page_table_lock ->&obj_hash[i].lock ->pool_lock#2 ->stock_lock FD: 33 BD: 578 +.+.: ptlock_ptr(ptdesc)#2 ->lock#4 ->key ->&____s->seqcount ->lock#5 ->&folio_wait_table[i] ->&obj_hash[i].lock FD: 257 BD: 5 +.+.: k-sk_lock-AF_RXRPC ->k-slock-AF_RXRPC ->&rxnet->local_mutex ->&local->services_lock ->fs_reclaim ->pool_lock#2 ->&c->lock ->&rx->incoming_lock ->&obj_hash[i].lock ->&rxnet->conn_lock ->&call->waitq ->(rxrpc_call_limiter).lock ->&rx->recvmsg_lock ->&rx->call_lock ->&rxnet->call_lock ->(&call->timer) ->&base->lock ->&list->lock#17 ->&n->list_lock ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&rq->__lock ->remove_cache_srcu FD: 1 BD: 6 +...: k-slock-AF_RXRPC FD: 246 BD: 6 +.+.: &rxnet->local_mutex ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->crngs.lock ->&c->lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&dir->lock ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->cpu_hotplug_lock ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&x->wait#21 ->&n->list_lock ->&cfs_rq->removed.lock FD: 6 BD: 57 +...: &table->hash[i].lock ->k-clock-AF_INET6 ->&table->hash2[i].lock ->k-clock-AF_INET ->clock-AF_INET ->clock-AF_INET6 FD: 1 BD: 58 +...: &table->hash2[i].lock FD: 228 BD: 2 +.+.: netstamp_work ->cpu_hotplug_lock FD: 26 BD: 7 ....: &x->wait#21 ->&p->pi_lock FD: 1 BD: 6 +.+.: &local->services_lock FD: 1 BD: 8 +.+.: &rxnet->conn_lock FD: 1 BD: 6 ....: &call->waitq FD: 1 BD: 6 +.+.: &rx->call_lock FD: 1 BD: 6 +.+.: &rxnet->call_lock FD: 64 BD: 2 ++++: init_user_ns.keyring_sem ->key_user_lock ->root_key_user.lock ->fs_reclaim ->pool_lock#2 ->crngs.lock ->key_serial_lock ->key_construction_mutex ->&type->lock_class ->keyring_serialise_link_lock ->rcu_node_0 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->init_user_ns.keyring_sem.wait_lock FD: 1 BD: 10 +.+.: root_key_user.lock FD: 1 BD: 9 +.+.: keyring_name_lock FD: 1 BD: 1 +.+.: template_list FD: 1 BD: 1 +.+.: idr_lock FD: 30 BD: 5 +.-.: (&rxnet->peer_keepalive_timer) FD: 50 BD: 1 +.+.: ima_extend_list_mutex ->fs_reclaim ->pool_lock#2 FD: 16 BD: 5 +.+.: (wq_completion)krxrpcd ->(work_completion)(&rxnet->peer_keepalive_work) ->(work_completion)(&rxnet->service_conn_reaper) FD: 13 BD: 6 +.+.: (work_completion)(&rxnet->peer_keepalive_work) ->&rxnet->peer_hash_lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 7 +.+.: &rxnet->peer_hash_lock FD: 26 BD: 4 +.+.: deferred_probe_work ->deferred_probe_mutex FD: 25 BD: 49 +.+.: &(&net->nexthop.notifier_chain)->rwsem ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 40 BD: 51 +.+.: k-sk_lock-AF_INET ->k-slock-AF_INET#2 ->&table->hash[i].lock ->&obj_hash[i].lock ->k-clock-AF_INET ->&rq->__lock FD: 22 BD: 52 +...: k-slock-AF_INET#2 ->&____s->seqcount ->pool_lock#2 ->batched_entropy_u32.lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount#2 ->&n->list_lock FD: 1 BD: 58 ++..: k-clock-AF_INET FD: 1 BD: 1 ....: power_off_handler_list.lock FD: 724 BD: 2 +.+.: reg_work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock FD: 1 BD: 49 +...: reg_pending_beacons_lock FD: 736 BD: 2 +.+.: (work_completion)(&fw_work->work) ->fs_reclaim ->pool_lock#2 ->&fw_cache.lock ->tk_core.seq.seqcount ->async_lock ->init_task.alloc_lock ->&dentry->d_lock ->&sb->s_type->i_mutex_key ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->(console_sem).lock ->&rq->__lock ->umhelper_sem ->fw_lock ->rtnl_mutex FD: 2 BD: 3 +.+.: &fw_cache.lock ->pool_lock#2 FD: 2 BD: 422 +.+.: &____s->seqcount#6 ->&____s->seqcount#6/1 FD: 1 BD: 1 +.+.: detector_work FD: 1 BD: 1 +.+.: acpi_gpio_deferred_req_irqs_lock FD: 3 BD: 4 +.+.: subsys mutex#73 ->&k->k_lock FD: 2 BD: 18 +.+.: fw_lock ->&x->wait#22 FD: 1 BD: 19 ....: &x->wait#22 FD: 1 BD: 1 ....: enable_lock FD: 1 BD: 1 +.+.: gpd_list_lock FD: 1 BD: 1 +.+.: cdev_lock FD: 302 BD: 3 +.+.: &tty->legacy_mutex ->&tty->read_wait ->&tty->write_wait ->&tty->ldisc_sem ->&tty->files_lock ->&port->lock ->&port->mutex ->&port_lock_key ->&f->f_lock ->&obj_hash[i].lock ->pool_lock#2 ->tasklist_lock ->&port->open_wait ->fs_reclaim ->&c->lock ->stock_lock ->tty_ldiscs_lock ->&k->list_lock ->&rq->__lock ->&k->k_lock ->&tty->ctrl.lock ->redirect_lock ->&tty->legacy_mutex/1 ->tty_mutex.wait_lock ->&p->pi_lock ->&n->list_lock ->(console_sem).lock ->console_lock ->&lock->wait_lock ->&cfs_rq->removed.lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 FD: 1 BD: 11 ....: &tty->read_wait FD: 26 BD: 164 ....: &tty->write_wait ->&p->pi_lock FD: 252 BD: 5 ++++: &tty->ldisc_sem ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&tty->write_wait ->&tty->read_wait ->&tty->termios_rwsem ->&mm->mmap_lock ->&port_lock_key ->&port->lock ->&tty->flow.lock ->&ldata->atomic_read_lock ->&buf->lock ->tty_ldiscs_lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->init_mm.page_table_lock ->&n->list_lock ->&rq->__lock ->&o_tty->termios_rwsem/1 ->&tty->ldisc_sem/1 ->rcu_node_0 ->remove_cache_srcu ->&rcu_state.expedited_wq ->batched_entropy_u8.lock ->kfence_freelist_lock ->&cfs_rq->removed.lock FD: 163 BD: 10 ++++: &tty->termios_rwsem ->&port->mutex ->&tty->write_wait ->&tty->read_wait ->&ldata->output_lock ->&port_lock_key ->vmap_area_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->pool_lock#2 ->&rq->__lock ->rcu_node_0 ->&meta->lock ->kfence_freelist_lock ->&____s->seqcount FD: 7 BD: 9 +.+.: &tty->files_lock ->&f->f_lock FD: 1 BD: 164 ....: &port->lock FD: 26 BD: 420 ....: &wq#2 ->&p->pi_lock FD: 165 BD: 1 +.+.: &type->s_umount_key#24/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->inode_hash_lock ->bdev_lock ->&disk->open_mutex ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&wq->mutex ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->wq_pool_mutex ->mmu_notifier_invalidate_range_start ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&xa->xa_lock#8 ->lock#4 ->&mapping->i_private_lock ->tk_core.seq.seqcount ->bit_wait_table + i ->wq_mayday_lock ->&sbi->old_work_lock ->(work_completion)(&(&sbi->old_work)->work) ->&x->wait#23 FD: 1 BD: 11 +.+.: &bdev->bd_holder_lock FD: 18 BD: 2 +.+.: (work_completion)(&s->destroy_work) ->&obj_hash[i].lock ->pool_lock#2 ->&rsp->gp_wait ->pcpu_lock ->&base->lock ->quarantine_lock ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 2 +.+.: &sbi->old_work_lock FD: 1 BD: 2 +.+.: (work_completion)(&(&sbi->old_work)->work) FD: 1 BD: 21 ....: &x->wait#23 FD: 11 BD: 317 ....: &xa->xa_lock#15 ->pool_lock#2 ->&____s->seqcount ->&c->lock ->&obj_hash[i].lock ->&n->list_lock FD: 163 BD: 1 +.+.: &type->s_umount_key#25/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->inode_hash_lock ->bdev_lock ->&disk->open_mutex ->&c->lock ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->&xa->xa_lock#8 ->lock#4 ->&mapping->i_private_lock ->tk_core.seq.seqcount ->bit_wait_table + i ->&rq->__lock ->&obj_hash[i].lock ->&sb->s_type->i_lock_key#3 ->lock#5 ->&lruvec->lru_lock ->crypto_alg_sem ->lock#2 ->&x->wait#23 FD: 333 BD: 1 +.+.: &type->s_umount_key#26/1 ->fs_reclaim ->pcpu_alloc_mutex ->&c->lock ->&____s->seqcount ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->inode_hash_lock ->bdev_lock ->&disk->open_mutex ->mmu_notifier_invalidate_range_start ->&xa->xa_lock#8 ->lock#4 ->&mapping->i_private_lock ->tk_core.seq.seqcount ->bit_wait_table + i ->&rq->__lock ->&obj_hash[i].lock ->&sb->s_type->i_lock_key#3 ->lock#5 ->&lruvec->lru_lock ->crypto_alg_sem ->pool_lock#2 ->percpu_counters_lock ->&sb->s_type->i_lock_key#22 ->&sb->s_type->i_mutex_key#8 ->proc_subdir_lock ->proc_inum_ida.xa_lock ->&journal->j_state_lock ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&journal->j_wait_done_commit ->&p->alloc_lock ->cpu_hotplug_lock ->wq_pool_mutex ->&ei->i_es_lock ->ext4_grpinfo_slab_create_mutex ->&s->s_inode_list_lock ->ext4_li_mtx ->lock ->&root->kernfs_rwsem ->(console_sem).lock ->&dentry->d_lock FD: 31 BD: 99 +.+.: &bgl->locks[i].lock ->&sbi->s_md_lock ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->&ei->i_prealloc_lock FD: 48 BD: 267 +.+.: &sb->s_type->i_lock_key#22 ->&dentry->d_lock ->&lru->node[i].lock ->bit_wait_table + i ->&xa->xa_lock#8 FD: 180 BD: 5 ++++: &sb->s_type->i_mutex_key#8 ->&ei->i_es_lock ->&ei->i_data_sem ->&ei->xattr_sem ->tk_core.seq.seqcount ->&mm->mmap_lock ->fs_reclaim ->&____s->seqcount ->&xa->xa_lock#8 ->lock#4 ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&mapping->i_private_lock ->&sb->s_type->i_lock_key#22 ->&wb->list_lock ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->&c->lock ->&rq->__lock ->rcu_node_0 ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->mapping.invalidate_lock ->free_vmap_area_lock ->vmap_area_lock ->init_mm.page_table_lock ->pcpu_alloc_mutex ->batched_entropy_u32.lock ->swap_cgroup_mutex ->&fq->mq_flush_lock ->&x->wait#26 ->&base->lock ->key ->pcpu_lock ->percpu_counters_lock ->(&timer.timer) ->&((cluster_info + ci)->lock)/1 ->swapon_mutex ->proc_poll_wait.lock ->&dentry->d_lock ->&n->list_lock ->&____s->seqcount#2 ->&meta->lock ->quarantine_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->bit_wait_table + i ->stock_lock FD: 21 BD: 98 ++++: &ei->i_es_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&sbi->s_es_lock ->&obj_hash[i].lock ->key#2 ->key#6 ->key#7 ->key#8 ->&____s->seqcount#2 ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->&base->lock FD: 92 BD: 97 ++++: &ei->i_data_sem ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&ei->i_es_lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&ei->i_prealloc_lock ->&sb->s_type->i_lock_key#22 ->&(ei->i_block_reservation_lock) ->&lg->lg_mutex ->&mapping->i_private_lock ->&ei->i_raw_lock ->&rq->__lock ->&wb->list_lock ->lock#4 ->&ret->b_state_lock ->&journal->j_revoke_lock ->key#15 ->&sbi->s_md_lock ->key#3 ->rcu_node_0 ->&____s->seqcount#2 ->&n->list_lock ->&bgl->locks[i].lock ->remove_cache_srcu ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->key ->pcpu_lock ->percpu_counters_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->stock_lock ->bit_wait_table + i ->&base->lock ->&wb->work_lock ->&pa->pa_lock#2 ->&xa->xa_lock#8 FD: 1 BD: 99 +.+.: &sbi->s_es_lock FD: 66 BD: 98 ++++: &journal->j_state_lock ->&journal->j_wait_done_commit ->&journal->j_wait_commit ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&base->lock ->&journal->j_wait_updates ->&journal->j_wait_transaction_locked ->&journal->j_list_lock ->&journal->j_wait_reserved FD: 12 BD: 5 +.-.: (&net->can.stattimer) ->&obj_hash[i].lock ->&base->lock FD: 26 BD: 99 ....: &journal->j_wait_done_commit ->&p->pi_lock FD: 26 BD: 99 ....: &journal->j_wait_commit ->&p->pi_lock FD: 106 BD: 2 +.+.: ext4_grpinfo_slab_create_mutex ->slab_mutex FD: 54 BD: 3 +.+.: ext4_li_mtx ->fs_reclaim ->pool_lock#2 ->batched_entropy_u16.lock ->&eli->li_list_mtx ->kthread_create_lock ->&p->pi_lock ->&x->wait ->&rq->__lock ->&obj_hash[i].lock FD: 1 BD: 1 ....: &rs->lock FD: 164 BD: 5 ++++: &type->i_mutex_dir_key#3 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->&ei->i_es_lock ->&ei->i_data_sem ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#8 ->lock#4 ->&mapping->i_private_lock ->tk_core.seq.seqcount ->bit_wait_table + i ->&rq->__lock ->inode_hash_lock ->&obj_hash[i].lock ->&journal->j_state_lock ->&sb->s_type->i_lock_key#22 ->&ei->xattr_sem ->namespace_sem ->&c->lock ->tomoyo_ss ->&s->s_inode_list_lock ->jbd2_handle ->&mm->mmap_lock ->&n->list_lock ->&____s->seqcount#2 ->quarantine_lock ->remove_cache_srcu ->&cfs_rq->removed.lock ->rcu_node_0 ->&sem->wait_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&base->lock FD: 4 BD: 7 ..-.: &rsp->gp_wait ->&obj_hash[i].lock ->pool_lock#2 FD: 75 BD: 97 ++++: &ei->xattr_sem ->mmu_notifier_invalidate_range_start ->lock#4 ->&mapping->i_private_lock ->pool_lock#2 ->&ret->b_state_lock ->&journal->j_revoke_lock ->tk_core.seq.seqcount ->&ei->i_raw_lock ->&____s->seqcount ->&xa->xa_lock#8 ->&rq->__lock ->stock_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->bit_wait_table + i ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock FD: 39 BD: 51 +.+.: rcu_state.barrier_mutex ->rcu_state.barrier_lock ->&x->wait#24 ->&rq->__lock ->rcu_state.barrier_mutex.wait_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->key ->pcpu_lock ->percpu_counters_lock ->&____s->seqcount ->stock_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->pool_lock FD: 26 BD: 52 ..-.: &x->wait#24 ->&p->pi_lock FD: 138 BD: 85 ++++: &vma->vm_lock->lock ->ptlock_ptr(ptdesc)#2 ->fs_reclaim ->&____s->seqcount ->mmu_notifier_invalidate_range_start ->mapping.invalidate_lock ->pool_lock#2 ->rcu_node_0 ->&rq->__lock ->&rcu_state.gp_wq ->&cfs_rq->removed.lock ->&c->lock ->remove_cache_srcu ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&mm->page_table_lock ->&lruvec->lru_lock ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rcu_state.expedited_wq ->stock_lock ->&xa->xa_lock#8 ->&sb->s_type->i_lock_key ->&info->lock ->lock#4 ->tk_core.seq.seqcount ->mount_lock ->&____s->seqcount#2 ->&n->list_lock ->key ->pcpu_lock ->percpu_counters_lock ->&base->lock FD: 46 BD: 577 ++++: &anon_vma->rwsem ->&mm->page_table_lock ->&obj_hash[i].lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&rq->__lock ->&sem->wait_lock ->quarantine_lock ->key ->pcpu_lock ->percpu_counters_lock ->ptlock_ptr(ptdesc)#2 ->&n->list_lock ->mmu_notifier_invalidate_range_start ->&meta->lock ->kfence_freelist_lock ->stock_lock ->&____s->seqcount#2 ->rcu_node_0 ->&cfs_rq->removed.lock ->batched_entropy_u8.lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&rcu_state.expedited_wq ->&base->lock FD: 192 BD: 1 +.+.: &sig->cred_guard_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->init_fs.lock ->&p->pi_lock ->mapping.invalidate_lock ->&folio_wait_table[i] ->&rq->__lock ->tomoyo_ss ->binfmt_lock ->init_binfmt_misc.entries_lock ->&dentry->d_lock ->&type->i_mutex_dir_key#3 ->&sb->s_type->i_lock_key#22 ->&obj_hash[i].lock ->&ei->xattr_sem ->&tsk->futex_exit_mutex ->&sig->exec_update_lock ->&fs->lock ->lock#4 ->&sb->s_type->i_mutex_key#8 ->&p->alloc_lock ->tk_core.seq.seqcount ->&mm->mmap_lock ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&n->list_lock ->&stopper->lock ->&stop_pi_lock ->&x->wait#8 ->remove_cache_srcu ->rcu_node_0 ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq ->key#5 FD: 131 BD: 87 ++++: mapping.invalidate_lock ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#8 ->lock#4 ->&ei->i_es_lock ->&ei->i_data_sem ->pool_lock#2 ->tk_core.seq.seqcount ->&c->lock ->&folio_wait_table[i] ->&rq->__lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&n->list_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->&mapping->i_mmap_rwsem ->&journal->j_state_lock ->jbd2_handle FD: 1 BD: 4 ++++: init_binfmt_misc.entries_lock FD: 170 BD: 2 +.+.: &sig->exec_update_lock ->&p->alloc_lock ->&sighand->siglock ->&newf->file_lock ->batched_entropy_u64.lock ->&mm->mmap_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock#2 ->pool_lock ->&rq->__lock ->&cfs_rq->removed.lock ->stock_lock ->rcu_node_0 ->quarantine_lock FD: 3 BD: 12 ..-.: batched_entropy_u16.lock ->crngs.lock FD: 1 BD: 579 +.+.: ptlock_ptr(ptdesc)#2/1 FD: 30 BD: 1 ..-.: &(&ipvs->defense_work)->timer FD: 30 BD: 1 ..-.: &(&gc_work->dwork)->timer FD: 37 BD: 2 +.+.: (work_completion)(&(&gc_work->dwork)->work) ->rcu_node_0 ->&rq->__lock ->&obj_hash[i].lock ->&base->lock ->&cfs_rq->removed.lock ->pool_lock#2 FD: 32 BD: 6 +.+.: (work_completion)(&(&ipvs->defense_work)->work) ->&s->s_inode_list_lock ->&ipvs->dropentry_lock ->&ipvs->droppacket_lock ->&ipvs->securetcp_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 146 ...-: &____s->seqcount#7 FD: 1 BD: 7 +...: &ipvs->dropentry_lock FD: 1 BD: 7 +...: &ipvs->droppacket_lock FD: 1 BD: 7 +...: &ipvs->securetcp_lock FD: 1 BD: 99 ....: key#2 FD: 1 BD: 19 ..-.: task_group_lock FD: 97 BD: 2 +.+.: &type->s_umount_key#27/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->&c->lock ->&____s->seqcount ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#23 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&sb->s_type->i_mutex_key#9 ->&dentry->d_lock ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->&n->list_lock ->&rq->__lock FD: 39 BD: 122 +.+.: &sb->s_type->i_lock_key#23 ->&dentry->d_lock ->&lru->node[i].lock ->bit_wait_table + i ->&p->pi_lock FD: 89 BD: 5 ++++: &sb->s_type->i_mutex_key#9 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#23 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->rename_lock.seqcount ->proc_subdir_lock ->&p->alloc_lock ->&pid->lock ->&c->lock ->&____s->seqcount ->sysctl_lock ->namespace_sem ->tomoyo_ss ->&rq->__lock ->&n->list_lock ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->&____s->seqcount#2 ->rcu_node_0 ->&cfs_rq->removed.lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rcu_state.gp_wq ->&rcu_state.expedited_wq ->key ->pcpu_lock ->percpu_counters_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 766 BD: 2 +.+.: &p->lock ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&c->lock ->&____s->seqcount ->file_systems_lock ->namespace_sem ->&of->mutex ->&n->list_lock ->remove_cache_srcu ->rcu_node_0 ->&rq->__lock ->module_mutex ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq FD: 97 BD: 2 +.+.: &type->s_umount_key#28/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#24 ->crngs.lock ->&root->kernfs_supers_rwsem ->&dentry->d_lock ->&c->lock ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock FD: 38 BD: 265 +.+.: &sb->s_type->i_lock_key#24 ->&dentry->d_lock ->&lru->node[i].lock FD: 170 BD: 3 ++++: &type->i_mutex_dir_key#4 ->fs_reclaim ->&c->lock ->&____s->seqcount ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->&root->kernfs_rwsem ->mmu_notifier_invalidate_range_start ->iattr_mutex ->&obj_hash[i].lock ->&sb->s_type->i_lock_key#24 ->namespace_sem ->&mm->mmap_lock ->vmap_area_lock ->tk_core.seq.seqcount ->&n->list_lock ->rcu_node_0 ->&rq->__lock ->quarantine_lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rcu_state.gp_wq ->rename_lock ->&sem->wait_lock ->&p->pi_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq ->&base->lock ->&meta->lock FD: 50 BD: 203 +.+.: iattr_mutex ->&rq->__lock ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->tk_core.seq.seqcount FD: 70 BD: 1 ++++: &type->s_umount_key#29 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->pool_lock#2 ->&dentry->d_lock ->rename_lock.seqcount ->&sb->s_type->i_lock_key#23 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 ->&rq->__lock FD: 26 BD: 135 ....: &x->wait#25 ->&p->pi_lock FD: 39 BD: 10 +.+.: &net->unx.table.locks[i] ->&net->unx.table.locks[i]/1 FD: 856 BD: 1 +.+.: &sb->s_type->i_mutex_key#10 ->&net->unx.table.locks[i] ->&u->lock ->&u->peer_wait ->rlock-AF_UNIX ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->sk_lock-AF_INET ->slock-AF_INET#2 ->clock-AF_INET ->nl_table_lock ->nl_table_wait.lock ->clock-AF_NETLINK ->&nlk->wait ->(netlink_chain).rwsem ->sk_lock-AF_INET6 ->slock-AF_INET6 ->clock-AF_INET6 ->&table->hash[i].lock ->&net->packet.sklist_lock ->&po->bind_lock ->sk_lock-AF_PACKET ->slock-AF_PACKET ->fanout_mutex ->&x->wait#2 ->&rq->__lock ->clock-AF_PACKET ->pcpu_lock ->key ->percpu_counters_lock ->rcu_node_0 ->&____s->seqcount ->cb_lock ->genl_sk_destructing_waitq.lock ->sk_lock-AF_BLUETOOTH-BTPROTO_HCI ->slock-AF_BLUETOOTH-BTPROTO_HCI ->hci_dev_list_lock ->&cfs_rq->removed.lock ->quarantine_lock ->pool_lock ->stock_lock ->sk_lock-AF_CAN ->slock-AF_CAN ->sk_lock-AF_X25 ->slock-AF_X25 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->clock-AF_NFC ->&rcu_state.expedited_wq ->&base->lock ->&rnp->exp_lock ->&rnp->exp_wq[1] ->&rnp->exp_wq[2] ->&rnp->exp_wq[3] ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->&meta->lock ->kfence_freelist_lock ->&net->ipv4.ra_mutex ->&hashinfo->lock ->(work_completion)(&msk->work) ->sk_lock-AF_BLUETOOTH-BTPROTO_SCO ->slock-AF_BLUETOOTH-BTPROTO_SCO ->clock-AF_BLUETOOTH ->sco_sk_list.lock ->sk_lock-AF_KCM ->slock-AF_KCM ->&mux->lock ->(work_completion)(&kcm->tx_work) ->&mux->rx_lock ->&knet->mutex ->sk_lock-AF_NFC ->slock-AF_NFC ->&bsd_socket_locks[i] ->rcu_state.exp_mutex ->sk_lock-AF_QIPCRTR ->slock-AF_QIPCRTR ->sk_lock-AF_INET/1 ->&net->sctp.addr_wq_lock ->unix_gc_lock ->rlock-AF_PACKET ->sk_lock-AF_AX25 ->slock-AF_AX25 ->clock-AF_NETROM ->sk_lock-AF_NETROM ->slock-AF_NETROM ->clock-AF_ROSE ->sk_lock-AF_ROSE ->slock-AF_ROSE ->sk_lock-AF_PPPOX ->slock-AF_PPPOX FD: 47 BD: 3 +.+.: &u->lock ->clock-AF_UNIX ->&u->lock/1 ->rlock-AF_UNIX ->&u->peer_wait ->&sk->sk_peer_lock ->&ei->socket.wq.wait FD: 1 BD: 4 +...: clock-AF_UNIX FD: 26 BD: 4 +.+.: &u->peer_wait ->&p->pi_lock FD: 1 BD: 5 +.+.: rlock-AF_UNIX FD: 187 BD: 1 .+.+: sb_writers#3 ->mount_lock ->tk_core.seq.seqcount ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->&sb->s_type->i_lock_key#22 ->&wb->list_lock ->&wb->work_lock ->&type->i_mutex_dir_key#3 ->&type->i_mutex_dir_key#3/1 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&zone->lock ->remove_cache_srcu ->&xa->xa_lock#8 ->lock#4 ->&mapping->i_private_lock ->&base->lock ->bit_wait_table + i ->&rq->__lock ->&n->list_lock ->&sb->s_type->i_mutex_key#8 ->tomoyo_ss ->&s->s_inode_list_lock ->sb_internal ->inode_hash_lock ->&fsnotify_mark_srcu ->quarantine_lock ->&____s->seqcount#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&sem->wait_lock ->&p->pi_lock ->key ->pcpu_lock ->percpu_counters_lock ->stock_lock ->fs_reclaim ->&mm->mmap_lock ->&p->alloc_lock ->&f->f_lock ->lock#5 ->&lruvec->lru_lock ->&folio_wait_table[i] FD: 168 BD: 2 .+.+: sb_writers#4 ->mount_lock ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#23 ->&wb->list_lock ->&sb->s_type->i_mutex_key#9 ->sysctl_lock ->&dentry->d_lock ->tomoyo_ss ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&obj_hash[i].lock ->oom_adj_mutex ->&p->pi_lock ->&c->lock ->&____s->seqcount#10 ->&(&net->ipv4.ping_group_range.lock)->lock ->oom_adj_mutex.wait_lock ->rcu_node_0 ->&rq->__lock ->&n->list_lock ->remove_cache_srcu ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock ->stock_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 1 BD: 6 +.+.: &pid->lock FD: 176 BD: 1 ++++: &type->s_umount_key#30 ->&lru->node[i].lock ->&dentry->d_lock ->&sb->s_type->i_lock_key#22 ->&sbi->s_writepages_rwsem ->&sem->waiters ->&rsp->gp_wait ->&journal->j_state_lock ->&p->alloc_lock ->(work_completion)(&sbi->s_sb_upd_work) ->key#3 ->key#4 ->&sbi->s_error_lock ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&base->lock ->bit_wait_table + i ->&rq->__lock ->ext4_li_mtx ->(console_sem).lock ->mount_lock ->&____s->seqcount ->&xa->xa_lock#8 ->lock#4 ->&eli->li_list_mtx ->jbd2_handle ->&wb->list_lock ->(wq_completion)ext4-rsv-conversion ->&wq->mutex ->&journal->j_wait_commit ->&journal->j_wait_done_commit ->&sb->s_type->i_lock_key#3 ->&c->lock ->lock#5 ->&rq_wait->wait ->&__ctx->lock ->rcu_node_0 ->&____s->seqcount#2 ->&folio_wait_table[i] ->&fq->mq_flush_lock ->&x->wait#26 ->&n->list_lock ->key#25 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&cfs_rq->removed.lock ->(&timer.timer) ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq FD: 126 BD: 4 ++++: &sbi->s_writepages_rwsem ->&rsp->gp_wait ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&c->lock ->remove_cache_srcu ->pool_lock#2 ->lock#4 ->lock#5 ->&journal->j_state_lock ->jbd2_handle ->tk_core.seq.seqcount ->&xa->xa_lock#8 ->&base->lock ->&rq_wait->wait ->&__ctx->lock ->rcu_node_0 ->&____s->seqcount#2 FD: 1 BD: 2 ....: &sem->waiters FD: 125 BD: 3 +.+.: (work_completion)(&sbi->s_sb_upd_work) ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock FD: 1 BD: 98 ....: key#3 FD: 1 BD: 97 ....: key#4 FD: 1 BD: 97 +.+.: &sbi->s_error_lock FD: 4 BD: 4 +.+.: &eli->li_list_mtx ->&obj_hash[i].lock ->pool_lock#2 FD: 124 BD: 96 ++++: jbd2_handle ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&ret->b_state_lock ->&journal->j_revoke_lock ->&ei->i_raw_lock ->&journal->j_wait_updates ->&mapping->i_private_lock ->&meta_group_info[i]->alloc_sem ->tk_core.seq.seqcount ->inode_hash_lock ->batched_entropy_u32.lock ->&ei->xattr_sem ->&obj_hash[i].lock ->&ei->i_es_lock ->&dentry->d_lock ->smack_known_lock ->&sb->s_type->i_lock_key#22 ->&journal->j_state_lock ->bit_wait_table + i ->&rq->__lock ->lock#4 ->lock#5 ->&ei->i_data_sem ->&xa->xa_lock#8 ->&sbi->s_orphan_lock ->&journal->j_list_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->key#3 ->key#4 ->&sbi->s_error_lock ->&n->list_lock ->stock_lock ->&____s->seqcount#2 ->rcu_node_0 ->quarantine_lock ->remove_cache_srcu ->&cfs_rq->removed.lock ->&bgl->locks[i].lock ->smack_known_lock.wait_lock ->&p->pi_lock ->&rcu_state.expedited_wq ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&lock->wait_lock ->&folio_wait_table[i] ->key ->pcpu_lock ->percpu_counters_lock ->&base->lock ->&ei->i_prealloc_lock FD: 61 BD: 101 +.+.: &ret->b_state_lock ->&journal->j_list_lock ->bit_wait_table + i ->&obj_hash[i].lock FD: 60 BD: 104 +.+.: &journal->j_list_lock ->&sb->s_type->i_lock_key#3 ->&wb->list_lock ->&wb->work_lock ->key#13 ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->quarantine_lock FD: 1 BD: 100 +.+.: &journal->j_revoke_lock FD: 1 BD: 100 +.+.: &ei->i_raw_lock FD: 26 BD: 99 ....: &journal->j_wait_updates ->&p->pi_lock FD: 12 BD: 300 -.-.: &wb->work_lock ->&obj_hash[i].lock ->&base->lock FD: 52 BD: 97 ++++: &meta_group_info[i]->alloc_sem ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&base->lock ->&c->lock ->&____s->seqcount ->&x->wait#26 ->&__ctx->lock ->rcu_node_0 ->&rq->__lock ->(&timer.timer) ->&fq->mq_flush_lock ->&bgl->locks[i].lock ->&cfs_rq->removed.lock FD: 129 BD: 3 .+.+: sb_internal ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->&rq->__lock ->&c->lock ->&n->list_lock ->remove_cache_srcu ->&____s->seqcount#2 ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&base->lock ->rcu_node_0 ->&cfs_rq->removed.lock ->quarantine_lock ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 2 BD: 101 ++++: &ei->i_prealloc_lock ->&pa->pa_lock#2 FD: 28 BD: 1 .+.+: file_rwsem ->&ctx->flc_lock ->&rq->__lock ->rcu_node_0 FD: 2 BD: 2 +.+.: &ctx->flc_lock ->&fll->lock FD: 1 BD: 3 +.+.: &fll->lock FD: 161 BD: 2 +.+.: &type->i_mutex_dir_key#3/1 ->rename_lock.seqcount ->&dentry->d_lock ->fs_reclaim ->&ei->i_es_lock ->&ei->i_data_sem ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#8 ->lock#4 ->pool_lock#2 ->&mapping->i_private_lock ->tk_core.seq.seqcount ->bit_wait_table + i ->&rq->__lock ->inode_hash_lock ->&obj_hash[i].lock ->&journal->j_state_lock ->&sb->s_type->i_lock_key#22 ->&ei->xattr_sem ->&c->lock ->tomoyo_ss ->&s->s_inode_list_lock ->jbd2_handle ->&sb->s_type->i_mutex_key#8 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&xa->xa_lock#15 ->stock_lock ->&____s->seqcount#2 ->&fsnotify_mark_srcu ->&type->i_mutex_dir_key#3 ->&wb->list_lock ->sb_internal ->remove_cache_srcu ->&n->list_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->quarantine_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&meta->lock ->key ->pcpu_lock ->percpu_counters_lock ->&u->bindlock ->&base->lock FD: 77 BD: 1 +.+.: &type->s_umount_key#31/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#25 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&sb->s_type->i_mutex_key#11 ->&dentry->d_lock FD: 38 BD: 3 +.+.: &sb->s_type->i_lock_key#25 ->&dentry->d_lock FD: 63 BD: 2 +.+.: &sb->s_type->i_mutex_key#11 ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#25 ->&s->s_inode_list_lock ->tk_core.seq.seqcount FD: 63 BD: 2 ++++: &type->s_umount_key#32 ->sb_lock ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->&lru->node[i].lock ->&obj_hash[i].lock ->&rq->__lock ->stock_lock ->&c->lock FD: 40 BD: 2 ++++: &type->s_umount_key#33 ->sb_lock ->&dentry->d_lock FD: 40 BD: 1 ++++: &type->s_umount_key#34 ->sb_lock ->&dentry->d_lock ->&rq->__lock FD: 76 BD: 1 +.+.: &type->s_umount_key#35/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#26 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock FD: 38 BD: 4 +.+.: &sb->s_type->i_lock_key#26 ->&dentry->d_lock FD: 40 BD: 1 ++++: &type->s_umount_key#36 ->sb_lock ->&dentry->d_lock FD: 1 BD: 4 +.+.: redirect_lock FD: 245 BD: 1 +.+.: &tty->atomic_write_lock ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&tty->termios_rwsem ->&tty->files_lock FD: 36 BD: 11 +.+.: &ldata->output_lock ->&port_lock_key ->&rq->__lock FD: 77 BD: 1 +.+.: &type->s_umount_key#37/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#27 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->fuse_mutex ->&dentry->d_lock FD: 38 BD: 2 +.+.: &sb->s_type->i_lock_key#27 ->&dentry->d_lock FD: 1 BD: 2 +.+.: fuse_mutex FD: 82 BD: 1 +.+.: &type->s_umount_key#38/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->&____s->seqcount ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#28 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->pstore_sb_lock ->&sb->s_type->i_mutex_key#12 ->&dentry->d_lock FD: 38 BD: 2 +.+.: &sb->s_type->i_lock_key#28 ->&dentry->d_lock FD: 57 BD: 4 +.+.: &sb->s_type->i_mutex_key#12 ->fs_reclaim ->&zone->lock ->&____s->seqcount ->&psinfo->read_mutex ->&obj_hash[i].lock FD: 56 BD: 5 +.+.: &psinfo->read_mutex ->(efivars_lock).lock ->fs_reclaim ->pool_lock#2 ->(efi_runtime_lock).lock ->&obj_hash[i].lock ->&x->wait#12 ->&rq->__lock FD: 82 BD: 2 +.+.: &type->s_umount_key#39/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#29 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->bpf_preload_lock ->&dentry->d_lock ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->&rq->__lock ->&c->lock ->bpf_preload_lock.wait_lock ->&p->pi_lock ->&n->list_lock FD: 38 BD: 4 +.+.: &sb->s_type->i_lock_key#29 ->&dentry->d_lock FD: 57 BD: 3 +.+.: bpf_preload_lock ->(kmod_concurrent_max).lock ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&x->wait#17 ->&rq->__lock ->key ->pcpu_lock ->percpu_counters_lock ->running_helpers_waitq.lock ->&c->lock ->bpf_preload_lock.wait_lock ->stock_lock ->&cfs_rq->removed.lock ->&n->list_lock FD: 26 BD: 1 ++++: uts_sem ->hostname_poll.wait.lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 13 BD: 11 +.-.: (&cb->timer) ->&obj_hash[i].lock ->&base->lock ->tk_core.seq.seqcount FD: 159 BD: 3 ++++: &type->i_mutex_dir_key#5 ->fs_reclaim ->&dentry->d_lock ->rename_lock.seqcount ->tomoyo_ss ->&sbinfo->stat_lock ->&____s->seqcount ->pool_lock#2 ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&xattrs->lock ->&obj_hash[i].lock ->&simple_offset_xa_lock ->smack_known_lock ->&c->lock ->&n->list_lock ->remove_cache_srcu ->&sem->wait_lock ->&rq->__lock ->rcu_node_0 ->&p->pi_lock ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&mm->mmap_lock ->vmap_area_lock ->&____s->seqcount#2 ->smack_known_lock.wait_lock ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&meta->lock ->&rcu_state.gp_wq ->&base->lock FD: 171 BD: 2 .+.+: sb_writers#5 ->mount_lock ->&type->i_mutex_dir_key#5 ->&type->i_mutex_dir_key#5/1 ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key ->&wb->list_lock ->&sb->s_type->i_mutex_key#13 ->&rq->__lock ->&sem->wait_lock ->&p->pi_lock ->&s->s_inode_list_lock ->&info->lock ->&obj_hash[i].lock ->pool_lock#2 ->&sbinfo->stat_lock ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->tomoyo_ss ->&xattrs->lock ->fs_reclaim ->&c->lock ->lock#4 ->lock#5 ->&lruvec->lru_lock ->&dentry->d_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->quarantine_lock ->&base->lock FD: 96 BD: 3 +.+.: &type->i_mutex_dir_key#5/1 ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->tomoyo_ss ->&sbinfo->stat_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->batched_entropy_u32.lock ->&xattrs->lock ->&obj_hash[i].lock ->&simple_offset_xa_lock ->smack_known_lock ->&u->bindlock ->pool_lock#2 ->&c->lock ->&sb->s_type->i_mutex_key#13/4 ->&____s->seqcount ->&sem->wait_lock ->&rq->__lock ->&n->list_lock ->&sb->s_type->i_mutex_key#13 ->&fsnotify_mark_srcu ->lock#4 ->lock#5 ->&lruvec->lru_lock ->&info->lock ->&xa->xa_lock#8 ->remove_cache_srcu ->&____s->seqcount#2 ->rcu_node_0 ->&cfs_rq->removed.lock ->smack_known_lock.wait_lock ->&p->pi_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rcu_state.expedited_wq ->&meta->lock ->key#9 FD: 6 BD: 63 +.+.: &f->f_lock ->fasync_lock FD: 1 BD: 2 ....: hostname_poll.wait.lock FD: 784 BD: 1 +.+.: &f->f_pos_lock ->&type->i_mutex_dir_key#3 ->&mm->mmap_lock ->&type->i_mutex_dir_key#4 ->sb_writers#5 ->&type->i_mutex_dir_key#5 ->&p->lock ->sysctl_lock ->fs_reclaim ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->&rq->__lock ->&cfs_rq->removed.lock ->pool_lock#2 ->key ->pcpu_lock ->percpu_counters_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->sb_writers#4 ->sb_writers#14 FD: 141 BD: 84 +.+.: &mm->mmap_lock/1 ->fs_reclaim ->pool_lock#2 ->&c->lock ->&vma->vm_lock->lock ->&mapping->i_mmap_rwsem ->&____s->seqcount ->&anon_vma->rwsem ->mmu_notifier_invalidate_range_start ->&mm->page_table_lock ->ptlock_ptr(ptdesc)#2 ->remove_cache_srcu ->&n->list_lock ->&rq->__lock ->&sem->wait_lock ->&p->pi_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->key ->pcpu_lock ->percpu_counters_lock ->&____s->seqcount#2 ->key#22 ->&rcu_state.expedited_wq ->pool_lock FD: 160 BD: 4 +.+.: &sb->s_type->i_mutex_key#13 ->&xattrs->lock ->tk_core.seq.seqcount ->&mm->mmap_lock ->fs_reclaim ->&____s->seqcount ->&xa->xa_lock#8 ->&sb->s_type->i_lock_key ->&info->lock ->lock#4 ->&wb->list_lock ->&rq->__lock ->key#9 ->&sb->s_type->i_mutex_key#13/4 ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&simple_offset_xa_lock ->&dentry->d_lock ->&mapping->i_mmap_rwsem ->lock#5 ->&lruvec->lru_lock ->&obj_hash[i].lock ->tomoyo_ss ->rcu_node_0 ->&cfs_rq->removed.lock ->pool_lock#2 ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 65 BD: 8 +.+.: &u->bindlock ->&net->unx.table.locks[i] ->&net->unx.table.locks[i]/1 ->&bsd_socket_locks[i] ->fs_reclaim ->pool_lock#2 ->batched_entropy_u32.lock ->&rq->__lock FD: 38 BD: 11 +.+.: &net->unx.table.locks[i]/1 ->&dentry->d_lock FD: 1 BD: 10 +.+.: &bsd_socket_locks[i] FD: 170 BD: 1 +.+.: &u->iolock ->rlock-AF_UNIX ->&mm->mmap_lock ->&obj_hash[i].lock ->pool_lock#2 ->&u->peer_wait ->&rq->__lock ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->&u->lock ->&dir->lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->&base->lock ->stock_lock FD: 28 BD: 283 ..-.: &ei->socket.wq.wait ->&p->pi_lock ->&ep->lock FD: 1 BD: 2 ....: key#5 FD: 1 BD: 99 ....: key#6 FD: 1 BD: 99 ....: key#7 FD: 1 BD: 99 ....: key#8 FD: 40 BD: 4 +.+.: &u->lock/1 ->&sk->sk_peer_lock ->&dentry->d_lock ->&sk->sk_peer_lock/1 FD: 163 BD: 1 +.+.: &pipe->mutex/1 ->&pipe->rd_wait ->&pipe->wr_wait ->&rq->__lock ->&lock->wait_lock ->fs_reclaim ->&____s->seqcount ->&mm->mmap_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->rcu_node_0 ->stock_lock ->pool_lock#2 ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq FD: 28 BD: 4 ....: &pipe->rd_wait ->&p->pi_lock ->&ep->lock FD: 28 BD: 4 ....: &pipe->wr_wait ->&p->pi_lock ->&ep->lock FD: 43 BD: 1 .+.+: sb_writers#6 ->tk_core.seq.seqcount ->mount_lock ->&rq->__lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 171 BD: 1 +.+.: sk_lock-AF_NETLINK ->slock-AF_NETLINK ->&mm->mmap_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->&c->lock ->pcpu_alloc_mutex ->&obj_hash[i].lock ->batched_entropy_u32.lock ->vmap_purge_lock ->&fp->aux->used_maps_mutex ->rcu_node_0 ->&rq->__lock ->stock_lock ->&____s->seqcount#2 ->&f->f_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock FD: 1 BD: 2 +...: slock-AF_NETLINK FD: 2 BD: 6 +.+.: &sk->sk_peer_lock ->&sk->sk_peer_lock/1 FD: 1 BD: 1 ....: &rs->lock#2 FD: 55 BD: 3 +.+.: oom_adj_mutex ->&p->alloc_lock ->&rq->__lock ->oom_adj_mutex.wait_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 83 BD: 3 +.+.: &group->mark_mutex ->&fsnotify_mark_srcu ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->lock ->ucounts_lock ->&mark->lock ->&conn->lock ->&sb->s_type->i_lock_key#22 ->&sb->s_type->i_lock_key ->remove_cache_srcu ->&____s->seqcount#2 ->&rq->__lock ->&lock->wait_lock ->&n->list_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->rcu_node_0 ->batched_entropy_u8.lock ->kfence_freelist_lock ->stock_lock FD: 13 BD: 197 +.+.: &group->inotify_data.idr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 3 BD: 4 +.+.: &mark->lock ->&fsnotify_mark_srcu ->&conn->lock FD: 1 BD: 8 +.+.: &conn->lock FD: 179 BD: 2 +.+.: &ep->mtx ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&f->f_lock ->&ei->socket.wq.wait ->&ep->lock ->&group->notification_waitq ->&group->notification_lock ->&sighand->signalfd_wqh ->&sighand->siglock ->&mm->mmap_lock ->&rq->__lock ->&pipe->rd_wait ->key#10 ->&obj_hash[i].lock ->sysctl_lock ->&pipe->wr_wait ->&lock->wait_lock ->rcu_node_0 ->&cfs_rq->removed.lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&rcu_state.expedited_wq ->&____s->seqcount#2 FD: 180 BD: 1 +.+.: epnested_mutex ->&ep->mtx FD: 27 BD: 298 ...-: &ep->lock ->&ep->wq FD: 28 BD: 8 ....: &group->notification_waitq ->&p->pi_lock ->&ep->lock FD: 1 BD: 8 +.+.: &group->notification_lock FD: 28 BD: 104 ....: &sighand->signalfd_wqh ->&p->pi_lock ->&ep->lock FD: 761 BD: 1 .+.+: sb_writers#7 ->mount_lock ->tk_core.seq.seqcount ->&sb->s_type->i_lock_key#24 ->&wb->list_lock ->&type->i_mutex_dir_key#4 ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&of->mutex ->&obj_hash[i].lock ->&____s->seqcount ->&c->lock ->&n->list_lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&rq->__lock ->&root->kernfs_iattr_rwsem ->&dentry->d_lock ->tomoyo_ss ->iattr_mutex ->&sb->s_type->i_mutex_key#14 ->&xattrs->lock ->&cfs_rq->removed.lock FD: 30 BD: 1 ..-.: &(&krcp->monitor_work)->timer FD: 30 BD: 1 ..-.: &(&tbl->managed_work)->timer FD: 33 BD: 2 +.+.: (work_completion)(&(&krcp->monitor_work)->work) ->krc.lock ->&obj_hash[i].lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 3 BD: 8 +.+.: swap_lock ->&p->lock#2 FD: 66 BD: 1 .+.+: kn->active ->fs_reclaim ->pool_lock#2 ->&kernfs_locks->open_file_mutex[count] ->&k->list_lock ->&c->lock ->&n->list_lock ->uevent_sock_mutex ->&obj_hash[i].lock ->&____s->seqcount ->remove_cache_srcu FD: 50 BD: 63 +.+.: &kernfs_locks->open_file_mutex[count] ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock ->&____s->seqcount ->remove_cache_srcu ->&rq->__lock ->&____s->seqcount#2 ->&lock->wait_lock ->&cfs_rq->removed.lock ->rcu_node_0 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&rcu_state.expedited_wq FD: 760 BD: 6 +.+.: &of->mutex ->&obj_hash[i].lock ->&rq->__lock ->cgroup_mutex ->&p->pi_lock ->cgroup_mutex.wait_lock ->&cfs_rq->removed.lock ->&root->deactivate_waitq FD: 1 BD: 275 ..-.: rlock-AF_NETLINK FD: 26 BD: 299 ..-.: &ep->wq ->&p->pi_lock FD: 65 BD: 1 .+.+: kn->active#2 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->pool_lock#2 ->uevent_sock_mutex ->&obj_hash[i].lock ->&____s->seqcount ->&n->list_lock ->quarantine_lock ->remove_cache_srcu ->&rq->__lock FD: 1 BD: 6 ....: &nlk->wait FD: 32 BD: 2 +.+.: (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) ->krc.lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 18 BD: 1 +.-.: (&vblank->disable_timer) ->&dev->vbl_lock FD: 30 BD: 1 +.-.: (&q->timeout) FD: 37 BD: 1 +.+.: (wq_completion)kblockd ->(work_completion)(&q->timeout_work) ->(work_completion)(&(&hctx->run_work)->work) ->(work_completion)(&(&q->requeue_work)->work) FD: 27 BD: 2 +.+.: (work_completion)(&q->timeout_work) ->&tags->lock FD: 65 BD: 1 .+.+: kn->active#3 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 ->uevent_sock_mutex ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->quarantine_lock ->&n->list_lock ->remove_cache_srcu FD: 115 BD: 9 ++++: kn->active#4 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 ->uevent_sock_mutex ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&device->physical_node_lock ->&____s->seqcount ->udc_lock ->&rq->__lock ->fw_lock ->remove_cache_srcu ->quarantine_lock ->&rfkill->lock ->&base->lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&cfs_rq->removed.lock ->uevent_sock_mutex.wait_lock ->&p->pi_lock ->rcu_node_0 ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&lock->wait_lock ->&root->deactivate_waitq ->&rcu_state.expedited_wq FD: 1 BD: 7 +.+.: &sk->sk_peer_lock/1 FD: 26 BD: 99 -.-.: &x->wait#26 ->&p->pi_lock FD: 1 BD: 102 +.+.: &__ctx->lock FD: 34 BD: 98 -.-.: &fq->mq_flush_lock ->tk_core.seq.seqcount ->&q->requeue_lock ->&obj_hash[i].lock ->&x->wait#26 ->bit_wait_table + i ->quarantine_lock FD: 1 BD: 102 -.-.: &q->requeue_lock FD: 30 BD: 1 ..-.: &(&ovs_net->masks_rebalance)->timer FD: 235 BD: 6 +.+.: (work_completion)(&(&ovs_net->masks_rebalance)->work) ->ovs_mutex ->&obj_hash[i].lock ->&base->lock FD: 56 BD: 1 .+.+: kn->active#5 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] ->param_lock ->pool_lock#2 ->&on->poll FD: 52 BD: 2 +.+.: &sb->s_type->i_mutex_key#14 ->tk_core.seq.seqcount ->&root->kernfs_iattr_rwsem ->&sem->wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 51 +.+.: disk_events_mutex FD: 64 BD: 5 +.+.: &sb->s_type->i_mutex_key#13/4 ->&dentry->d_lock ->&simple_offset_xa_lock ->fs_reclaim ->tk_core.seq.seqcount ->rename_lock ->&rq->__lock ->rcu_node_0 ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 30 BD: 420 +.+.: &dentry->d_lock/2 ->&dentry->d_lock/3 FD: 29 BD: 421 +.+.: &dentry->d_lock/3 ->&____s->seqcount#6 ->&wq FD: 1 BD: 423 +.+.: &____s->seqcount#6/1 FD: 52 BD: 1 .+.+: kn->active#6 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 50 BD: 1 .+.+: kn->active#7 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->remove_cache_srcu FD: 50 BD: 1 .+.+: kn->active#8 ->fs_reclaim ->&c->lock ->&n->list_lock ->&kernfs_locks->open_file_mutex[count] ->remove_cache_srcu FD: 52 BD: 1 .+.+: kn->active#9 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 52 BD: 1 .+.+: kn->active#10 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 50 BD: 1 .+.+: kn->active#11 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->remove_cache_srcu ->&n->list_lock FD: 52 BD: 1 .+.+: kn->active#12 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 52 BD: 1 .+.+: kn->active#13 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock FD: 52 BD: 1 .+.+: kn->active#14 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->remove_cache_srcu ->&____s->seqcount#2 ->rcu_node_0 ->&rq->__lock FD: 52 BD: 1 .+.+: kn->active#15 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 52 BD: 1 .+.+: kn->active#16 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&c->lock ->&n->list_lock ->remove_cache_srcu ->&____s->seqcount#2 ->&____s->seqcount FD: 52 BD: 1 .+.+: kn->active#17 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] FD: 51 BD: 1 .+.+: kn->active#18 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->&rq->__lock ->remove_cache_srcu FD: 52 BD: 1 .+.+: kn->active#19 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 53 BD: 1 .+.+: kn->active#20 ->fs_reclaim ->&c->lock ->&____s->seqcount ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock ->&n->list_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount#2 ->remove_cache_srcu FD: 52 BD: 1 .+.+: kn->active#21 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] FD: 54 BD: 1 .+.+: kn->active#22 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock FD: 52 BD: 1 .+.+: kn->active#23 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] FD: 52 BD: 1 .+.+: kn->active#24 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 52 BD: 1 .+.+: kn->active#25 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 52 BD: 1 .+.+: kn->active#26 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 52 BD: 1 .+.+: kn->active#27 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] FD: 52 BD: 1 .+.+: kn->active#28 ->fs_reclaim ->&c->lock ->&kernfs_locks->open_file_mutex[count] FD: 54 BD: 1 .+.+: kn->active#29 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->dev_base_lock FD: 59 BD: 1 .+.+: kn->active#30 ->fs_reclaim ->&c->lock ->&n->list_lock ->&kernfs_locks->open_file_mutex[count] ->&dev->power.lock ->pci_lock FD: 52 BD: 1 .+.+: kn->active#31 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 52 BD: 1 .+.+: kn->active#32 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 73 BD: 53 +.-.: slock-AF_INET/1 ->tk_core.seq.seqcount ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&hashinfo->ehash_locks[i] ->&tcp_hashinfo.bhash[i].lock ->&c->lock ->&meta->lock ->kfence_freelist_lock ->&n->list_lock ->quarantine_lock ->batched_entropy_u8.lock ->&____s->seqcount#2 ->&sctp_ep_hashtable[i].lock ->clock-AF_INET FD: 104 BD: 49 ++++: devnet_rename_sem ->(console_sem).lock ->fs_reclaim ->pool_lock#2 ->&k->list_lock ->&root->kernfs_rwsem ->kernfs_rename_lock ->uevent_sock_mutex ->&obj_hash[i].lock ->&c->lock ->&rq->__lock ->&n->list_lock ->&____s->seqcount ->&sem->wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock ->uevent_sock_mutex.wait_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->batched_entropy_u8.lock ->&____s->seqcount#2 ->rcu_node_0 ->&rcu_state.expedited_wq ->remove_cache_srcu FD: 1 BD: 199 ....: kernfs_rename_lock FD: 52 BD: 1 .+.+: kn->active#33 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 52 BD: 1 .+.+: kn->active#34 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 1 BD: 1 +.+.: &evdev->client_lock FD: 28 BD: 1 +.+.: &evdev->mutex ->&dev->mutex#2 FD: 1 BD: 1 +.+.: &sb->s_type->i_mutex_key#15 FD: 47 BD: 1 .+.+: mapping.invalidate_lock#2 ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&xa->xa_lock#8 ->lock#4 ->pool_lock#2 ->tk_core.seq.seqcount ->&c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 25 BD: 53 +.+.: &nft_net->commit_mutex ->&rq->__lock FD: 1 BD: 53 ....: target_list_lock FD: 210 BD: 2 +.+.: sk_lock-AF_INET ->slock-AF_INET#2 ->&table->hash[i].lock ->&tcp_hashinfo.bhash[i].lock ->&h->lhash2[i].lock ->&icsk->icsk_accept_queue.rskq_lock ->clock-AF_INET ->&obj_hash[i].lock ->&base->lock ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&mm->mmap_lock ->tk_core.seq.seqcount ->&sd->defer_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->remove_cache_srcu ->&n->list_lock ->&rq->__lock ->rcu_node_0 ->once_mutex ->batched_entropy_u32.lock ->batched_entropy_u16.lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->&rcu_state.expedited_wq ->&meta->lock ->quarantine_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->&msk->pm.lock FD: 75 BD: 7 +.-.: slock-AF_INET#2 ->&obj_hash[i].lock ->batched_entropy_u16.lock ->&tcp_hashinfo.bhash[i].lock ->&hashinfo->ehash_locks[i] ->tk_core.seq.seqcount ->(&req->rsk_timer) ->&base->lock ->&icsk->icsk_accept_queue.rskq_lock ->pool_lock#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&sk->sk_lock.wq ->&____s->seqcount#2 ->key#23 FD: 1 BD: 74 ++..: clock-AF_INET FD: 1 BD: 5 ....: key#9 FD: 52 BD: 1 .+.+: kn->active#35 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 1 BD: 3 ....: key#10 FD: 52 BD: 1 .+.+: kn->active#36 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 52 BD: 1 .+.+: kn->active#37 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 25 BD: 1 +.+.: fh->state->lock ->&rq->__lock FD: 1 BD: 5 ....: &vdev->fh_lock FD: 55 BD: 2 +.+.: &dev->dev_mutex ->fs_reclaim ->&c->lock ->pool_lock#2 ->vim2m:1183:(hdl)->_lock ->&obj_hash[i].lock ->&vdev->fh_lock ->&m2m_dev->job_spinlock ->&q->done_wq ->&q->mmap_lock FD: 55 BD: 2 +.+.: &dev_instance->mutex ->fs_reclaim ->&c->lock ->pool_lock#2 ->vicodec_core:1851:(hdl)->_lock ->&vdev->fh_lock ->&m2m_dev->job_spinlock ->&q->done_wq ->&q->mmap_lock ->&obj_hash[i].lock ->&____s->seqcount FD: 4 BD: 3 +.+.: vim2m:1183:(hdl)->_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 4 BD: 3 +.+.: vicodec_core:1851:(hdl)->_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 60 BD: 1 +.+.: &mdev->req_queue_mutex ->&dev_instance->mutex ->&vdev->fh_lock ->&mdev->graph_mutex ->vim2m:1183:(hdl)->_lock ->&dev->dev_mutex ->vicodec_core:1851:(hdl)->_lock ->&obj_hash[i].lock ->pool_lock#2 ->&dev->mutex#3 FD: 1 BD: 4 ....: &m2m_dev->job_spinlock FD: 1 BD: 6 ....: &q->done_wq FD: 1 BD: 6 +.+.: &q->mmap_lock FD: 50 BD: 1 .+.+: kn->active#38 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->remove_cache_srcu ->&____s->seqcount FD: 3 BD: 1 +.+.: &vcapture->lock ->&q->done_wq ->&q->mmap_lock FD: 4 BD: 2 +.+.: &dev->mutex#3 ->&vdev->fh_lock ->&q->done_wq ->&q->mmap_lock FD: 25 BD: 11 +.+.: &lo->lo_mutex ->&rq->__lock FD: 71 BD: 14 +.+.: &nbd->config_lock ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->&bdev->bd_size_lock ->&q->queue_lock ->&ACCESS_PRIVATE(sdp, lock) ->set->srcu ->&obj_hash[i].lock ->&rq->__lock ->&x->wait#3 ->&c->lock ->&____s->seqcount ->&n->list_lock ->&cfs_rq->removed.lock ->fs_reclaim ->uevent_sock_mutex ->&lock->wait_lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 30 BD: 8 ....: &ACCESS_PRIVATE(ssp->srcu_sup, lock) ->&obj_hash[i].lock ->&base->lock FD: 2 BD: 10 +.+.: &new->lock ->&mtdblk->cache_mutex FD: 1 BD: 11 +.+.: &mtdblk->cache_mutex FD: 52 BD: 1 .+.+: kn->active#39 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 159 BD: 1 +.+.: &mtd->master.chrdev_lock ->&mm->mmap_lock FD: 30 BD: 1 ..-.: &(&wb->dwork)->timer FD: 131 BD: 1 +.+.: (wq_completion)writeback ->(work_completion)(&(&wb->dwork)->work) ->(work_completion)(&(&wb->bw_dwork)->work) ->(work_completion)(&barr->work) ->&rq->__lock FD: 128 BD: 2 +.+.: (work_completion)(&(&wb->dwork)->work) ->&wb->work_lock ->&wb->list_lock ->&p->sequence ->key#11 ->&pl->lock ->&rq->__lock FD: 2 BD: 4 +.-.: &p->sequence ->key#14 FD: 1 BD: 304 -.-.: key#11 FD: 52 BD: 1 .+.+: kn->active#40 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 1 BD: 6 +.+.: destroy_lock FD: 63 BD: 2 +.+.: connector_reaper_work ->destroy_lock ->&ACCESS_PRIVATE(sdp, lock) ->&fsnotify_mark_srcu ->&obj_hash[i].lock ->&x->wait#3 ->&rq->__lock ->pool_lock#2 ->&base->lock ->pool_lock ->&cfs_rq->removed.lock ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->quarantine_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&meta->lock ->kfence_freelist_lock FD: 30 BD: 1 ..-.: fs/notify/mark.c:89 FD: 63 BD: 2 +.+.: (reaper_work).work ->destroy_lock ->&ACCESS_PRIVATE(sdp, lock) ->&fsnotify_mark_srcu ->&obj_hash[i].lock ->&x->wait#3 ->&rq->__lock ->pool_lock#2 ->&cfs_rq->removed.lock ->&base->lock ->rcu_node_0 ->pool_lock ->&meta->lock ->kfence_freelist_lock ->&rcu_state.expedited_wq ->&ACCESS_PRIVATE(ssp->srcu_sup, lock) ->quarantine_lock ->stock_lock FD: 26 BD: 1 +.-.: (&journal->j_commit_timer) ->&p->pi_lock FD: 75 BD: 1 +.+.: &journal->j_checkpoint_mutex ->mmu_notifier_invalidate_range_start ->pool_lock#2 ->tk_core.seq.seqcount ->bit_wait_table + i ->&rq->__lock ->&journal->j_state_lock FD: 26 BD: 99 ....: &journal->j_wait_transaction_locked ->&p->pi_lock FD: 1 BD: 106 -.-.: &memcg->move_lock FD: 1 BD: 100 +.+.: &sbi->s_md_lock FD: 1 BD: 1 ....: &journal->j_fc_wait FD: 1 BD: 1 +.+.: &journal->j_history_lock FD: 11 BD: 50 +...: fib_info_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 77 BD: 50 +...: &net->sctp.local_addr_lock ->&net->sctp.addr_wq_lock FD: 76 BD: 52 +.-.: &net->sctp.addr_wq_lock ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&____s->seqcount#2 ->k-slock-AF_INET6/1 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->slock-AF_INET6/1 ->slock-AF_INET/1 FD: 1 BD: 2 +...: clock-AF_NETLINK FD: 55 BD: 52 +.+.: (work_completion)(&ht->run_work) ->&ht->mutex ->&rq->__lock FD: 54 BD: 53 +.+.: &ht->mutex ->fs_reclaim ->pool_lock#2 ->batched_entropy_u32.lock ->rhashtable_bucket ->&ht->lock ->&c->lock ->&n->list_lock ->&obj_hash[i].lock ->&____s->seqcount ->&meta->lock ->kfence_freelist_lock ->&rq->__lock ->batched_entropy_u8.lock ->remove_cache_srcu ->&base->lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->quarantine_lock ->rcu_node_0 FD: 1 BD: 104 ....: rhashtable_bucket/1 FD: 10 BD: 54 +.+.: &ht->lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 49 +...: _xmit_LOOPBACK FD: 25 BD: 55 .+.+: netpoll_srcu ->&rq->__lock FD: 13 BD: 60 +.-.: &in_dev->mc_tomb_lock ->&c->lock ->&n->list_lock ->pool_lock#2 ->&obj_hash[i].lock ->&____s->seqcount#2 ->&____s->seqcount ->quarantine_lock FD: 19 BD: 56 +.-.: &im->lock ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&n->list_lock ->&____s->seqcount#2 ->init_task.mems_allowed_seq.seqcount FD: 1 BD: 55 +.+.: cbs_list_lock FD: 11 BD: 53 +...: &net->ipv6.addrconf_hash_lock ->&obj_hash[i].lock FD: 31 BD: 84 +...: &ifa->lock ->batched_entropy_u32.lock ->crngs.lock ->&obj_hash[i].lock ->&base->lock FD: 41 BD: 85 +...: &tb->tb6_lock ->&net->ipv6.fib6_walker_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->rlock-AF_NETLINK ->rt6_exception_lock ->&n->list_lock ->&data->fib_event_queue_lock ->quarantine_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&base->lock FD: 1 BD: 86 ++..: &net->ipv6.fib6_walker_lock FD: 232 BD: 2 +.+.: sk_lock-AF_INET6 ->slock-AF_INET6 ->&table->hash[i].lock ->batched_entropy_u32.lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&obj_hash[i].lock ->batched_entropy_u16.lock ->&tcp_hashinfo.bhash[i].lock ->&h->lhash2[i].lock ->fs_reclaim ->&mm->mmap_lock ->once_lock ->clock-AF_INET6 ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&msk->pm.lock ->tcpv6_prot_mutex ->device_spinlock ->stock_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&dir->lock ->k-sk_lock-AF_INET6/1 ->k-slock-AF_INET6 ->k-clock-AF_INET6 ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&n->list_lock ->&____s->seqcount#2 ->key#26 ->(kmod_concurrent_max).lock ->&x->wait#17 ->running_helpers_waitq.lock ->key ->pcpu_lock ->percpu_counters_lock ->&sighand->siglock ->k-sk_lock-AF_INET6 ->remove_cache_srcu ->vmap_area_lock ->crngs.lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 35 BD: 4 +...: slock-AF_INET6 ->&obj_hash[i].lock ->&tcp_hashinfo.bhash[i].lock ->pool_lock#2 ->key#23 ->&sk->sk_lock.wq FD: 1 BD: 75 ++..: clock-AF_INET6 FD: 1 BD: 1 +.+.: userns_state_mutex FD: 2 BD: 1 +.+.: sk_lock-AF_UNIX ->slock-AF_UNIX FD: 1 BD: 2 +...: slock-AF_UNIX FD: 1 BD: 8 +.+.: vmap_purge_lock.wait_lock FD: 52 BD: 1 .+.+: kn->active#41 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 165 BD: 49 ++++: dev_addr_sem ->net_rwsem ->&tn->lock ->&sdata->sec_mtx ->fs_reclaim ->pool_lock#2 ->nl_table_lock ->rlock-AF_NETLINK ->nl_table_wait.lock ->&tbl->lock ->&pn->hash_lock ->&obj_hash[i].lock ->input_pool.lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->&br->lock ->&n->list_lock ->quarantine_lock ->remove_cache_srcu ->_xmit_ETHER ->&hard_iface->bat_iv.ogm_buff_mutex ->rcu_node_0 ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&rcu_state.expedited_wq ->team->team_lock_key#101 ->team->team_lock_key#115 ->team->team_lock_key#116 ->team->team_lock_key#117 ->team->team_lock_key#118 FD: 734 BD: 3 +.+.: nlk_cb_mutex-GENERIC ->fs_reclaim ->pool_lock#2 ->&____s->seqcount ->rtnl_mutex ->&rdev->wiphy.mtx ->rlock-AF_NETLINK ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&rq->__lock ->&____s->seqcount#2 ->remove_cache_srcu ->&cfs_rq->removed.lock ->&devlink->lock_key#101 ->&devlink->lock_key#115 ->&devlink->lock_key#116 ->&devlink->lock_key#117 ->&devlink->lock_key#118 FD: 19 BD: 53 +...: &rdev->bss_lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->quarantine_lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&base->lock FD: 77 BD: 1 +.-.: (&net->sctp.addr_wq_timer) ->&net->sctp.addr_wq_lock FD: 1 BD: 50 +.+.: napi_hash_lock FD: 12 BD: 49 ++..: lapb_list_lock ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&c->lock FD: 726 BD: 2 +.+.: (work_completion)(&aux->work) ->vmap_area_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->vmap_purge_lock ->pool_lock#2 ->pcpu_lock ->stock_lock ->quarantine_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&base->lock ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->rcu_node_0 ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&cfs_rq->removed.lock FD: 4 BD: 49 ++.-: x25_neigh_list_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 49 +...: _xmit_SLIP FD: 13 BD: 1 +.-.: (&eql->timer) ->&eql->queue.lock ->&obj_hash[i].lock ->&base->lock FD: 4 BD: 52 +.-.: &eql->queue.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 52 BD: 91 +.+.: xps_map_mutex ->fs_reclaim ->pool_lock#2 ->jump_label_mutex ->&rq->__lock FD: 1 BD: 53 +.+.: &data->mutex FD: 16 BD: 70 +...: &local->filter_lock ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->krc.lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 30 BD: 71 ..-.: &rdev->wiphy_work_lock FD: 376 BD: 7 +.+.: (work_completion)(&rdev->wiphy_work) ->&rdev->wiphy.mtx ->&lock->wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 FD: 1 BD: 49 +...: _xmit_VOID FD: 1 BD: 49 +...: _xmit_X25 FD: 4 BD: 50 +...: &lapbeth->up_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 61 BD: 50 +.-.: &lapb->lock ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&c->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount ->&list->lock#20 FD: 1 BD: 145 +.-.: &ul->lock FD: 74 BD: 49 +...: dev->qdisc_tx_busylock ?: &qdisc_tx_busylock ->_xmit_ETHER#2 ->&obj_hash[i].lock ->pool_lock#2 ->&sch->q.lock ->&r->producer_lock#3 ->_xmit_SLIP#2 ->_xmit_NETROM ->quarantine_lock ->&base->lock ->&meta->lock ->kfence_freelist_lock FD: 58 BD: 75 +.-.: _xmit_ETHER#2 ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->&____s->seqcount ->quarantine_lock ->&base->lock FD: 2 BD: 98 +.+.: &(ei->i_block_reservation_lock) ->key#15 FD: 32 BD: 2 +.+.: (work_completion)(&work->work) ->devices_rwsem ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->quarantine_lock ->&base->lock ->&cfs_rq->removed.lock FD: 724 BD: 2 +.+.: (work_completion)(&(&ifa->dad_work)->work) ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->rcu_node_0 FD: 5 BD: 64 +.+.: fasync_lock ->&new->fa_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 105 ....: &tty->ctrl.lock FD: 164 BD: 6 +.+.: &buf->lock ->&tty->termios_rwsem ->&rq->__lock FD: 1 BD: 11 ....: &tty->flow.lock FD: 1 BD: 86 +.-.: rt6_exception_lock FD: 165 BD: 6 +.+.: &ldata->atomic_read_lock ->&tty->termios_rwsem ->(work_completion)(&buf->work) ->&rq->__lock FD: 31 BD: 1 +.-.: &tx->clean_lock ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->&base->lock FD: 30 BD: 1 ..-.: &(&idev->mc_dad_work)->timer FD: 136 BD: 1 +.+.: (wq_completion)mld ->(work_completion)(&(&idev->mc_dad_work)->work) ->(work_completion)(&(&idev->mc_ifc_work)->work) FD: 134 BD: 2 +.+.: (work_completion)(&(&idev->mc_dad_work)->work) ->&idev->mc_lock ->&rq->__lock FD: 1 BD: 54 ....: class FD: 1 BD: 54 ....: (&tbl->proxy_timer) FD: 30 BD: 1 ..-.: &(&idev->mc_ifc_work)->timer FD: 134 BD: 2 +.+.: (work_completion)(&(&idev->mc_ifc_work)->work) ->&idev->mc_lock ->&rq->__lock FD: 15 BD: 58 +.-.: &ul->lock#2 ->pool_lock#2 ->&dir->lock#2 ->&c->lock ->&n->list_lock FD: 14 BD: 130 ++--: &n->lock ->&____s->seqcount#8 ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&(&n->ha_lock)->lock FD: 1 BD: 132 +.--: &____s->seqcount#8 FD: 30 BD: 1 ..-.: net/core/link_watch.c:31 FD: 82 BD: 50 +.-.: &dev->tx_global_lock ->_xmit_ETHER#2 ->&obj_hash[i].lock ->&base->lock ->_xmit_NETROM ->_xmit_NONE#2 ->_xmit_TUNNEL6#2 ->_xmit_SIT#2 ->_xmit_TUNNEL#2 ->_xmit_IPGRE#2 ->&qdisc_xmit_lock_key ->&qdisc_xmit_lock_key#2 ->&vlan_netdev_xmit_lock_key ->&batadv_netdev_xmit_lock_key ->&qdisc_xmit_lock_key#3 ->&qdisc_xmit_lock_key#4 ->_xmit_LOOPBACK#2 FD: 1 BD: 57 +.-.: &sch->q.lock FD: 1 BD: 129 +.-.: lock#8 FD: 1 BD: 129 ..-.: id_table_lock FD: 38 BD: 2 +.+.: (work_completion)(&w->work)#2 ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->nf_conntrack_mutex ->nf_conntrack_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 143 ...-: &____s->seqcount#9 FD: 30 BD: 1 ..-.: &(&ifa->dad_work)->timer FD: 1 BD: 2 +.+.: &net->packet.sklist_lock FD: 179 BD: 2 +.+.: sk_lock-AF_PACKET ->slock-AF_PACKET ->&po->bind_lock ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->&mm->mmap_lock ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->pcpu_alloc_mutex ->&c->lock ->batched_entropy_u32.lock ->vmap_purge_lock ->&fp->aux->used_maps_mutex ->&n->list_lock ->pool_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->rcu_state.exp_mutex.wait_lock ->&p->pi_lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock FD: 1 BD: 3 +...: slock-AF_PACKET FD: 16 BD: 3 +.+.: &po->bind_lock ->ptype_lock ->pool_lock#2 ->&dir->lock#2 FD: 1 BD: 76 +.-.: rlock-AF_PACKET FD: 1 BD: 1 +...: wlock-AF_PACKET FD: 1 BD: 2 +.+.: fanout_mutex FD: 1 BD: 2 +...: clock-AF_PACKET FD: 30 BD: 1 ..-.: drivers/base/dd.c:321 FD: 36 BD: 2 +.+.: (deferred_probe_timeout_work).work ->device_links_lock ->deferred_probe_mutex ->&rq->__lock ->deferred_probe_work FD: 84 BD: 1 ++++: &type->s_umount_key#40 ->&sb->s_type->i_lock_key#3 ->&xa->xa_lock#8 ->mmu_notifier_invalidate_range_start ->batched_entropy_u8.lock ->kfence_freelist_lock ->tk_core.seq.seqcount ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->lock#4 ->lock#5 ->&wb->list_lock ->&rq->__lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->&x->wait#23 ->shrinker_mutex ->rename_lock.seqcount ->&dentry->d_lock ->&s->s_inode_list_lock ->&fsnotify_mark_srcu ->&dentry->d_lock/1 FD: 1 BD: 304 -.-.: &s->s_inode_wblist_lock FD: 1 BD: 305 -.-.: key#12 FD: 77 BD: 98 +.+.: &lg->lg_mutex ->&ei->i_prealloc_lock ->mmu_notifier_invalidate_range_start ->&____s->seqcount ->&c->lock ->pool_lock#2 ->lock#4 ->&mapping->i_private_lock ->&ret->b_state_lock ->&journal->j_revoke_lock ->&pa->pa_lock ->&lg->lg_prealloc_lock FD: 1 BD: 99 +.+.: &pa->pa_lock FD: 1 BD: 99 +.+.: &lg->lg_prealloc_lock FD: 30 BD: 4 -...: &ei->i_completed_io_lock FD: 129 BD: 2 +.+.: (wq_completion)ext4-rsv-conversion ->(work_completion)(&ei->i_rsv_conversion_work) ->&rq->__lock FD: 128 BD: 3 +.+.: (work_completion)(&ei->i_rsv_conversion_work) ->&ei->i_completed_io_lock ->&journal->j_state_lock ->jbd2_handle ->&obj_hash[i].lock ->pool_lock#2 ->&ext4__ioend_wq[i] ->&ret->b_uptodate_lock ->&folio_wait_table[i] ->&rq->__lock ->&lruvec->lru_lock FD: 1 BD: 99 ....: &journal->j_wait_reserved FD: 1 BD: 4 ....: &ext4__ioend_wq[i] FD: 30 BD: 1 ..-.: &(&wb->bw_dwork)->timer FD: 56 BD: 2 +.+.: (work_completion)(&(&wb->bw_dwork)->work) ->&wb->list_lock FD: 1 BD: 60 +.-.: &ct->lock FD: 1 BD: 105 ....: key#13 FD: 62 BD: 1 +.-.: (&lapb->t1timer) ->&lapb->lock FD: 83 BD: 1 +.-.: (&dev->watchdog_timer) ->&dev->tx_global_lock FD: 14 BD: 1 +.-.: (&dom->period_timer) ->key#14 ->&p->sequence ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 306 -.-.: key#14 FD: 2 BD: 304 -.-.: &pl->lock ->key#12 FD: 30 BD: 1 ..-.: &(&tbl->gc_work)->timer FD: 46 BD: 2 +.+.: (work_completion)(&(&tbl->gc_work)->work) ->&tbl->lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 FD: 5 BD: 127 +.-.: &nf_conntrack_locks[i] ->&nf_conntrack_locks[i]/1 ->batched_entropy_u8.lock FD: 4 BD: 128 +.-.: &nf_conntrack_locks[i]/1 ->batched_entropy_u8.lock FD: 1 BD: 71 +.-.: &hashinfo->ehash_locks[i] FD: 2 BD: 131 +.-.: &(&n->ha_lock)->lock ->&____s->seqcount#8 FD: 1 BD: 8 ..-.: (&req->rsk_timer) FD: 1 BD: 8 +.-.: &icsk->icsk_accept_queue.rskq_lock FD: 1 BD: 3 +.-.: &sd->defer_lock FD: 76 BD: 1 +.-.: (&icsk->icsk_retransmit_timer) ->slock-AF_INET#2 FD: 76 BD: 1 +.-.: (&icsk->icsk_delack_timer) ->slock-AF_INET#2 FD: 2 BD: 54 +.-.: tcp_metrics_lock ->pool_lock#2 FD: 1 BD: 99 ....: key#15 FD: 73 BD: 97 +.+.: &sbi->s_orphan_lock ->&rq->__lock ->mmu_notifier_invalidate_range_start ->&ei->i_raw_lock ->&lock->wait_lock ->rcu_node_0 ->bit_wait_table + i ->&ret->b_state_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 ->&mapping->i_private_lock ->&____s->seqcount ->&rcu_state.expedited_wq ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock FD: 52 BD: 1 .+.+: kn->active#42 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 1 BD: 1 +.+.: &futex_queues[i].lock FD: 1 BD: 4 ....: &on->poll FD: 1 BD: 3 +.+.: module_mutex FD: 3 BD: 3 +.+.: once_mutex ->crngs.lock FD: 26 BD: 62 ....: &sk->sk_lock.wq ->&p->pi_lock FD: 30 BD: 1 ..-.: &(&dm_bufio_cleanup_old_work)->timer FD: 14 BD: 1 +.+.: (wq_completion)dm_bufio_cache ->(work_completion)(&(&dm_bufio_cleanup_old_work)->work) FD: 13 BD: 2 +.+.: (work_completion)(&(&dm_bufio_cleanup_old_work)->work) ->dm_bufio_clients_lock ->&obj_hash[i].lock ->&base->lock FD: 97 BD: 1 +.+.: &type->s_umount_key#41/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&root->kernfs_rwsem ->&____s->seqcount ->&sb->s_type->i_lock_key#30 ->crngs.lock ->&root->kernfs_supers_rwsem ->&dentry->d_lock FD: 38 BD: 248 +.+.: &sb->s_type->i_lock_key#30 ->&dentry->d_lock FD: 756 BD: 1 .+.+: sb_writers#8 ->mount_lock ->&type->i_mutex_dir_key#6 ->fs_reclaim ->&mm->mmap_lock ->&of->mutex ->&obj_hash[i].lock ->&type->i_mutex_dir_key#6/1 ->&c->lock ->&rq->__lock ->&n->list_lock ->remove_cache_srcu FD: 52 BD: 2 ++++: &type->i_mutex_dir_key#6 ->tomoyo_ss ->tk_core.seq.seqcount ->&root->kernfs_iattr_rwsem ->rename_lock.seqcount ->fs_reclaim ->&c->lock ->&dentry->d_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#30 FD: 52 BD: 1 ++++: kn->active#43 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->pool_lock#2 FD: 97 BD: 1 +.+.: &type->s_umount_key#42/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#31 ->crngs.lock ->&root->kernfs_supers_rwsem ->&dentry->d_lock ->&c->lock ->&n->list_lock FD: 38 BD: 248 +.+.: &sb->s_type->i_lock_key#31 ->&dentry->d_lock FD: 78 BD: 1 ++++: &type->s_umount_key#43 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->percpu_ref_switch_lock ->&root->kernfs_supers_rwsem ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#31 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->inode_hash_lock ->pool_lock#2 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 ->&rq->__lock FD: 1 BD: 1 ..-.: percpu_ref_switch_waitq.lock FD: 733 BD: 2 +.+.: (work_completion)(&cgrp->bpf.release_work) ->cgroup_mutex ->cgroup_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->percpu_ref_switch_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 18 +.+.: cgroup_mutex.wait_lock FD: 738 BD: 1 +.+.: (wq_completion)cgroup_destroy ->(work_completion)(&css->destroy_work) ->(work_completion)(&(&css->destroy_rwork)->work) FD: 733 BD: 2 +.+.: (work_completion)(&css->destroy_work) ->cgroup_mutex ->&obj_hash[i].lock ->pool_lock#2 ->cgroup_mutex.wait_lock ->&p->pi_lock FD: 736 BD: 2 +.+.: (work_completion)(&(&css->destroy_rwork)->work) ->percpu_ref_switch_lock ->&obj_hash[i].lock ->pool_lock#2 ->&cgrp->pidlist_mutex ->(wq_completion)cgroup_pidlist_destroy ->&wq->mutex ->(work_completion)(&cgrp->release_agent_work) ->cgroup_mutex ->cgroup_rstat_lock ->pcpu_lock ->&root->kernfs_rwsem ->kernfs_idr_lock ->krc.lock ->cgroup_mutex.wait_lock ->&p->pi_lock FD: 1 BD: 3 +.+.: &cgrp->pidlist_mutex FD: 1 BD: 3 +.+.: (wq_completion)cgroup_pidlist_destroy FD: 1 BD: 3 +.+.: (work_completion)(&cgrp->release_agent_work) FD: 756 BD: 1 .+.+: sb_writers#9 ->mount_lock ->&type->i_mutex_dir_key#7 ->fs_reclaim ->&mm->mmap_lock ->&of->mutex ->&obj_hash[i].lock ->&type->i_mutex_dir_key#7/1 ->&c->lock ->&n->list_lock ->remove_cache_srcu FD: 65 BD: 2 ++++: &type->i_mutex_dir_key#7 ->tomoyo_ss ->tk_core.seq.seqcount ->&root->kernfs_iattr_rwsem ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->&root->kernfs_rwsem ->&sb->s_type->i_lock_key#31 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->pool_lock#2 ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->&rq->__lock FD: 1 BD: 18 +.+.: &dom->lock FD: 30 BD: 1 ..-.: drivers/regulator/core.c:6335 FD: 4 BD: 2 +.+.: (regulator_init_complete_work).work ->&k->list_lock ->&k->k_lock FD: 52 BD: 1 .+.+: kn->active#44 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] FD: 228 BD: 1 .+.+: kn->active#45 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->cpu_hotplug_lock FD: 160 BD: 1 .+.+: sb_writers#10 ->&mm->mmap_lock ->&attr->mutex FD: 159 BD: 2 +.+.: &attr->mutex ->&mm->mmap_lock FD: 40 BD: 2 ++++: &type->s_umount_key#44 ->sb_lock ->&dentry->d_lock FD: 78 BD: 2 +.+.: &sb->s_type->i_mutex_key#16 ->namespace_sem ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#26 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->init_binfmt_misc.entries_lock FD: 168 BD: 1 .+.+: sb_writers#11 ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&sb->s_type->i_mutex_key#16 FD: 1 BD: 49 +.+.: &wpan_dev->association_lock FD: 1 BD: 51 +...: &pn->hash_lock FD: 42 BD: 1 +...: &net->ipv6.fib6_gc_lock ->&obj_hash[i].lock FD: 1 BD: 49 +...: _xmit_IEEE802154 FD: 1 BD: 3 +.+.: &xa->xa_lock#16 FD: 1 BD: 6 ....: genl_sk_destructing_waitq.lock FD: 1 BD: 3 +...: &rdev->beacon_registrations_lock FD: 1 BD: 54 +...: &rdev->mgmt_registrations_lock FD: 1 BD: 54 +...: &wdev->pmsr_lock FD: 1 BD: 50 +.+.: reg_indoor_lock FD: 139 BD: 2 +.+.: (work_completion)(&w->w) ->nfc_devlist_mutex ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock ->nfc_devlist_mutex.wait_lock ->&p->pi_lock ->&base->lock ->&rq->__lock ->quarantine_lock ->rcu_node_0 FD: 25 BD: 5 +.+.: &genl_data->genl_data_mutex ->&rq->__lock FD: 50 BD: 6 +.+.: swap_cgroup_mutex ->fs_reclaim ->&____s->seqcount FD: 1 BD: 6 +.+.: &((cluster_info + ci)->lock)/1 FD: 67 BD: 6 +.+.: swapon_mutex ->fs_reclaim ->pool_lock#2 ->swap_lock ->percpu_ref_switch_lock ->(console_sem).lock FD: 2 BD: 9 +.+.: &p->lock#2 ->swap_avail_lock FD: 1 BD: 10 +.+.: swap_avail_lock FD: 1 BD: 6 ....: proc_poll_wait.lock FD: 231 BD: 1 +.+.: swap_slots_cache_enable_mutex ->cpu_hotplug_lock ->swap_lock FD: 1 BD: 93 +.+.: swap_slots_cache_mutex FD: 1 BD: 3 +.+.: tomoyo_policy_lock.wait_lock FD: 1 BD: 136 ....: &newf->resize_wait FD: 13 BD: 115 ....: &kcov->lock ->kcov_remote_lock FD: 106 BD: 1 +.+.: pid_caches_mutex ->slab_mutex FD: 40 BD: 2 ++++: &type->s_umount_key#45 ->sb_lock ->&dentry->d_lock FD: 75 BD: 1 +.+.: &sb->s_type->i_mutex_key#17 ->namespace_sem ->&rq->__lock FD: 1 BD: 30 ++++: hci_sk_list.lock FD: 1 BD: 1 +.+.: (work_completion)(&(&data->open_timeout)->work) FD: 260 BD: 1 +.+.: &data->open_mutex ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->&x->wait#9 ->hci_index_ida.xa_lock ->pcpu_alloc_mutex ->cpu_hotplug_lock ->wq_pool_mutex ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->&k->list_lock ->gdp_mutex ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&n->list_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->subsys mutex#74 ->&dev->devres_lock ->triggers_list_lock ->leds_list_lock ->rfkill_global_mutex ->rfkill_global_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&rfkill->lock ->hci_dev_list_lock ->tk_core.seq.seqcount ->hci_sk_list.lock ->(pm_chain_head).rwsem ->&list->lock#5 ->&data->read_wait ->remove_cache_srcu ->&____s->seqcount#2 ->&sem->wait_lock ->&base->lock ->&cfs_rq->removed.lock ->quarantine_lock FD: 1 BD: 2 ....: hci_index_ida.xa_lock FD: 27 BD: 27 +.+.: subsys mutex#74 ->&k->k_lock ->&rq->__lock FD: 1 BD: 16 ++++: hci_dev_list_lock FD: 1 BD: 27 ....: &list->lock#5 FD: 26 BD: 27 ....: &data->read_wait ->&p->pi_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#145 FD: 177 BD: 11 +.+.: (work_completion)(&hdev->power_on) ->&hdev->req_lock ->fs_reclaim ->pool_lock#2 ->tk_core.seq.seqcount ->hci_sk_list.lock ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->rcu_node_0 ->&rq->__lock ->&____s->seqcount#2 ->&____s->seqcount FD: 176 BD: 13 +.+.: &hdev->req_lock ->&obj_hash[i].lock ->pool_lock#2 ->&c->lock ->&list->lock#6 ->&list->lock#7 ->&hdev->req_wait_q ->&base->lock ->&rq->__lock ->(&timer.timer) ->tk_core.seq.seqcount ->hci_sk_list.lock ->&____s->seqcount ->&cfs_rq->removed.lock ->pool_lock ->(work_completion)(&(&hdev->interleave_scan)->work) ->hci_dev_list_lock ->(work_completion)(&hdev->tx_work) ->(work_completion)(&hdev->rx_work) ->&wq->mutex ->&hdev->lock ->&list->lock#5 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&lock->wait_lock ->&p->pi_lock ->rcu_node_0 ->quarantine_lock ->(wq_completion)hci1#6 ->(wq_completion)hci0#6 ->(wq_completion)hci4#4 ->(wq_completion)hci3#6 ->(wq_completion)hci5#2 FD: 1 BD: 14 ....: &list->lock#6 FD: 1 BD: 28 ....: &list->lock#7 FD: 26 BD: 25 ....: &hdev->req_wait_q ->&p->pi_lock FD: 51 BD: 2 +.+.: sk_lock-AF_BLUETOOTH-BTPROTO_HCI ->slock-AF_BLUETOOTH-BTPROTO_HCI ->sock_cookie_ida.xa_lock ->&p->alloc_lock ->pool_lock#2 ->&c->lock ->&n->list_lock ->tk_core.seq.seqcount ->hci_sk_list.lock ->&obj_hash[i].lock ->clock-AF_BLUETOOTH ->&____s->seqcount ->&____s->seqcount#2 ->&rq->__lock FD: 1 BD: 3 +...: slock-AF_BLUETOOTH-BTPROTO_HCI FD: 1 BD: 3 ....: sock_cookie_ida.xa_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#149 FD: 53 BD: 24 +.+.: (work_completion)(&hdev->cmd_work) ->&list->lock#7 ->fs_reclaim ->pool_lock#2 ->tk_core.seq.seqcount ->&list->lock#5 ->&data->read_wait ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->remove_cache_srcu ->&____s->seqcount#2 ->quarantine_lock ->rcu_node_0 ->&n->list_lock ->&rcu_state.expedited_wq ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock FD: 158 BD: 24 +.+.: (work_completion)(&hdev->rx_work) ->&list->lock#7 ->lock#6 ->fs_reclaim ->pool_lock#2 ->free_vmap_area_lock ->vmap_area_lock ->&____s->seqcount ->init_mm.page_table_lock ->&c->lock ->&hdev->lock ->(console_sem).lock ->console_owner_lock ->console_owner ->&rq->__lock ->&obj_hash[i].lock ->&hdev->req_wait_q ->&base->lock ->chan_list_lock ->&n->list_lock ->&____s->seqcount#2 ->rcu_node_0 ->&meta->lock ->kfence_freelist_lock ->&rcu_state.expedited_wq ->remove_cache_srcu ->batched_entropy_u8.lock ->quarantine_lock ->&cfs_rq->removed.lock FD: 145 BD: 25 +.+.: &hdev->lock ->&xa->xa_lock#17 ->fs_reclaim ->pool_lock#2 ->&obj_hash[i].lock ->&x->wait#9 ->&k->list_lock ->lock ->&root->kernfs_rwsem ->bus_type_sem ->sysfs_symlink_target_lock ->&c->lock ->&dev->power.lock ->dpm_list_mtx ->&____s->seqcount ->uevent_sock_mutex ->&k->k_lock ->subsys mutex#74 ->&list->lock#7 ->&hdev->unregister_lock ->hci_cb_list_lock ->&base->lock ->tk_core.seq.seqcount ->hci_sk_list.lock ->pool_lock ->&n->list_lock ->(work_completion)(&(&conn->disc_work)->work) ->(work_completion)(&(&conn->auto_accept_work)->work) ->(work_completion)(&(&conn->idle_work)->work) ->&x->wait#2 ->&rq->__lock ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->deferred_probe_mutex ->device_links_lock ->mmu_notifier_invalidate_range_start ->&____s->seqcount#2 ->quarantine_lock ->&cfs_rq->removed.lock ->&sem->wait_lock ->&p->pi_lock ->hci_cb_list_lock.wait_lock ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->sk_lock-AF_BLUETOOTH-BTPROTO_SCO ->slock-AF_BLUETOOTH-BTPROTO_SCO ->&lock->wait_lock ->rcu_node_0 ->uevent_sock_mutex.wait_lock ->&rcu_state.expedited_wq ->dev_pm_qos_sysfs_mtx.wait_lock ->&meta->lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#415 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#585 ->&rq->__lock FD: 13 BD: 26 ....: &xa->xa_lock#17 ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->&____s->seqcount ->&base->lock FD: 52 BD: 27 +.+.: &hdev->unregister_lock ->fs_reclaim ->pool_lock#2 ->&hdev->cmd_sync_work_lock ->&c->lock ->&rq->__lock ->&n->list_lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->remove_cache_srcu ->rcu_node_0 ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 1 BD: 28 +.+.: &hdev->cmd_sync_work_lock FD: 177 BD: 11 +.+.: (work_completion)(&hdev->cmd_sync_work) ->&hdev->cmd_sync_work_lock ->&hdev->req_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->quarantine_lock ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 3 +...: clock-AF_BLUETOOTH FD: 1 BD: 1 +.+.: &undo_list->lock FD: 1 BD: 27 +.+.: &conn->ident_lock FD: 1 BD: 28 ....: &list->lock#8 FD: 1 BD: 29 +.+.: &conn->chan_lock FD: 1 BD: 49 +...: &nr_netdev_addr_lock_key FD: 1 BD: 49 +...: listen_lock FD: 30 BD: 24 +.+.: (work_completion)(&hdev->tx_work) ->&list->lock#8 ->tk_core.seq.seqcount ->&list->lock#5 ->&data->read_wait ->&list->lock#7 ->&rq->__lock FD: 26 BD: 24 +.+.: (work_completion)(&conn->pending_rx_work) ->&list->lock#9 ->&rq->__lock FD: 1 BD: 25 ....: &list->lock#9 FD: 1 BD: 172 +.+.: uevent_sock_mutex.wait_lock FD: 10 BD: 53 +...: &idev->mc_query_lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#210 FD: 2 BD: 6 +.+.: rdma_nets.xa_lock ->pool_lock#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#489 FD: 1 BD: 95 +.+.: gdp_mutex.wait_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#57 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#27 ->&rq->__lock FD: 1 BD: 4 +.+.: &____s->seqcount#10 FD: 2 BD: 3 +.+.: &(&net->ipv4.ping_group_range.lock)->lock ->&____s->seqcount#10 FD: 2 BD: 49 +.+.: &r->consumer_lock ->&r->producer_lock FD: 1 BD: 59 +...: &r->producer_lock FD: 16 BD: 57 +...: &bridge_netdev_addr_lock_key ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock ->&obj_hash[i].lock ->krc.lock FD: 40 BD: 52 +.-.: &br->hash_lock ->&____s->seqcount ->&c->lock ->pool_lock#2 ->nl_table_lock ->&obj_hash[i].lock ->nl_table_wait.lock ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->&base->lock ->rlock-AF_NETLINK FD: 52 BD: 50 +.+.: j1939_netdev_lock ->&rq->__lock ->fs_reclaim ->&____s->seqcount#2 ->&____s->seqcount ->&c->lock ->pool_lock#2 ->&net->can.rcvlists_lock ->&obj_hash[i].lock ->&priv->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&cfs_rq->removed.lock FD: 10 BD: 57 +...: &dev_addr_list_lock_key#2 ->pool_lock#2 ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock FD: 8 BD: 49 +...: &bat_priv->tvlv.handler_list_lock ->pool_lock#2 ->&c->lock ->&n->list_lock FD: 14 BD: 56 +...: &bat_priv->tvlv.container_list_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock ->&____s->seqcount#2 ->&____s->seqcount FD: 9 BD: 57 +...: &batadv_netdev_addr_lock_key ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 8 BD: 58 +...: &bat_priv->softif_vlan_list_lock ->pool_lock#2 ->&c->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 17 BD: 57 +...: key#16 ->&bat_priv->softif_vlan_list_lock ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 6 BD: 57 +.-.: &bat_priv->tt.changes_list_lock ->&obj_hash[i].lock ->pool_lock#2 ->&meta->lock ->kfence_freelist_lock FD: 30 BD: 1 ..-.: &(&bat_priv->nc.work)->timer FD: 61 BD: 1 +.+.: (wq_completion)bat_events ->(work_completion)(&(&bat_priv->nc.work)->work) ->(work_completion)(&(&bat_priv->mcast.work)->work) ->(work_completion)(&(&bat_priv->orig_work)->work) ->(work_completion)(&(&forw_packet_aggr->delayed_work)->work) ->(work_completion)(&(&bat_priv->tt.work)->work) ->(work_completion)(&(&bat_priv->dat.work)->work) ->(work_completion)(&(&bat_priv->bla.work)->work) ->(work_completion)(&barr->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 31 BD: 6 +.+.: (work_completion)(&(&bat_priv->nc.work)->work) ->key#17 ->key#18 ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->rcu_node_0 ->&rq->__lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 1 BD: 7 +...: key#17 FD: 1 BD: 7 +...: key#18 FD: 106 BD: 50 +.+.: init_lock ->slab_mutex ->fs_reclaim ->&zone->lock ->&____s->seqcount ->&obj_hash[i].lock ->&base->lock ->crngs.lock FD: 1 BD: 69 +.-.: deferred_lock FD: 724 BD: 2 +.+.: deferred_process_work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 54 BD: 51 +.-.: &br->lock ->&br->hash_lock ->lweventlist_lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&dir->lock#2 ->deferred_lock ->(console_sem).lock ->console_owner_lock ->console_owner ->nl_table_lock ->nl_table_wait.lock ->&br->multicast_lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->quarantine_lock FD: 1 BD: 50 +.+.: &bond->stats_lock/1 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1087 ->&rq->__lock FD: 42 BD: 59 +.+.: (work_completion)(&(&slave->notify_work)->work) ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->rtnl_mutex.wait_lock ->&p->pi_lock FD: 30 BD: 1 ..-.: &(&slave->notify_work)->timer FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#451 ->&rq->__lock FD: 1 BD: 2 ....: key#25 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#454 FD: 1 BD: 49 +...: &pmc->lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#194 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#28 FD: 40 BD: 52 +.+.: &hard_iface->bat_iv.ogm_buff_mutex ->crngs.lock ->pool_lock#2 ->batched_entropy_u8.lock ->&bat_priv->forw_bat_list_lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->&bat_priv->tt.commit_lock ->&bat_priv->tvlv.container_list_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->kfence_freelist_lock ->&n->list_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock FD: 12 BD: 53 +...: &bat_priv->forw_bat_list_lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#68 FD: 30 BD: 1 ..-.: &(&bat_priv->mcast.work)->timer FD: 37 BD: 6 +.+.: (work_completion)(&(&bat_priv->mcast.work)->work) ->pool_lock#2 ->&bat_priv->mcast.mla_lock ->&obj_hash[i].lock ->&base->lock ->rcu_node_0 ->&rq->__lock ->kfence_freelist_lock ->&meta->lock ->&cfs_rq->removed.lock ->quarantine_lock FD: 22 BD: 7 +.+.: &bat_priv->mcast.mla_lock ->pool_lock#2 ->key#16 ->&bat_priv->tt.changes_list_lock ->&bat_priv->tvlv.container_list_lock ->&obj_hash[i].lock ->&base->lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->&meta->lock ->kfence_freelist_lock FD: 1 BD: 49 +...: _xmit_NONE FD: 1 BD: 49 +...: lock#9 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#216 FD: 1 BD: 50 ...-: &____s->seqcount#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#74 FD: 1 BD: 61 +.-.: &hsr->list_lock FD: 30 BD: 1 ..-.: &(&bat_priv->orig_work)->timer FD: 29 BD: 6 +.+.: (work_completion)(&(&bat_priv->orig_work)->work) ->key#19 ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->&cfs_rq->removed.lock ->rcu_node_0 ->&rcu_state.expedited_wq ->pool_lock#2 FD: 1 BD: 51 +...: key#19 FD: 30 BD: 1 ..-.: drivers/net/wireguard/ratelimiter.c:20 FD: 29 BD: 2 +.+.: (gc_work).work ->tk_core.seq.seqcount ->"ratelimiter_table_lock" ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->&cfs_rq->removed.lock ->rcu_node_0 ->pool_lock#2 ->&rcu_state.expedited_wq FD: 1 BD: 3 +.+.: "ratelimiter_table_lock" FD: 9 BD: 57 +...: &vlan_netdev_addr_lock_key ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 15 BD: 49 +.-.: (&app->join_timer) ->&app->lock ->&list->lock#10 ->batched_entropy_u32.lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 50 +.-.: &app->lock FD: 1 BD: 50 ..-.: &list->lock#10 FD: 15 BD: 49 +.-.: (&app->join_timer)#2 ->&app->lock#2 ->&list->lock#11 FD: 13 BD: 51 +.-.: &app->lock#2 ->batched_entropy_u32.lock ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 50 ..-.: &list->lock#11 FD: 9 BD: 57 +...: &macvlan_netdev_addr_lock_key ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 9 BD: 57 +...: &dev_addr_list_lock_key#3 ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock FD: 1 BD: 49 ....: &xa->xa_lock#18 FD: 30 BD: 1 ..-.: &(&forw_packet_aggr->delayed_work)->timer FD: 42 BD: 50 +.+.: (work_completion)(&(&forw_packet_aggr->delayed_work)->work) ->&hard_iface->bat_iv.ogm_buff_mutex ->&bat_priv->forw_bat_list_lock ->&obj_hash[i].lock ->pool_lock#2 ->&c->lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&____s->seqcount#2 ->&____s->seqcount ->&meta->lock ->&n->list_lock ->quarantine_lock ->&rq->__lock ->&cfs_rq->removed.lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&base->lock FD: 16 BD: 57 +...: &dev_addr_list_lock_key#3/1 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&n->list_lock FD: 2 BD: 49 +.+.: &tap_major->minor_lock ->pool_lock#2 FD: 3 BD: 49 +.+.: subsys mutex#75 ->&k->k_lock FD: 744 BD: 1 .+.+: kn->active#46 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->nsim_bus_dev_list_lock ->&c->lock ->&____s->seqcount ->nsim_bus_dev_list_lock.wait_lock ->&p->pi_lock ->&rq->__lock ->&____s->seqcount#2 ->&n->list_lock FD: 743 BD: 9 +.+.: nsim_bus_dev_list_lock ->fs_reclaim ->pool_lock#2 ->nsim_bus_dev_ids.xa_lock ->&x->wait#9 ->&obj_hash[i].lock ->&k->list_lock ->lock ->&root->kernfs_rwsem ->&sem->wait_lock ->&p->pi_lock ->bus_type_sem ->&c->lock ->&____s->seqcount ->sysfs_symlink_target_lock ->&k->k_lock ->&dev->power.lock ->dpm_list_mtx ->uevent_sock_mutex ->uevent_sock_mutex.wait_lock ->device_links_lock ->&rq->__lock ->nsim_bus_dev_list_lock.wait_lock ->deferred_probe_mutex ->subsys mutex#76 ->dev_pm_qos_sysfs_mtx ->kernfs_idr_lock ->mmu_notifier_invalidate_range_start ->&lock->wait_lock ->&____s->seqcount#2 ->&n->list_lock ->&cfs_rq->removed.lock ->remove_cache_srcu ->quarantine_lock FD: 744 BD: 1 .+.+: kn->active#47 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->nsim_bus_dev_list_lock ->nsim_bus_dev_list_lock.wait_lock ->&p->pi_lock ->&rq->__lock ->&c->lock ->&n->list_lock FD: 1 BD: 10 ....: nsim_bus_dev_ids.xa_lock FD: 10 BD: 18 +.+.: devlinks.xa_lock ->pool_lock#2 ->&c->lock ->&n->list_lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#168 FD: 10 BD: 18 +.+.: &xa->xa_lock#19 ->pool_lock#2 ->&c->lock ->&n->list_lock ->&obj_hash[i].lock FD: 1 BD: 10 +.+.: nsim_bus_dev_list_lock.wait_lock FD: 1 BD: 87 +...: &data->fib_event_queue_lock FD: 58 BD: 2 +.+.: (work_completion)(&data->fib_event_work) ->&data->fib_event_queue_lock ->&data->fib_lock ->&rq->__lock FD: 56 BD: 3 +.+.: &data->fib_lock ->fs_reclaim ->&____s->seqcount ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->(&timer.timer) ->remove_cache_srcu ->batched_entropy_u8.lock ->kfence_freelist_lock ->pool_lock ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&cfs_rq->removed.lock ->&meta->lock ->quarantine_lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 18 ....: &(&fn_net->fib_chain)->lock FD: 54 BD: 53 +.+.: bpf_devs_lock ->fs_reclaim ->&c->lock ->pool_lock#2 ->&obj_hash[i].lock ->rcu_node_0 ->&rq->__lock ->&n->list_lock ->remove_cache_srcu FD: 25 BD: 49 +.+.: (work_completion)(&(&devlink_port->type_warn_dw)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 49 +...: &devlink_port->type_lock FD: 1 BD: 49 +.+.: &vn->sock_lock FD: 1 BD: 10 +.+.: subsys mutex#76 FD: 1 BD: 3 ....: rlock-AF_AX25 FD: 30 BD: 1 ..-.: &(&nsim_dev->trap_data->trap_report_dw)->timer FD: 33 BD: 2 +.+.: (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->&cfs_rq->removed.lock FD: 30 BD: 20 +.+.: &nsim_trap_data->trap_lock ->pool_lock#2 ->crngs.lock ->&nsim_dev->fa_cookie_lock ->&obj_hash[i].lock ->&c->lock ->&____s->seqcount ->&n->list_lock ->quarantine_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&meta->lock ->&____s->seqcount#2 ->&base->lock FD: 1 BD: 21 +...: &nsim_dev->fa_cookie_lock FD: 30 BD: 1 ..-.: &(&hwstats->traffic_dw)->timer FD: 26 BD: 2 +.+.: (work_completion)(&(&hwstats->traffic_dw)->work) ->&hwstats->hwsdev_list_lock ->&obj_hash[i].lock ->&base->lock ->&rq->__lock FD: 25 BD: 50 +.+.: &hwstats->hwsdev_list_lock ->&rq->__lock FD: 30 BD: 1 ..-.: &(&hdev->cmd_timer)->timer FD: 41 BD: 24 +.+.: (work_completion)(&(&hdev->cmd_timer)->work) ->(console_sem).lock ->console_owner_lock ->console_owner ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 337 BD: 49 +.+.: &wg->device_update_lock ->&wg->static_identity.lock ->fs_reclaim ->&____s->seqcount ->&c->lock ->pool_lock#2 ->pcpu_alloc_mutex ->&handshake->lock ->&obj_hash[i].lock ->tk_core.seq.seqcount ->&table->lock ->&peer->endpoint_lock ->pool_lock ->&zone->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#8 ->&dir->lock ->k-slock-AF_INET/1 ->k-sk_lock-AF_INET ->k-slock-AF_INET#2 ->cpu_hotplug_lock ->k-sk_lock-AF_INET6 ->k-slock-AF_INET6 ->&wg->socket_update_lock ->rcu_node_0 ->&rq->__lock ->&list->lock#14 ->&rnp->exp_wq[3] ->&rnp->exp_wq[0] ->&rnp->exp_wq[1] ->&rnp->exp_wq[2] ->&____s->seqcount#2 ->&x->wait#2 ->&table->hash[i].lock ->k-clock-AF_INET ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->k-clock-AF_INET6 ->(&peer->timer_retransmit_handshake) ->&base->lock ->(&peer->timer_send_keepalive) ->(&peer->timer_new_handshake) ->(&peer->timer_zero_key_material) ->(&peer->timer_persistent_keepalive) ->(work_completion)(&peer->clear_peer_work) ->&peer->keypairs.keypair_update_lock ->&wq->mutex ->napi_hash_lock ->&table->lock#2 ->wq_pool_mutex ->wq_mayday_lock ->&p->pi_lock ->&x->wait ->pcpu_lock ->&r->consumer_lock#2 ->rcu_state.barrier_mutex ->init_lock ->&cfs_rq->removed.lock ->&n->list_lock ->&rnp->exp_lock ->rcu_state.exp_mutex ->rcu_state.exp_mutex.wait_lock ->quarantine_lock ->rcu_state.barrier_mutex.wait_lock ->&rcu_state.expedited_wq ->batched_entropy_u8.lock ->kfence_freelist_lock ->remove_cache_srcu ->(wq_completion)wg-crypt-wg2#103 ->(wq_completion)wg-kex-wg2#205 ->(wq_completion)wg-kex-wg2#206 ->(wq_completion)wg-crypt-wg1#103 ->(wq_completion)wg-kex-wg1#205 ->(wq_completion)wg-kex-wg1#206 ->(wq_completion)wg-crypt-wg0#106 ->(wq_completion)wg-kex-wg0#207 ->(wq_completion)wg-kex-wg0#208 ->(wq_completion)wg-crypt-wg2#102 ->(wq_completion)wg-kex-wg2#203 ->(wq_completion)wg-kex-wg2#204 ->(wq_completion)wg-crypt-wg1#102 ->(wq_completion)wg-kex-wg1#203 ->(wq_completion)wg-kex-wg1#204 ->(wq_completion)wg-crypt-wg0#105 ->(wq_completion)wg-kex-wg0#205 ->(wq_completion)wg-kex-wg0#206 ->(wq_completion)wg-crypt-wg2#104 ->(wq_completion)wg-kex-wg2#208 ->(wq_completion)wg-kex-wg2#207 ->(wq_completion)wg-crypt-wg1#104 ->(wq_completion)wg-kex-wg1#207 ->(wq_completion)wg-kex-wg1#208 ->(wq_completion)wg-crypt-wg0#107 ->(wq_completion)wg-kex-wg0#209 ->(wq_completion)wg-kex-wg0#210 ->(wq_completion)wg-crypt-wg2#96 ->(wq_completion)wg-kex-wg2#191 ->(wq_completion)wg-kex-wg2#192 ->(wq_completion)wg-crypt-wg1#96 ->(wq_completion)wg-kex-wg1#191 ->(wq_completion)wg-kex-wg1#192 ->(wq_completion)wg-crypt-wg0#99 ->(wq_completion)wg-kex-wg0#193 ->(wq_completion)wg-kex-wg0#194 ->(wq_completion)wg-crypt-wg2#100 ->(wq_completion)wg-kex-wg2#199 ->(wq_completion)wg-kex-wg2#200 ->(wq_completion)wg-crypt-wg1#100 ->(wq_completion)wg-kex-wg1#199 ->(wq_completion)wg-kex-wg1#200 ->(wq_completion)wg-crypt-wg0#103 ->(wq_completion)wg-kex-wg0#201 ->(wq_completion)wg-kex-wg0#202 FD: 52 BD: 142 ++++: &wg->static_identity.lock ->&handshake->lock ->&rq->__lock ->&sem->wait_lock ->&p->pi_lock ->stock_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock#2 ->&cfs_rq->removed.lock ->rcu_node_0 FD: 51 BD: 143 ++++: &handshake->lock ->&rq->__lock ->crngs.lock ->tk_core.seq.seqcount ->&table->lock#2 ->fs_reclaim ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&n->list_lock ->&sem->wait_lock ->&____s->seqcount#2 ->batched_entropy_u8.lock ->kfence_freelist_lock ->remove_cache_srcu ->&cfs_rq->removed.lock ->rcu_node_0 ->&obj_hash[i].lock FD: 25 BD: 50 +.+.: &table->lock ->&rq->__lock FD: 53 BD: 144 ++-.: &peer->endpoint_lock ->pool_lock#2 ->&obj_hash[i].lock FD: 1 BD: 49 +...: _xmit_SIT FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#175 FD: 16 BD: 57 +...: &bridge_netdev_addr_lock_key/1 ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->&n->list_lock ->&____s->seqcount#2 FD: 1 BD: 49 +...: _xmit_TUNNEL FD: 16 BD: 49 +...: _xmit_IPGRE ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&c->lock ->&n->list_lock FD: 39 BD: 49 +.-.: (&brmctx->ip6_own_query.timer) ->&br->multicast_lock FD: 38 BD: 68 +.-.: &br->multicast_lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&____s->seqcount ->&dir->lock#2 ->deferred_lock ->&c->lock ->nl_table_lock ->nl_table_wait.lock ->&n->list_lock ->&____s->seqcount#2 ->quarantine_lock FD: 39 BD: 49 +.-.: (&brmctx->ip4_own_query.timer) ->&br->multicast_lock FD: 55 BD: 1 +.-.: (&in_dev->mr_ifc_timer) ->&obj_hash[i].lock ->batched_entropy_u32.lock ->&base->lock FD: 30 BD: 1 ..-.: &(&br->gc_work)->timer FD: 12 BD: 50 +.+.: (work_completion)(&(&br->gc_work)->work) ->&obj_hash[i].lock ->&base->lock FD: 1 BD: 49 +...: _xmit_TUNNEL6 FD: 48 BD: 52 +.-.: _xmit_TUNNEL6#2 ->&obj_hash[i].lock ->pool_lock#2 ->&base->lock FD: 1 BD: 5 +.+.: &pnetids_ndev->lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#495 ->&rq->__lock FD: 14 BD: 49 +.-.: (&app->periodic_timer) ->&app->lock#2 FD: 43 BD: 57 +...: &dev_addr_list_lock_key/1 ->_xmit_ETHER ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock FD: 43 BD: 57 +...: &dev_addr_list_lock_key#2/1 ->_xmit_ETHER ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock FD: 39 BD: 1 +.-.: (&pmctx->ip6_own_query.timer) ->&br->multicast_lock FD: 39 BD: 1 +.-.: (&pmctx->ip4_own_query.timer) ->&br->multicast_lock FD: 1 BD: 179 ..-.: &list->lock#12 FD: 43 BD: 57 +...: _xmit_ETHER/1 ->_xmit_ETHER ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->pool_lock#2 ->&n->list_lock FD: 21 BD: 59 +.-.: &hsr->seqnr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&____s->seqcount ->&meta->lock ->kfence_freelist_lock ->&base->lock ->quarantine_lock FD: 1 BD: 60 +.-.: &new_node->seq_out_lock FD: 22 BD: 49 +.-.: (&hsr->announce_timer) FD: 1 BD: 49 +.+.: &nn->netlink_tap_lock FD: 16 BD: 57 +...: &batadv_netdev_addr_lock_key/1 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->&n->list_lock FD: 43 BD: 57 +...: &vlan_netdev_addr_lock_key/1 ->_xmit_ETHER ->pool_lock#2 ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->&n->list_lock ->&____s->seqcount#2 FD: 43 BD: 57 +...: &macvlan_netdev_addr_lock_key/1 ->_xmit_ETHER ->&c->lock ->&____s->seqcount ->&obj_hash[i].lock ->krc.lock ->&n->list_lock ->&____s->seqcount#2 FD: 16 BD: 50 +...: &ipvlan->addrs_lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock FD: 1 BD: 51 +.-.: &list->lock#13 FD: 30 BD: 50 +.+.: (work_completion)(&port->bc_work) ->&list->lock#13 ->&obj_hash[i].lock ->pool_lock#2 ->rcu_node_0 ->&rq->__lock ->&meta->lock ->kfence_freelist_lock ->quarantine_lock ->&base->lock FD: 43 BD: 57 +...: &macsec_netdev_addr_lock_key/1 ->_xmit_ETHER ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock ->pool_lock#2 ->&____s->seqcount#2 ->&____s->seqcount FD: 12 BD: 49 +...: dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 ->&sch->q.lock FD: 16 BD: 59 +...: key#20 ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock FD: 23 BD: 53 +...: &bat_priv->tt.commit_lock ->key#16 ->&bat_priv->softif_vlan_list_lock ->&bat_priv->tt.changes_list_lock ->&bat_priv->tt.last_changeset_lock ->pool_lock#2 ->&bat_priv->tvlv.container_list_lock ->&obj_hash[i].lock ->&c->lock ->&meta->lock ->kfence_freelist_lock ->&n->list_lock FD: 1 BD: 59 +...: &entry->crc_lock FD: 1 BD: 50 +.+.: &wg->socket_update_lock FD: 4 BD: 114 +.-.: &list->lock#14 ->&obj_hash[i].lock ->pool_lock#2 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#836 ->&rq->__lock FD: 52 BD: 80 +.+.: (work_completion)(&peer->transmit_handshake_work) ->tk_core.seq.seqcount ->&wg->static_identity.lock ->&cookie->lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&peer->endpoint_lock ->batched_entropy_u8.lock ->&c->lock ->&____s->seqcount ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&____s->seqcount#2 ->&n->list_lock ->&cfs_rq->removed.lock ->kfence_freelist_lock FD: 1 BD: 145 +...: &table->lock#2 FD: 25 BD: 142 ++++: &cookie->lock ->&rq->__lock FD: 76 BD: 1 +.-.: (&sk->sk_timer) ->slock-AF_INET#2 FD: 1 BD: 1 ....: &rs->rs_rdma_lock FD: 1 BD: 114 +.-.: &r->producer_lock#2 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1054 ->&rq->__lock FD: 86 BD: 110 +.+.: (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->&r->consumer_lock#2 ->&wg->static_identity.lock ->&peer->endpoint_lock ->tk_core.seq.seqcount ->&cookie->lock ->&handshake->lock ->&obj_hash[i].lock ->&base->lock ->pool_lock#2 ->&rq->__lock ->&c->lock ->&____s->seqcount ->&list->lock#14 ->&____s->seqcount#2 ->&n->list_lock ->rcu_node_0 ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock ->&meta->lock ->kfence_freelist_lock ->&sem->wait_lock ->&p->pi_lock FD: 1 BD: 111 +.+.: &r->consumer_lock#2 FD: 5 BD: 144 +.-.: &peer->keypairs.keypair_update_lock ->&table->lock#2 ->&obj_hash[i].lock ->pool_lock#2 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#578 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#340 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#743 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#886 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#898 ->&rq->__lock FD: 1 BD: 1 +.+.: &xs->mutex FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#472 ->&rq->__lock FD: 26 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#514 ->rcu_node_0 ->&rq->__lock FD: 28 BD: 80 +.+.: (work_completion)(&peer->transmit_packet_work) ->&obj_hash[i].lock ->&base->lock ->&peer->endpoint_lock ->batched_entropy_u8.lock ->&rq->__lock ->&cfs_rq->removed.lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#406 FD: 1 BD: 1 +.-.: &keypair->receiving_counter.lock FD: 1 BD: 2 +.+.: loop_validate_mutex.wait_lock FD: 50 BD: 3 +.+.: smk_net6addr_lock ->fs_reclaim ->pool_lock#2 ->&c->lock FD: 25 BD: 14 +.+.: &data->mtx ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1058 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#507 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#979 FD: 1 BD: 4 +.+.: genl_mutex.wait_lock FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#95 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (work_completion)(&td->dispatch_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1068 FD: 11 BD: 51 ++.-: &priv->lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock ->quarantine_lock FD: 1 BD: 1 ....: &lo->lo_work_lock FD: 1 BD: 3 +...: clock-AF_QIPCRTR FD: 1 BD: 53 ....: &wdev->event_lock FD: 1 BD: 53 ....: (&dwork->timer) FD: 1 BD: 53 ....: (&dwork->timer)#2 FD: 1 BD: 53 +.+.: (work_completion)(&(&link->color_collision_detect_work)->work) FD: 1 BD: 54 ..-.: &list->lock#15 FD: 1 BD: 53 +.-.: &ifibss->incomplete_lock FD: 726 BD: 1 +.+.: (wq_completion)cfg80211 ->(work_completion)(&rdev->event_work) ->(work_completion)(&(&rdev->dfs_update_channels_wk)->work) ->(work_completion)(&barr->work) FD: 376 BD: 6 +.+.: (work_completion)(&rdev->event_work) ->&rdev->wiphy.mtx ->&lock->wait_lock ->&p->pi_lock ->&rq->__lock FD: 39 BD: 2 +.+.: wireless_nlevent_work ->net_rwsem FD: 25 BD: 1 +.+.: (wq_completion)nfc18_nci_tx_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1007 FD: 1 BD: 3 ....: &tags->lock FD: 10 BD: 54 +...: &bat_priv->tt.last_changeset_lock ->pool_lock#2 ->&c->lock ->&obj_hash[i].lock ->&n->list_lock FD: 1 BD: 5 +.+.: &hn->hn_lock FD: 1 BD: 77 +.-.: &local->active_txq_lock[i] FD: 39 BD: 76 +.-.: &local->handle_wake_tx_queue_lock ->&local->active_txq_lock[i] ->&local->queue_stop_reason_lock ->&fq->lock ->tk_core.seq.seqcount ->hwsim_radio_lock ->&list->lock#16 FD: 1 BD: 77 ..-.: &local->queue_stop_reason_lock FD: 1 BD: 79 ..-.: &list->lock#16 FD: 1 BD: 5 +.+.: (work_completion)(&(&bat_priv->bat_v.ogm_wq)->work) FD: 28 BD: 49 +.+.: &caifn->caifdevs.lock ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->pool_lock#2 ->&this->info_list_lock FD: 4 BD: 5 +.+.: &bat_priv->bat_v.ogm_buff_mutex ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 5 +...: &bat_priv->gw.list_lock FD: 1 BD: 51 +...: &this->info_list_lock FD: 59 BD: 2 +.+.: (work_completion)(&fqdir->destroy_work) ->(work_completion)(&ht->run_work) ->&ht->mutex FD: 41 BD: 2 +.+.: fqdir_free_work ->rcu_state.barrier_mutex ->&obj_hash[i].lock ->pool_lock#2 ->rcu_state.barrier_mutex.wait_lock ->&p->pi_lock ->quarantine_lock FD: 60 BD: 1 +.+.: (wq_completion)inet_frag_wq ->(work_completion)(&fqdir->destroy_work) FD: 33 BD: 1 +.-.: &local->rx_path_lock ->&obj_hash[i].lock ->pool_lock#2 ->&list->lock#15 ->&rdev->wiphy_work_lock ->quarantine_lock FD: 16 BD: 53 +...: &sta->lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock FD: 16 BD: 53 +...: &sta->rate_ctrl_lock ->pool_lock#2 ->&obj_hash[i].lock ->krc.lock ->&c->lock ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 82 BD: 1 +.+.: &type->s_umount_key#46/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#32 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount ->binderfs_minors_mutex ->&dentry->d_lock ->&sb->s_type->i_mutex_key#18 ->&____s->seqcount#2 ->&n->list_lock ->&rq->__lock FD: 38 BD: 5 +.+.: &sb->s_type->i_lock_key#32 ->&dentry->d_lock FD: 26 BD: 3 +.+.: binderfs_minors_mutex ->binderfs_minors.xa_lock ->&rq->__lock FD: 1 BD: 4 ....: binderfs_minors.xa_lock FD: 66 BD: 3 +.+.: &sb->s_type->i_mutex_key#18 ->&sb->s_type->i_lock_key#32 ->rename_lock.seqcount ->fs_reclaim ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->rcu_node_0 ->&rq->__lock ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock FD: 1 BD: 17 +.+.: iunique_lock FD: 667 BD: 2 +.+.: &type->i_mutex_dir_key#6/1 ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->&root->kernfs_rwsem ->tomoyo_ss ->&root->kernfs_iattr_rwsem ->cgroup_mutex FD: 52 BD: 1 .+.+: kn->active#48 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->&rq->__lock ->&____s->seqcount#2 ->&____s->seqcount FD: 51 BD: 1 ++++: kn->active#49 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->remove_cache_srcu FD: 3 BD: 93 ..-.: cgroup_threadgroup_rwsem.rss.gp_wait.lock ->&obj_hash[i].lock FD: 13 BD: 49 +.-.: (&hsr->prune_timer) ->&hsr->list_lock ->&obj_hash[i].lock ->&base->lock FD: 30 BD: 1 ..-.: &(&conn->info_timer)->timer FD: 2 BD: 28 +.+.: (work_completion)(&(&conn->info_timer)->work) ->&conn->chan_lock FD: 26 BD: 93 ....: cgroup_threadgroup_rwsem.waiters.lock ->&p->pi_lock FD: 1 BD: 18 +.+.: (wq_completion)cpuset_migrate_mm FD: 667 BD: 2 +.+.: &type->i_mutex_dir_key#7/1 ->rename_lock.seqcount ->fs_reclaim ->&dentry->d_lock ->&root->kernfs_rwsem ->tomoyo_ss ->&root->kernfs_iattr_rwsem ->cgroup_mutex ->pool_lock#2 ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->cgroup_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 50 BD: 1 ++++: kn->active#50 ->fs_reclaim ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&kernfs_locks->open_file_mutex[count] ->stock_lock ->&n->list_lock ->remove_cache_srcu ->&rq->__lock FD: 1 BD: 94 ....: cpuset_attach_wq.lock FD: 2 BD: 652 ..-.: stock_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) FD: 50 BD: 1 .+.+: kn->active#51 ->fs_reclaim ->stock_lock ->pool_lock#2 ->&kernfs_locks->open_file_mutex[count] ->&c->lock ->&n->list_lock ->remove_cache_srcu ->&____s->seqcount#2 ->&____s->seqcount FD: 53 BD: 1 .+.+: kn->active#52 ->fs_reclaim ->stock_lock ->&c->lock ->&kernfs_locks->open_file_mutex[count] ->memcg_max_mutex ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->&rq->__lock FD: 1 BD: 8 +.+.: memcg_max_mutex FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#233 FD: 1 BD: 6 ....: &per_cpu(xt_recseq, i) FD: 228 BD: 5 +.+.: nf_nat_proto_mutex ->fs_reclaim ->pool_lock#2 ->nf_hook_mutex ->cpu_hotplug_lock ->&obj_hash[i].lock ->stock_lock ->&c->lock ->&____s->seqcount ->&____s->seqcount#2 ->krc.lock ->&n->list_lock ->&rq->__lock ->nf_nat_proto_mutex.wait_lock ->quarantine_lock FD: 31 BD: 1 +.+.: loop_validate_mutex ->&lo->lo_mutex ->&rq->__lock ->loop_validate_mutex.wait_lock ->stock_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock#2 ->&cfs_rq->removed.lock FD: 2 BD: 1 +.-.: (&tun->flow_gc_timer) ->&tun->lock FD: 1 BD: 50 +.-.: &tun->lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#802 ->&rq->__lock FD: 1 BD: 53 +.+.: (work_completion)(&sta->drv_deliver_wk) FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg2#213 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 2 +...: clock-AF_CAIF FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#871 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#412 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#976 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#144 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#410 ->&rq->__lock FD: 30 BD: 1 +.-.: (&sdp->delay_work) FD: 787 BD: 1 +.+.: &fc->uapi_mutex ->&rq->__lock ->fs_reclaim ->&c->lock ->pool_lock#2 ->sb_lock ->&type->s_umount_key#65 ->remove_cache_srcu ->&n->list_lock ->&type->s_umount_key#67/1 ->&____s->seqcount#2 ->&____s->seqcount ->&type->s_umount_key#45 ->&type->s_umount_key#63 ->&obj_hash[i].lock ->&type->s_umount_key#28/1 ->crngs.lock ->krc.lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->&type->s_umount_key#69/1 ->unnamed_dev_ida.xa_lock ->list_lrus_mutex ->&xa->xa_lock#15 ->&type->s_umount_key#44 ->v9fs_trans_lock ->&type->s_umount_key/1 ->&type->s_umount_key#18/1 ->&type->s_umount_key#6 ->uuid_mutex ->&type->s_umount_key#33 ->&type->s_umount_key#27/1 ->(console_sem).lock ->&type->s_umount_key#70/1 ->&type->s_umount_key#3/1 ->&type->s_umount_key#72/1 ->stock_lock ->mnt_id_ida.xa_lock ->pcpu_alloc_mutex ->&dentry->d_lock ->mount_lock ->ucounts_lock ->&newf->file_lock ->bit_wait_table + i ->&lock->wait_lock ->&type->s_umount_key#68 ->&type->s_umount_key#32 ->&type->s_umount_key#74/1 ->&type->s_umount_key#39/1 ->&type->s_umount_key#14 ->&type->s_umount_key#76/1 ->cgroup_mutex ->virtio_fs_mutex FD: 184 BD: 49 +.+.: sk_lock-AF_CAN ->slock-AF_CAN ->&mm->mmap_lock ->clock-AF_CAN ->&rq->__lock ->pcpu_lock ->&jsk->filters_lock ->&obj_hash[i].lock ->pool_lock#2 ->j1939_netdev_lock ->fs_reclaim ->&c->lock ->&priv->lock ->&priv->j1939_socks_lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->&jsk->sk_session_queue_lock ->(console_sem).lock ->remove_cache_srcu ->stock_lock ->&f->f_lock ->rcu_node_0 ->&list->lock#21 ->&priv->active_session_list_lock ->hrtimer_bases.lock ->&jsk->waitq FD: 27 BD: 50 +...: slock-AF_CAN ->&sk->sk_lock.wq FD: 1 BD: 59 .+.-: &table->lock#3 FD: 165 BD: 2 +.+.: sk_lock-AF_X25 ->slock-AF_X25 ->&mm->mmap_lock ->&obj_hash[i].lock ->x25_list_lock ->rlock-AF_X25 ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq ->fs_reclaim ->stock_lock ->&sem->wait_lock ->&p->pi_lock ->&f->f_lock FD: 1 BD: 3 +...: slock-AF_X25 FD: 1 BD: 125 +.-.: &nf_nat_locks[i] FD: 1 BD: 50 +...: clock-AF_CAN FD: 1 BD: 3 +.-.: x25_list_lock FD: 1 BD: 3 ....: rlock-AF_X25 FD: 59 BD: 2 +.+.: sk_lock-AF_NFC ->slock-AF_NFC ->&rq->__lock ->&k->list_lock ->&k->k_lock ->llcp_devices_lock ->fs_reclaim ->pool_lock#2 ->&local->sdp_lock ->&local->sockets.lock ->&c->lock ->&local->raw_sockets.lock ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 27 BD: 3 +...: slock-AF_NFC ->&sk->sk_lock.wq FD: 1919 BD: 1 +.+.: &ndev->req_lock ->&wq->mutex ->(&ndev->cmd_timer) ->&obj_hash[i].lock ->&base->lock ->(&ndev->data_timer) ->&rq->__lock ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock#2 ->(wq_completion)nfc2_nci_cmd_wq#394 ->(wq_completion)nfc3_nci_cmd_wq#135 ->(wq_completion)nfc2_nci_cmd_wq#395 ->(wq_completion)nfc3_nci_cmd_wq#136 ->(wq_completion)nfc4_nci_cmd_wq#45 ->(wq_completion)nfc2_nci_cmd_wq#396 ->(wq_completion)nfc3_nci_cmd_wq#137 ->(wq_completion)nfc2_nci_cmd_wq#397 ->(wq_completion)nfc3_nci_cmd_wq#138 ->(wq_completion)nfc2_nci_cmd_wq#398 ->(wq_completion)nfc2_nci_cmd_wq#399 ->(wq_completion)nfc3_nci_cmd_wq#139 ->(wq_completion)nfc2_nci_cmd_wq#400 ->(wq_completion)nfc2_nci_cmd_wq#401 ->(wq_completion)nfc3_nci_cmd_wq#140 ->(wq_completion)nfc2_nci_cmd_wq#402 ->(wq_completion)nfc2_nci_cmd_wq#403 ->(wq_completion)nfc2_nci_cmd_wq#404 ->(wq_completion)nfc3_nci_cmd_wq#141 ->(wq_completion)nfc2_nci_cmd_wq#405 ->(wq_completion)nfc2_nci_cmd_wq#406 ->(wq_completion)nfc2_nci_cmd_wq#407 ->(wq_completion)nfc3_nci_cmd_wq#142 ->(wq_completion)nfc4_nci_cmd_wq#46 ->(wq_completion)nfc2_nci_cmd_wq#408 ->(wq_completion)nfc3_nci_cmd_wq#143 ->(wq_completion)nfc3_nci_cmd_wq#144 ->(wq_completion)nfc2_nci_cmd_wq#409 ->(wq_completion)nfc4_nci_cmd_wq#47 ->(wq_completion)nfc3_nci_cmd_wq#145 ->(wq_completion)nfc2_nci_cmd_wq#410 ->(wq_completion)nfc2_nci_cmd_wq#411 ->(wq_completion)nfc3_nci_cmd_wq#146 ->(wq_completion)nfc4_nci_cmd_wq#50 ->(wq_completion)nfc2_nci_cmd_wq#412 ->(wq_completion)nfc3_nci_cmd_wq#147 ->(wq_completion)nfc2_nci_cmd_wq#413 ->(wq_completion)nfc4_nci_cmd_wq#51 ->(wq_completion)nfc2_nci_cmd_wq#414 ->(wq_completion)nfc2_nci_cmd_wq#415 ->(wq_completion)nfc3_nci_cmd_wq#148 ->(wq_completion)nfc2_nci_cmd_wq#416 ->(wq_completion)nfc3_nci_cmd_wq#149 ->(wq_completion)nfc4_nci_cmd_wq#52 ->(wq_completion)nfc2_nci_cmd_wq#417 ->(wq_completion)nfc2_nci_cmd_wq#418 ->(wq_completion)nfc3_nci_cmd_wq#150 ->(wq_completion)nfc2_nci_cmd_wq#419 ->(wq_completion)nfc3_nci_cmd_wq#151 ->(wq_completion)nfc2_nci_cmd_wq#420 ->(wq_completion)nfc2_nci_cmd_wq#421 ->(wq_completion)nfc3_nci_cmd_wq#152 ->(wq_completion)nfc2_nci_cmd_wq#422 ->(wq_completion)nfc2_nci_cmd_wq#423 ->(wq_completion)nfc2_nci_cmd_wq#424 ->(wq_completion)nfc2_nci_cmd_wq#425 ->(wq_completion)nfc2_nci_cmd_wq#426 ->(wq_completion)nfc2_nci_cmd_wq#427 ->(wq_completion)nfc2_nci_cmd_wq#428 ->(wq_completion)nfc3_nci_cmd_wq#153 ->(wq_completion)nfc2_nci_cmd_wq#429 ->(wq_completion)nfc2_nci_cmd_wq#430 ->(wq_completion)nfc4_nci_cmd_wq#53 ->(wq_completion)nfc3_nci_cmd_wq#154 ->(wq_completion)nfc5_nci_cmd_wq#24 ->(wq_completion)nfc3_nci_cmd_wq#155 ->(wq_completion)nfc2_nci_cmd_wq#431 ->(wq_completion)nfc5_nci_cmd_wq#25 ->(wq_completion)nfc4_nci_cmd_wq#54 ->(wq_completion)nfc3_nci_cmd_wq#156 ->(wq_completion)nfc2_nci_cmd_wq#432 ->(wq_completion)nfc2_nci_cmd_wq#433 ->(wq_completion)nfc2_nci_cmd_wq#434 ->(wq_completion)nfc2_nci_cmd_wq#435 ->(wq_completion)nfc3_nci_cmd_wq#157 ->(wq_completion)nfc2_nci_cmd_wq#436 ->(wq_completion)nfc4_nci_cmd_wq#55 ->(wq_completion)nfc5_nci_cmd_wq#26 ->(wq_completion)nfc3_nci_cmd_wq#158 ->(wq_completion)nfc2_nci_cmd_wq#437 ->(wq_completion)nfc3_nci_cmd_wq#159 ->(wq_completion)nfc2_nci_cmd_wq#438 ->(wq_completion)nfc4_nci_cmd_wq#56 ->(wq_completion)nfc3_nci_cmd_wq#160 ->(wq_completion)nfc2_nci_cmd_wq#439 ->(wq_completion)nfc5_nci_cmd_wq#28 ->(wq_completion)nfc4_nci_cmd_wq#57 ->(wq_completion)nfc2_nci_cmd_wq#440 ->(wq_completion)nfc3_nci_cmd_wq#161 ->(wq_completion)nfc4_nci_cmd_wq#58 ->(wq_completion)nfc2_nci_cmd_wq#441 ->(wq_completion)nfc2_nci_cmd_wq#442 ->(wq_completion)nfc4_nci_cmd_wq#59 ->(wq_completion)nfc3_nci_cmd_wq#162 ->(wq_completion)nfc5_nci_cmd_wq#29 ->(wq_completion)nfc2_nci_cmd_wq#443 ->(wq_completion)nfc2_nci_cmd_wq#444 ->(wq_completion)nfc2_nci_cmd_wq#445 ->(wq_completion)nfc3_nci_cmd_wq#163 ->(wq_completion)nfc3_nci_cmd_wq#164 ->&cfs_rq->removed.lock ->(wq_completion)nfc2_nci_cmd_wq#446 ->(wq_completion)nfc3_nci_cmd_wq#165 ->(wq_completion)nfc2_nci_cmd_wq#447 ->(wq_completion)nfc3_nci_cmd_wq#166 ->(wq_completion)nfc2_nci_cmd_wq#448 ->(wq_completion)nfc3_nci_cmd_wq#167 ->(wq_completion)nfc4_nci_cmd_wq#60 ->(wq_completion)nfc3_nci_cmd_wq#168 ->(wq_completion)nfc4_nci_cmd_wq#61 ->(wq_completion)nfc3_nci_cmd_wq#169 ->(wq_completion)nfc3_nci_cmd_wq#170 ->(wq_completion)nfc3_nci_cmd_wq#171 ->(wq_completion)nfc3_nci_cmd_wq#172 ->(wq_completion)nfc2_nci_cmd_wq#449 ->(wq_completion)nfc2_nci_cmd_wq#450 ->(wq_completion)nfc3_nci_cmd_wq#173 ->(wq_completion)nfc2_nci_cmd_wq#451 ->(wq_completion)nfc3_nci_cmd_wq#174 ->(wq_completion)nfc2_nci_cmd_wq#452 ->(wq_completion)nfc3_nci_cmd_wq#175 ->(wq_completion)nfc2_nci_cmd_wq#453 ->(wq_completion)nfc4_nci_cmd_wq#62 ->(wq_completion)nfc2_nci_cmd_wq#454 ->(wq_completion)nfc2_nci_cmd_wq#455 ->(wq_completion)nfc3_nci_cmd_wq#176 ->(wq_completion)nfc2_nci_cmd_wq#456 ->(wq_completion)nfc2_nci_cmd_wq#457 ->(wq_completion)nfc3_nci_cmd_wq#177 ->(wq_completion)nfc4_nci_cmd_wq#63 ->(wq_completion)nfc2_nci_cmd_wq#458 ->(wq_completion)nfc2_nci_cmd_wq#459 ->(wq_completion)nfc3_nci_cmd_wq#178 ->(wq_completion)nfc4_nci_cmd_wq#64 ->(wq_completion)nfc2_nci_cmd_wq#460 ->(wq_completion)nfc2_nci_cmd_wq#461 ->(wq_completion)nfc3_nci_cmd_wq#179 ->(wq_completion)nfc4_nci_cmd_wq#65 ->(wq_completion)nfc2_nci_cmd_wq#462 ->(wq_completion)nfc2_nci_cmd_wq#463 ->(wq_completion)nfc3_nci_cmd_wq#180 ->(wq_completion)nfc2_nci_cmd_wq#464 ->(wq_completion)nfc4_nci_cmd_wq#66 ->(wq_completion)nfc3_nci_cmd_wq#181 ->(wq_completion)nfc2_nci_cmd_wq#465 ->(wq_completion)nfc2_nci_cmd_wq#466 ->(wq_completion)nfc4_nci_cmd_wq#67 ->(wq_completion)nfc2_nci_cmd_wq#467 ->(wq_completion)nfc3_nci_cmd_wq#182 ->(wq_completion)nfc2_nci_cmd_wq#468 ->(wq_completion)nfc3_nci_cmd_wq#183 ->(wq_completion)nfc4_nci_cmd_wq#68 ->(wq_completion)nfc3_nci_cmd_wq#184 ->(wq_completion)nfc2_nci_cmd_wq#469 ->(wq_completion)nfc4_nci_cmd_wq#69 ->(wq_completion)nfc2_nci_cmd_wq#470 ->(wq_completion)nfc2_nci_cmd_wq#471 ->(wq_completion)nfc3_nci_cmd_wq#185 ->(wq_completion)nfc2_nci_cmd_wq#472 ->(wq_completion)nfc3_nci_cmd_wq#186 ->(wq_completion)nfc2_nci_cmd_wq#473 ->(wq_completion)nfc3_nci_cmd_wq#187 ->(wq_completion)nfc3_nci_cmd_wq#188 ->(wq_completion)nfc2_nci_cmd_wq#474 ->(wq_completion)nfc4_nci_cmd_wq#70 ->(wq_completion)nfc2_nci_cmd_wq#475 ->(wq_completion)nfc3_nci_cmd_wq#189 ->(wq_completion)nfc2_nci_cmd_wq#476 ->(wq_completion)nfc3_nci_cmd_wq#190 ->(wq_completion)nfc2_nci_cmd_wq#477 ->(wq_completion)nfc4_nci_cmd_wq#71 ->(wq_completion)nfc2_nci_cmd_wq#478 ->(wq_completion)nfc3_nci_cmd_wq#191 ->(wq_completion)nfc5_nci_cmd_wq#30 ->(wq_completion)nfc4_nci_cmd_wq#72 ->(wq_completion)nfc2_nci_cmd_wq#479 ->(wq_completion)nfc2_nci_cmd_wq#480 ->(wq_completion)nfc3_nci_cmd_wq#192 ->(wq_completion)nfc2_nci_cmd_wq#481 ->(wq_completion)nfc2_nci_cmd_wq#482 ->(wq_completion)nfc3_nci_cmd_wq#193 ->(wq_completion)nfc2_nci_cmd_wq#483 ->(wq_completion)nfc2_nci_cmd_wq#484 ->(wq_completion)nfc3_nci_cmd_wq#194 ->(wq_completion)nfc3_nci_cmd_wq#195 ->(wq_completion)nfc2_nci_cmd_wq#485 ->(wq_completion)nfc3_nci_cmd_wq#196 ->(wq_completion)nfc2_nci_cmd_wq#486 ->(wq_completion)nfc3_nci_cmd_wq#197 ->(wq_completion)nfc2_nci_cmd_wq#487 ->(wq_completion)nfc3_nci_cmd_wq#198 ->(wq_completion)nfc4_nci_cmd_wq#73 ->(wq_completion)nfc2_nci_cmd_wq#488 ->(wq_completion)nfc3_nci_cmd_wq#199 ->(wq_completion)nfc2_nci_cmd_wq#489 ->(wq_completion)nfc2_nci_cmd_wq#490 ->(wq_completion)nfc3_nci_cmd_wq#200 ->(wq_completion)nfc2_nci_cmd_wq#491 ->(wq_completion)nfc2_nci_cmd_wq#492 ->(wq_completion)nfc4_nci_cmd_wq#74 ->(wq_completion)nfc5_nci_cmd_wq#31 ->(wq_completion)nfc2_nci_cmd_wq#493 ->(wq_completion)nfc3_nci_cmd_wq#201 ->(wq_completion)nfc4_nci_cmd_wq#75 ->(wq_completion)nfc2_nci_cmd_wq#494 ->(wq_completion)nfc2_nci_cmd_wq#495 ->(wq_completion)nfc4_nci_cmd_wq#76 ->(wq_completion)nfc5_nci_cmd_wq#32 ->(wq_completion)nfc3_nci_cmd_wq#202 ->(wq_completion)nfc2_nci_cmd_wq#496 ->(wq_completion)nfc4_nci_cmd_wq#78 ->(wq_completion)nfc3_nci_cmd_wq#203 ->(wq_completion)nfc2_nci_cmd_wq#497 ->(wq_completion)nfc3_nci_cmd_wq#204 ->(wq_completion)nfc2_nci_cmd_wq#498 ->(wq_completion)nfc3_nci_cmd_wq#205 ->(wq_completion)nfc2_nci_cmd_wq#499 ->(wq_completion)nfc4_nci_cmd_wq#79 ->(wq_completion)nfc2_nci_cmd_wq#500 ->(wq_completion)nfc3_nci_cmd_wq#206 ->(wq_completion)nfc2_nci_cmd_wq#501 ->(wq_completion)nfc4_nci_cmd_wq#80 ->(wq_completion)nfc3_nci_cmd_wq#207 ->(wq_completion)nfc2_nci_cmd_wq#502 ->(wq_completion)nfc2_nci_cmd_wq#503 ->(wq_completion)nfc3_nci_cmd_wq#208 ->(wq_completion)nfc2_nci_cmd_wq#504 ->(wq_completion)nfc2_nci_cmd_wq#505 ->(wq_completion)nfc2_nci_cmd_wq#506 ->(wq_completion)nfc2_nci_cmd_wq#507 ->(wq_completion)nfc2_nci_cmd_wq#508 ->(wq_completion)nfc2_nci_cmd_wq#509 ->(wq_completion)nfc2_nci_cmd_wq#510 ->(wq_completion)nfc2_nci_cmd_wq#511 ->(wq_completion)nfc2_nci_cmd_wq#512 ->(wq_completion)nfc3_nci_cmd_wq#209 ->(wq_completion)nfc4_nci_cmd_wq#81 ->(wq_completion)nfc2_nci_cmd_wq#513 ->(wq_completion)nfc2_nci_cmd_wq#514 ->(wq_completion)nfc3_nci_cmd_wq#210 ->(wq_completion)nfc2_nci_cmd_wq#515 ->(wq_completion)nfc3_nci_cmd_wq#211 ->(wq_completion)nfc4_nci_cmd_wq#82 ->(wq_completion)nfc2_nci_cmd_wq#516 ->(wq_completion)nfc2_nci_cmd_wq#517 ->(wq_completion)nfc3_nci_cmd_wq#212 ->(wq_completion)nfc2_nci_cmd_wq#518 ->(wq_completion)nfc3_nci_cmd_wq#213 ->(wq_completion)nfc2_nci_cmd_wq#519 ->(wq_completion)nfc3_nci_cmd_wq#214 ->(wq_completion)nfc2_nci_cmd_wq#520 ->(wq_completion)nfc3_nci_cmd_wq#215 ->(wq_completion)nfc2_nci_cmd_wq#521 ->(wq_completion)nfc2_nci_cmd_wq#522 ->(wq_completion)nfc3_nci_cmd_wq#216 ->(wq_completion)nfc2_nci_cmd_wq#523 ->(wq_completion)nfc3_nci_cmd_wq#217 ->(wq_completion)nfc2_nci_cmd_wq#524 ->(wq_completion)nfc2_nci_cmd_wq#525 ->(wq_completion)nfc3_nci_cmd_wq#218 ->(wq_completion)nfc3_nci_cmd_wq#219 ->(wq_completion)nfc2_nci_cmd_wq#526 ->(wq_completion)nfc4_nci_cmd_wq#83 ->(wq_completion)nfc5_nci_cmd_wq#33 ->(wq_completion)nfc2_nci_cmd_wq#527 ->(wq_completion)nfc3_nci_cmd_wq#220 ->(wq_completion)nfc4_nci_cmd_wq#84 ->(wq_completion)nfc3_nci_cmd_wq#221 ->(wq_completion)nfc2_nci_cmd_wq#528 ->(wq_completion)nfc2_nci_cmd_wq#529 ->(wq_completion)nfc3_nci_cmd_wq#222 ->(wq_completion)nfc2_nci_cmd_wq#530 ->(wq_completion)nfc4_nci_cmd_wq#85 ->(wq_completion)nfc2_nci_cmd_wq#531 ->(wq_completion)nfc3_nci_cmd_wq#223 ->(wq_completion)nfc2_nci_cmd_wq#532 ->(wq_completion)nfc2_nci_cmd_wq#533 ->(wq_completion)nfc2_nci_cmd_wq#534 ->(wq_completion)nfc4_nci_cmd_wq#86 ->(wq_completion)nfc3_nci_cmd_wq#224 ->(wq_completion)nfc2_nci_cmd_wq#535 ->(wq_completion)nfc2_nci_cmd_wq#536 ->(wq_completion)nfc3_nci_cmd_wq#225 ->(wq_completion)nfc4_nci_cmd_wq#87 ->(wq_completion)nfc2_nci_cmd_wq#537 ->(wq_completion)nfc3_nci_cmd_wq#226 ->(wq_completion)nfc2_nci_cmd_wq#538 ->(wq_completion)nfc3_nci_cmd_wq#227 ->(wq_completion)nfc3_nci_cmd_wq#228 ->(wq_completion)nfc2_nci_cmd_wq#539 ->(wq_completion)nfc3_nci_cmd_wq#229 ->(wq_completion)nfc2_nci_cmd_wq#540 ->(wq_completion)nfc3_nci_cmd_wq#230 ->(wq_completion)nfc3_nci_cmd_wq#231 ->(wq_completion)nfc2_nci_cmd_wq#541 ->(wq_completion)nfc4_nci_cmd_wq#89 ->(wq_completion)nfc2_nci_cmd_wq#542 ->(wq_completion)nfc3_nci_cmd_wq#232 ->(wq_completion)nfc2_nci_cmd_wq#543 ->(wq_completion)nfc4_nci_cmd_wq#90 ->(wq_completion)nfc3_nci_cmd_wq#233 ->(wq_completion)nfc2_nci_cmd_wq#544 ->(wq_completion)nfc5_nci_cmd_wq#34 ->(wq_completion)nfc2_nci_cmd_wq#545 ->(wq_completion)nfc2_nci_cmd_wq#546 ->(wq_completion)nfc3_nci_cmd_wq#234 ->(wq_completion)nfc2_nci_cmd_wq#547 ->(wq_completion)nfc3_nci_cmd_wq#235 ->(wq_completion)nfc4_nci_cmd_wq#91 ->(wq_completion)nfc2_nci_cmd_wq#548 ->(wq_completion)nfc2_nci_cmd_wq#549 ->(wq_completion)nfc3_nci_cmd_wq#236 ->(wq_completion)nfc2_nci_cmd_wq#550 ->(wq_completion)nfc2_nci_cmd_wq#551 ->(wq_completion)nfc3_nci_cmd_wq#237 ->(wq_completion)nfc2_nci_cmd_wq#552 ->(wq_completion)nfc3_nci_cmd_wq#238 ->(wq_completion)nfc2_nci_cmd_wq#553 ->(wq_completion)nfc2_nci_cmd_wq#554 ->(wq_completion)nfc2_nci_cmd_wq#555 ->(wq_completion)nfc2_nci_cmd_wq#556 ->(wq_completion)nfc3_nci_cmd_wq#239 ->(wq_completion)nfc2_nci_cmd_wq#557 ->(wq_completion)nfc3_nci_cmd_wq#240 ->(wq_completion)nfc2_nci_cmd_wq#558 ->(wq_completion)nfc2_nci_cmd_wq#559 ->(wq_completion)nfc2_nci_cmd_wq#560 ->(wq_completion)nfc3_nci_cmd_wq#241 ->(wq_completion)nfc2_nci_cmd_wq#561 ->(wq_completion)nfc3_nci_cmd_wq#242 ->(wq_completion)nfc2_nci_cmd_wq#562 ->(wq_completion)nfc2_nci_cmd_wq#563 ->(wq_completion)nfc2_nci_cmd_wq#564 ->(wq_completion)nfc4_nci_cmd_wq#92 ->(wq_completion)nfc3_nci_cmd_wq#243 ->(wq_completion)nfc2_nci_cmd_wq#565 ->(wq_completion)nfc2_nci_cmd_wq#566 ->(wq_completion)nfc3_nci_cmd_wq#244 ->(wq_completion)nfc2_nci_cmd_wq#567 ->(wq_completion)nfc2_nci_cmd_wq#568 ->(wq_completion)nfc4_nci_cmd_wq#93 ->(wq_completion)nfc11_nci_cmd_wq#12 ->(wq_completion)nfc11_nci_cmd_wq#13 ->(wq_completion)nfc19_nci_cmd_wq#10 ->(wq_completion)nfc20_nci_cmd_wq#10 ->(wq_completion)nfc23_nci_cmd_wq#11 ->(wq_completion)nfc24_nci_cmd_wq#9 ->(wq_completion)nfc26_nci_cmd_wq#11 ->(wq_completion)nfc29_nci_cmd_wq#9 ->(wq_completion)nfc30_nci_cmd_wq#9 ->(wq_completion)nfc35_nci_cmd_wq#7 ->(wq_completion)nfc3_nci_cmd_wq#245 ->(wq_completion)nfc3_nci_cmd_wq#246 ->(wq_completion)nfc37_nci_cmd_wq#4 ->(wq_completion)nfc36_nci_cmd_wq#4 ->(wq_completion)nfc34_nci_cmd_wq#9 ->(wq_completion)nfc33_nci_cmd_wq#10 ->(wq_completion)nfc3_nci_cmd_wq#247 ->(wq_completion)nfc32_nci_cmd_wq#9 ->(wq_completion)nfc30_nci_cmd_wq#10 ->(wq_completion)nfc35_nci_cmd_wq#9 ->(wq_completion)nfc29_nci_cmd_wq#10 ->(wq_completion)nfc31_nci_cmd_wq#9 ->(wq_completion)nfc26_nci_cmd_wq#12 ->(wq_completion)nfc28_nci_cmd_wq#10 ->(wq_completion)nfc27_nci_cmd_wq#11 ->(wq_completion)nfc24_nci_cmd_wq#10 ->(wq_completion)nfc29_nci_cmd_wq#11 ->(wq_completion)nfc3_nci_cmd_wq#248 ->(wq_completion)nfc23_nci_cmd_wq#12 ->(wq_completion)nfc25_nci_cmd_wq#9 ->(wq_completion)nfc22_nci_cmd_wq#11 ->(wq_completion)nfc20_nci_cmd_wq#11 ->(wq_completion)nfc21_nci_cmd_wq#9 ->(wq_completion)nfc19_nci_cmd_wq#11 ->(wq_completion)nfc18_nci_cmd_wq#11 ->(wq_completion)nfc17_nci_cmd_wq#10 ->(wq_completion)nfc16_nci_cmd_wq#12 ->(wq_completion)nfc15_nci_cmd_wq#11 ->(wq_completion)nfc11_nci_cmd_wq#14 ->(wq_completion)nfc14_nci_cmd_wq#10 ->(wq_completion)nfc13_nci_cmd_wq#10 ->(wq_completion)nfc12_nci_cmd_wq#11 ->(wq_completion)nfc10_nci_cmd_wq#10 ->(wq_completion)nfc9_nci_cmd_wq#12 ->(wq_completion)nfc8_nci_cmd_wq#11 ->(wq_completion)nfc7_nci_cmd_wq#11 ->(wq_completion)nfc4_nci_cmd_wq#94 ->(wq_completion)nfc2_nci_cmd_wq#569 ->(wq_completion)nfc6_nci_cmd_wq#14 ->(wq_completion)nfc5_nci_cmd_wq#35 ->(wq_completion)nfc2_nci_cmd_wq#570 ->(wq_completion)nfc3_nci_cmd_wq#249 ->(wq_completion)nfc4_nci_cmd_wq#95 ->(wq_completion)nfc2_nci_cmd_wq#571 ->(wq_completion)nfc2_nci_cmd_wq#572 ->(wq_completion)nfc3_nci_cmd_wq#250 ->(wq_completion)nfc4_nci_cmd_wq#96 ->(wq_completion)nfc2_nci_cmd_wq#573 ->(wq_completion)nfc4_nci_cmd_wq#97 ->(wq_completion)nfc3_nci_cmd_wq#251 ->(wq_completion)nfc5_nci_cmd_wq#36 ->(wq_completion)nfc2_nci_cmd_wq#574 ->(wq_completion)nfc3_nci_cmd_wq#252 ->(wq_completion)nfc2_nci_cmd_wq#575 ->(wq_completion)nfc2_nci_cmd_wq#576 ->(wq_completion)nfc4_nci_cmd_wq#98 ->(wq_completion)nfc3_nci_cmd_wq#253 ->(wq_completion)nfc2_nci_cmd_wq#577 ->(wq_completion)nfc4_nci_cmd_wq#99 ->(wq_completion)nfc5_nci_cmd_wq#37 ->(wq_completion)nfc3_nci_cmd_wq#254 ->(wq_completion)nfc2_nci_cmd_wq#578 ->(wq_completion)nfc4_nci_cmd_wq#100 ->(wq_completion)nfc2_nci_cmd_wq#579 ->(wq_completion)nfc3_nci_cmd_wq#255 ->(wq_completion)nfc5_nci_cmd_wq#38 ->(wq_completion)nfc2_nci_cmd_wq#580 ->(wq_completion)nfc3_nci_cmd_wq#256 ->(wq_completion)nfc2_nci_cmd_wq#581 ->(wq_completion)nfc2_nci_cmd_wq#582 ->(wq_completion)nfc3_nci_cmd_wq#257 ->(wq_completion)nfc2_nci_cmd_wq#583 ->(wq_completion)nfc3_nci_cmd_wq#258 ->(wq_completion)nfc2_nci_cmd_wq#584 ->(wq_completion)nfc2_nci_cmd_wq#585 ->(wq_completion)nfc3_nci_cmd_wq#259 ->(wq_completion)nfc2_nci_cmd_wq#586 ->(wq_completion)nfc2_nci_cmd_wq#587 ->(wq_completion)nfc3_nci_cmd_wq#260 ->(wq_completion)nfc2_nci_cmd_wq#588 ->(wq_completion)nfc2_nci_cmd_wq#589 ->(wq_completion)nfc3_nci_cmd_wq#261 ->(wq_completion)nfc2_nci_cmd_wq#590 ->(wq_completion)nfc2_nci_cmd_wq#591 ->(wq_completion)nfc2_nci_cmd_wq#592 ->(wq_completion)nfc2_nci_cmd_wq#593 ->(wq_completion)nfc2_nci_cmd_wq#594 ->(wq_completion)nfc2_nci_cmd_wq#595 ->(wq_completion)nfc3_nci_cmd_wq#263 ->(wq_completion)nfc2_nci_cmd_wq#596 ->(wq_completion)nfc2_nci_cmd_wq#597 ->(wq_completion)nfc3_nci_cmd_wq#264 ->(wq_completion)nfc2_nci_cmd_wq#598 ->(wq_completion)nfc4_nci_cmd_wq#101 ->(wq_completion)nfc2_nci_cmd_wq#599 ->(wq_completion)nfc2_nci_cmd_wq#600 ->(wq_completion)nfc2_nci_cmd_wq#601 ->(wq_completion)nfc3_nci_cmd_wq#265 ->(wq_completion)nfc2_nci_cmd_wq#602 ->(wq_completion)nfc3_nci_cmd_wq#266 ->(wq_completion)nfc2_nci_cmd_wq#603 ->(wq_completion)nfc3_nci_cmd_wq#267 ->(wq_completion)nfc2_nci_cmd_wq#604 ->(wq_completion)nfc2_nci_cmd_wq#605 ->(wq_completion)nfc3_nci_cmd_wq#268 ->(wq_completion)nfc2_nci_cmd_wq#606 ->(wq_completion)nfc2_nci_cmd_wq#607 ->(wq_completion)nfc3_nci_cmd_wq#269 ->(wq_completion)nfc2_nci_cmd_wq#608 ->(wq_completion)nfc2_nci_cmd_wq#609 ->(wq_completion)nfc3_nci_cmd_wq#270 ->(wq_completion)nfc2_nci_cmd_wq#610 ->(wq_completion)nfc3_nci_cmd_wq#271 ->(wq_completion)nfc2_nci_cmd_wq#611 ->(wq_completion)nfc2_nci_cmd_wq#612 ->(wq_completion)nfc2_nci_cmd_wq#613 ->(wq_completion)nfc3_nci_cmd_wq#272 ->(wq_completion)nfc2_nci_cmd_wq#614 ->(wq_completion)nfc3_nci_cmd_wq#273 ->(wq_completion)nfc2_nci_cmd_wq#615 ->(wq_completion)nfc3_nci_cmd_wq#274 ->(wq_completion)nfc2_nci_cmd_wq#616 ->(wq_completion)nfc2_nci_cmd_wq#617 ->(wq_completion)nfc3_nci_cmd_wq#275 ->(wq_completion)nfc3_nci_cmd_wq#276 ->(wq_completion)nfc3_nci_cmd_wq#277 ->(wq_completion)nfc2_nci_cmd_wq#618 ->(wq_completion)nfc4_nci_cmd_wq#102 ->(wq_completion)nfc2_nci_cmd_wq#619 ->(wq_completion)nfc2_nci_cmd_wq#620 ->(wq_completion)nfc3_nci_cmd_wq#278 ->(wq_completion)nfc2_nci_cmd_wq#621 ->(wq_completion)nfc4_nci_cmd_wq#103 ->(wq_completion)nfc3_nci_cmd_wq#279 ->(wq_completion)nfc2_nci_cmd_wq#622 ->(wq_completion)nfc3_nci_cmd_wq#280 ->(wq_completion)nfc2_nci_cmd_wq#623 ->(wq_completion)nfc2_nci_cmd_wq#624 ->(wq_completion)nfc3_nci_cmd_wq#281 ->(wq_completion)nfc4_nci_cmd_wq#104 ->(wq_completion)nfc2_nci_cmd_wq#625 ->(wq_completion)nfc2_nci_cmd_wq#626 ->(wq_completion)nfc3_nci_cmd_wq#282 ->(wq_completion)nfc2_nci_cmd_wq#627 ->(wq_completion)nfc8_nci_cmd_wq#12 ->(wq_completion)nfc9_nci_cmd_wq#13 ->(wq_completion)nfc10_nci_cmd_wq#11 ->(wq_completion)nfc11_nci_cmd_wq#15 ->(wq_completion)nfc14_nci_cmd_wq#11 ->(wq_completion)nfc16_nci_cmd_wq#13 ->(wq_completion)nfc18_nci_cmd_wq#12 ->(wq_completion)nfc21_nci_cmd_wq#10 ->(wq_completion)nfc26_nci_cmd_wq#13 ->(wq_completion)nfc28_nci_cmd_wq#11 ->(wq_completion)nfc25_nci_cmd_wq#10 ->(wq_completion)nfc26_nci_cmd_wq#14 ->(wq_completion)nfc26_nci_cmd_wq#15 ->(wq_completion)nfc2_nci_cmd_wq#628 ->(wq_completion)nfc47_nci_cmd_wq#3 ->(wq_completion)nfc46_nci_cmd_wq#3 ->(wq_completion)nfc45_nci_cmd_wq#3 ->(wq_completion)nfc44_nci_cmd_wq#3 ->(wq_completion)nfc43_nci_cmd_wq#3 ->(wq_completion)nfc42_nci_cmd_wq#4 ->(wq_completion)nfc41_nci_cmd_wq#4 ->(wq_completion)nfc40_nci_cmd_wq#5 ->(wq_completion)nfc39_nci_cmd_wq#4 ->(wq_completion)nfc38_nci_cmd_wq#4 ->(wq_completion)nfc37_nci_cmd_wq#5 ->(wq_completion)nfc36_nci_cmd_wq#5 ->(wq_completion)nfc35_nci_cmd_wq#10 ->(wq_completion)nfc34_nci_cmd_wq#10 ->(wq_completion)nfc33_nci_cmd_wq#11 ->(wq_completion)nfc32_nci_cmd_wq#10 ->(wq_completion)nfc31_nci_cmd_wq#10 ->(wq_completion)nfc26_nci_cmd_wq#16 ->(wq_completion)nfc30_nci_cmd_wq#11 ->(wq_completion)nfc29_nci_cmd_wq#12 ->(wq_completion)nfc25_nci_cmd_wq#11 ->(wq_completion)nfc28_nci_cmd_wq#12 ->(wq_completion)nfc27_nci_cmd_wq#12 ->(wq_completion)nfc24_nci_cmd_wq#11 ->(wq_completion)nfc21_nci_cmd_wq#11 ->(wq_completion)nfc23_nci_cmd_wq#13 ->(wq_completion)nfc22_nci_cmd_wq#12 ->(wq_completion)nfc20_nci_cmd_wq#12 ->(wq_completion)nfc19_nci_cmd_wq#12 ->(wq_completion)nfc18_nci_cmd_wq#13 ->(wq_completion)nfc16_nci_cmd_wq#14 ->(wq_completion)nfc17_nci_cmd_wq#11 ->(wq_completion)nfc14_nci_cmd_wq#12 ->(wq_completion)nfc15_nci_cmd_wq#12 ->(wq_completion)nfc11_nci_cmd_wq#16 ->(wq_completion)nfc13_nci_cmd_wq#11 ->(wq_completion)nfc12_nci_cmd_wq#12 ->(wq_completion)nfc10_nci_cmd_wq#12 ->(wq_completion)nfc9_nci_cmd_wq#14 ->(wq_completion)nfc8_nci_cmd_wq#13 ->(wq_completion)nfc7_nci_cmd_wq#12 ->(wq_completion)nfc6_nci_cmd_wq#15 ->(wq_completion)nfc5_nci_cmd_wq#39 ->(wq_completion)nfc4_nci_cmd_wq#105 ->(wq_completion)nfc3_nci_cmd_wq#283 ->(wq_completion)nfc2_nci_cmd_wq#629 ->(wq_completion)nfc2_nci_cmd_wq#630 ->(wq_completion)nfc2_nci_cmd_wq#631 ->(wq_completion)nfc3_nci_cmd_wq#284 ->(wq_completion)nfc4_nci_cmd_wq#106 ->(wq_completion)nfc2_nci_cmd_wq#632 ->(wq_completion)nfc5_nci_cmd_wq#40 ->(wq_completion)nfc2_nci_cmd_wq#633 ->(wq_completion)nfc3_nci_cmd_wq#285 ->(wq_completion)nfc2_nci_cmd_wq#634 ->(wq_completion)nfc2_nci_cmd_wq#635 ->(wq_completion)nfc2_nci_cmd_wq#636 ->(wq_completion)nfc2_nci_cmd_wq#637 ->(wq_completion)nfc3_nci_cmd_wq#286 ->(wq_completion)nfc2_nci_cmd_wq#638 ->(wq_completion)nfc2_nci_cmd_wq#639 ->(wq_completion)nfc3_nci_cmd_wq#287 ->(wq_completion)nfc2_nci_cmd_wq#640 ->(wq_completion)nfc4_nci_cmd_wq#107 ->(wq_completion)nfc2_nci_cmd_wq#641 ->(wq_completion)nfc2_nci_cmd_wq#642 ->(wq_completion)nfc2_nci_cmd_wq#643 ->(wq_completion)nfc2_nci_cmd_wq#644 ->(wq_completion)nfc2_nci_cmd_wq#645 ->(wq_completion)nfc3_nci_cmd_wq#288 ->(wq_completion)nfc2_nci_cmd_wq#646 ->(wq_completion)nfc3_nci_cmd_wq#289 ->(wq_completion)nfc2_nci_cmd_wq#647 ->(wq_completion)nfc2_nci_cmd_wq#648 ->(wq_completion)nfc3_nci_cmd_wq#290 ->(wq_completion)nfc3_nci_cmd_wq#291 ->(wq_completion)nfc2_nci_cmd_wq#649 ->(wq_completion)nfc3_nci_cmd_wq#292 ->(wq_completion)nfc4_nci_cmd_wq#108 ->(wq_completion)nfc5_nci_cmd_wq#41 ->(wq_completion)nfc2_nci_cmd_wq#650 ->(wq_completion)nfc3_nci_cmd_wq#293 ->(wq_completion)nfc2_nci_cmd_wq#651 ->(wq_completion)nfc4_nci_cmd_wq#109 ->(wq_completion)nfc2_nci_cmd_wq#652 ->(wq_completion)nfc2_nci_cmd_wq#653 ->(wq_completion)nfc4_nci_cmd_wq#110 ->(wq_completion)nfc3_nci_cmd_wq#296 ->(wq_completion)nfc2_nci_cmd_wq#654 ->(wq_completion)nfc2_nci_cmd_wq#655 ->(wq_completion)nfc3_nci_cmd_wq#297 ->(wq_completion)nfc2_nci_cmd_wq#656 ->(wq_completion)nfc3_nci_cmd_wq#298 ->(wq_completion)nfc2_nci_cmd_wq#657 ->(wq_completion)nfc2_nci_cmd_wq#658 ->(wq_completion)nfc3_nci_cmd_wq#299 ->(wq_completion)nfc2_nci_cmd_wq#659 ->(wq_completion)nfc4_nci_cmd_wq#111 ->(wq_completion)nfc3_nci_cmd_wq#300 ->(wq_completion)nfc2_nci_cmd_wq#660 ->(wq_completion)nfc4_nci_cmd_wq#112 ->(wq_completion)nfc2_nci_cmd_wq#661 ->(wq_completion)nfc2_nci_cmd_wq#662 ->(wq_completion)nfc2_nci_cmd_wq#663 ->(wq_completion)nfc3_nci_cmd_wq#301 ->(wq_completion)nfc2_nci_cmd_wq#664 ->(wq_completion)nfc2_nci_cmd_wq#665 ->(wq_completion)nfc3_nci_cmd_wq#302 ->(wq_completion)nfc2_nci_cmd_wq#666 ->(wq_completion)nfc2_nci_cmd_wq#667 ->(wq_completion)nfc2_nci_cmd_wq#668 ->(wq_completion)nfc2_nci_cmd_wq#669 ->(wq_completion)nfc3_nci_cmd_wq#303 ->(wq_completion)nfc2_nci_cmd_wq#670 ->(wq_completion)nfc2_nci_cmd_wq#671 ->(wq_completion)nfc2_nci_cmd_wq#672 ->(wq_completion)nfc2_nci_cmd_wq#673 ->(wq_completion)nfc3_nci_cmd_wq#304 ->(wq_completion)nfc2_nci_cmd_wq#674 ->(wq_completion)nfc4_nci_cmd_wq#113 ->(wq_completion)nfc2_nci_cmd_wq#675 ->(wq_completion)nfc2_nci_cmd_wq#676 ->(wq_completion)nfc4_nci_cmd_wq#114 ->(wq_completion)nfc3_nci_cmd_wq#305 ->(wq_completion)nfc2_nci_cmd_wq#677 ->(wq_completion)nfc2_nci_cmd_wq#678 ->(wq_completion)nfc2_nci_cmd_wq#679 ->(wq_completion)nfc3_nci_cmd_wq#306 ->(wq_completion)nfc2_nci_cmd_wq#680 ->(wq_completion)nfc3_nci_cmd_wq#307 ->(wq_completion)nfc2_nci_cmd_wq#681 ->(wq_completion)nfc2_nci_cmd_wq#682 ->(wq_completion)nfc2_nci_cmd_wq#683 ->(wq_completion)nfc3_nci_cmd_wq#308 ->(wq_completion)nfc2_nci_cmd_wq#684 ->(wq_completion)nfc4_nci_cmd_wq#115 ->(wq_completion)nfc3_nci_cmd_wq#309 ->(wq_completion)nfc2_nci_cmd_wq#685 ->(wq_completion)nfc2_nci_cmd_wq#686 ->(wq_completion)nfc4_nci_cmd_wq#116 ->(wq_completion)nfc3_nci_cmd_wq#310 ->(wq_completion)nfc2_nci_cmd_wq#687 ->(wq_completion)nfc2_nci_cmd_wq#688 ->(wq_completion)nfc3_nci_cmd_wq#311 ->(wq_completion)nfc2_nci_cmd_wq#689 ->(wq_completion)nfc3_nci_cmd_wq#312 ->(wq_completion)nfc2_nci_cmd_wq#690 ->(wq_completion)nfc3_nci_cmd_wq#313 ->(wq_completion)nfc4_nci_cmd_wq#117 ->(wq_completion)nfc2_nci_cmd_wq#691 ->(wq_completion)nfc2_nci_cmd_wq#692 ->(wq_completion)nfc3_nci_cmd_wq#314 ->(wq_completion)nfc2_nci_cmd_wq#693 ->(wq_completion)nfc4_nci_cmd_wq#118 ->(wq_completion)nfc3_nci_cmd_wq#315 ->(wq_completion)nfc2_nci_cmd_wq#694 ->(wq_completion)nfc2_nci_cmd_wq#695 ->(wq_completion)nfc3_nci_cmd_wq#316 ->(wq_completion)nfc4_nci_cmd_wq#119 ->(wq_completion)nfc3_nci_cmd_wq#317 ->(wq_completion)nfc3_nci_cmd_wq#318 ->(wq_completion)nfc2_nci_cmd_wq#696 ->(wq_completion)nfc4_nci_cmd_wq#120 ->(wq_completion)nfc2_nci_cmd_wq#697 ->(wq_completion)nfc4_nci_cmd_wq#121 ->(wq_completion)nfc3_nci_cmd_wq#319 ->(wq_completion)nfc2_nci_cmd_wq#698 ->(wq_completion)nfc2_nci_cmd_wq#699 ->(wq_completion)nfc3_nci_cmd_wq#320 ->(wq_completion)nfc8_nci_cmd_wq#14 ->(wq_completion)nfc9_nci_cmd_wq#15 ->(wq_completion)nfc7_nci_cmd_wq#13 ->(wq_completion)nfc6_nci_cmd_wq#16 ->(wq_completion)nfc5_nci_cmd_wq#42 ->(wq_completion)nfc4_nci_cmd_wq#122 ->(wq_completion)nfc3_nci_cmd_wq#321 ->(wq_completion)nfc5_nci_cmd_wq#43 ->(wq_completion)nfc2_nci_cmd_wq#700 ->(wq_completion)nfc2_nci_cmd_wq#701 ->(wq_completion)nfc3_nci_cmd_wq#322 ->(wq_completion)nfc3_nci_cmd_wq#323 ->(wq_completion)nfc4_nci_cmd_wq#123 ->(wq_completion)nfc2_nci_cmd_wq#702 ->(wq_completion)nfc2_nci_cmd_wq#703 ->(wq_completion)nfc3_nci_cmd_wq#324 ->(wq_completion)nfc2_nci_cmd_wq#704 ->(wq_completion)nfc2_nci_cmd_wq#705 ->(wq_completion)nfc2_nci_cmd_wq#706 ->(wq_completion)nfc4_nci_cmd_wq#124 ->(wq_completion)nfc3_nci_cmd_wq#325 ->(wq_completion)nfc2_nci_cmd_wq#707 ->(wq_completion)nfc2_nci_cmd_wq#708 ->(wq_completion)nfc3_nci_cmd_wq#326 ->(wq_completion)nfc4_nci_cmd_wq#125 ->(wq_completion)nfc5_nci_cmd_wq#45 ->(wq_completion)nfc2_nci_cmd_wq#709 ->(wq_completion)nfc3_nci_cmd_wq#327 ->(wq_completion)nfc6_nci_cmd_wq#17 ->(wq_completion)nfc2_nci_cmd_wq#710 ->(wq_completion)nfc3_nci_cmd_wq#328 ->(wq_completion)nfc2_nci_cmd_wq#711 ->(wq_completion)nfc3_nci_cmd_wq#329 ->(wq_completion)nfc2_nci_cmd_wq#712 ->(wq_completion)nfc4_nci_cmd_wq#126 ->(wq_completion)nfc5_nci_cmd_wq#46 ->(wq_completion)nfc2_nci_cmd_wq#714 ->(wq_completion)nfc3_nci_cmd_wq#330 ->(wq_completion)nfc2_nci_cmd_wq#715 ->(wq_completion)nfc3_nci_cmd_wq#332 ->(wq_completion)nfc2_nci_cmd_wq#716 ->(wq_completion)nfc4_nci_cmd_wq#127 ->(wq_completion)nfc2_nci_cmd_wq#717 ->(wq_completion)nfc2_nci_cmd_wq#718 ->(wq_completion)nfc3_nci_cmd_wq#333 ->(wq_completion)nfc4_nci_cmd_wq#128 ->(wq_completion)nfc6_nci_cmd_wq#18 ->(wq_completion)nfc2_nci_cmd_wq#719 ->(wq_completion)nfc5_nci_cmd_wq#47 ->(wq_completion)nfc2_nci_cmd_wq#720 ->(wq_completion)nfc2_nci_cmd_wq#721 ->(wq_completion)nfc3_nci_cmd_wq#335 ->(wq_completion)nfc2_nci_cmd_wq#722 ->(wq_completion)nfc3_nci_cmd_wq#336 ->(wq_completion)nfc2_nci_cmd_wq#723 ->(wq_completion)nfc3_nci_cmd_wq#337 ->(wq_completion)nfc2_nci_cmd_wq#724 ->(wq_completion)nfc3_nci_cmd_wq#338 ->(wq_completion)nfc2_nci_cmd_wq#725 ->(wq_completion)nfc2_nci_cmd_wq#726 ->(wq_completion)nfc2_nci_cmd_wq#727 ->(wq_completion)nfc2_nci_cmd_wq#728 ->(wq_completion)nfc2_nci_cmd_wq#729 ->(wq_completion)nfc3_nci_cmd_wq#339 ->(wq_completion)nfc2_nci_cmd_wq#730 ->(wq_completion)nfc2_nci_cmd_wq#731 ->(wq_completion)nfc3_nci_cmd_wq#340 ->(wq_completion)nfc2_nci_cmd_wq#732 ->(wq_completion)nfc3_nci_cmd_wq#341 ->(wq_completion)nfc2_nci_cmd_wq#733 ->(wq_completion)nfc2_nci_cmd_wq#734 ->(wq_completion)nfc2_nci_cmd_wq#735 ->(wq_completion)nfc3_nci_cmd_wq#342 ->(wq_completion)nfc2_nci_cmd_wq#736 ->(wq_completion)nfc2_nci_cmd_wq#737 ->(wq_completion)nfc3_nci_cmd_wq#343 ->(wq_completion)nfc2_nci_cmd_wq#738 ->(wq_completion)nfc4_nci_cmd_wq#129 ->(wq_completion)nfc2_nci_cmd_wq#739 ->(wq_completion)nfc3_nci_cmd_wq#344 ->(wq_completion)nfc2_nci_cmd_wq#740 ->(wq_completion)nfc4_nci_cmd_wq#130 ->(wq_completion)nfc2_nci_cmd_wq#741 ->(wq_completion)nfc3_nci_cmd_wq#345 ->(wq_completion)nfc2_nci_cmd_wq#742 ->(wq_completion)nfc2_nci_cmd_wq#743 ->(wq_completion)nfc2_nci_cmd_wq#744 ->(wq_completion)nfc3_nci_cmd_wq#346 ->(wq_completion)nfc2_nci_cmd_wq#745 ->(wq_completion)nfc2_nci_cmd_wq#746 ->(wq_completion)nfc2_nci_cmd_wq#747 ->(wq_completion)nfc2_nci_cmd_wq#748 ->(wq_completion)nfc3_nci_cmd_wq#347 ->(wq_completion)nfc2_nci_cmd_wq#750 ->(wq_completion)nfc3_nci_cmd_wq#348 ->(wq_completion)nfc2_nci_cmd_wq#751 ->(wq_completion)nfc2_nci_cmd_wq#752 ->(wq_completion)nfc3_nci_cmd_wq#349 ->(wq_completion)nfc4_nci_cmd_wq#131 ->(wq_completion)nfc2_nci_cmd_wq#753 ->(wq_completion)nfc2_nci_cmd_wq#754 ->(wq_completion)nfc3_nci_cmd_wq#350 ->(wq_completion)nfc2_nci_cmd_wq#755 ->(wq_completion)nfc3_nci_cmd_wq#351 ->(wq_completion)nfc4_nci_cmd_wq#132 ->(wq_completion)nfc2_nci_cmd_wq#756 ->(wq_completion)nfc4_nci_cmd_wq#133 ->(wq_completion)nfc3_nci_cmd_wq#352 ->(wq_completion)nfc2_nci_cmd_wq#757 ->(wq_completion)nfc2_nci_cmd_wq#758 ->(wq_completion)nfc3_nci_cmd_wq#353 ->(wq_completion)nfc2_nci_cmd_wq#759 ->(wq_completion)nfc4_nci_cmd_wq#134 ->(wq_completion)nfc3_nci_cmd_wq#354 ->(wq_completion)nfc2_nci_cmd_wq#760 ->(wq_completion)nfc2_nci_cmd_wq#761 ->(wq_completion)nfc3_nci_cmd_wq#355 ->(wq_completion)nfc3_nci_cmd_wq#356 ->(wq_completion)nfc3_nci_cmd_wq#357 ->(wq_completion)nfc2_nci_cmd_wq#763 ->(wq_completion)nfc2_nci_cmd_wq#764 ->(wq_completion)nfc3_nci_cmd_wq#358 ->(wq_completion)nfc3_nci_cmd_wq#359 ->(wq_completion)nfc4_nci_cmd_wq#135 ->(wq_completion)nfc2_nci_cmd_wq#765 ->(wq_completion)nfc4_nci_cmd_wq#136 ->(wq_completion)nfc3_nci_cmd_wq#360 ->(wq_completion)nfc2_nci_cmd_wq#766 ->(wq_completion)nfc5_nci_cmd_wq#48 ->(wq_completion)nfc2_nci_cmd_wq#767 ->(wq_completion)nfc3_nci_cmd_wq#361 ->(wq_completion)nfc2_nci_cmd_wq#768 ->(wq_completion)nfc3_nci_cmd_wq#362 ->(wq_completion)nfc4_nci_cmd_wq#137 ->(wq_completion)nfc2_nci_cmd_wq#769 ->(wq_completion)nfc2_nci_cmd_wq#770 ->(wq_completion)nfc3_nci_cmd_wq#363 ->(wq_completion)nfc2_nci_cmd_wq#771 ->(wq_completion)nfc2_nci_cmd_wq#772 ->(wq_completion)nfc3_nci_cmd_wq#364 ->(wq_completion)nfc2_nci_cmd_wq#773 ->(wq_completion)nfc3_nci_cmd_wq#365 ->(wq_completion)nfc3_nci_cmd_wq#366 ->(wq_completion)nfc2_nci_cmd_wq#774 ->(wq_completion)nfc2_nci_cmd_wq#775 ->(wq_completion)nfc2_nci_cmd_wq#776 ->(wq_completion)nfc2_nci_cmd_wq#777 ->(wq_completion)nfc2_nci_cmd_wq#778 ->(wq_completion)nfc3_nci_cmd_wq#367 ->(wq_completion)nfc2_nci_cmd_wq#779 ->(wq_completion)nfc3_nci_cmd_wq#368 ->(wq_completion)nfc4_nci_cmd_wq#138 ->(wq_completion)nfc2_nci_cmd_wq#780 ->(wq_completion)nfc3_nci_cmd_wq#369 ->(wq_completion)nfc4_nci_cmd_wq#139 ->(wq_completion)nfc2_nci_cmd_wq#781 ->(wq_completion)nfc5_nci_cmd_wq#49 ->(wq_completion)nfc2_nci_cmd_wq#782 ->(wq_completion)nfc2_nci_cmd_wq#783 ->(wq_completion)nfc2_nci_cmd_wq#784 ->(wq_completion)nfc3_nci_cmd_wq#370 ->(wq_completion)nfc2_nci_cmd_wq#785 ->(wq_completion)nfc2_nci_cmd_wq#786 ->(wq_completion)nfc3_nci_cmd_wq#371 ->(wq_completion)nfc2_nci_cmd_wq#787 ->(wq_completion)nfc3_nci_cmd_wq#372 ->(wq_completion)nfc3_nci_cmd_wq#373 ->(wq_completion)nfc2_nci_cmd_wq#788 ->(wq_completion)nfc4_nci_cmd_wq#140 ->(wq_completion)nfc2_nci_cmd_wq#789 ->(wq_completion)nfc2_nci_cmd_wq#790 ->(wq_completion)nfc2_nci_cmd_wq#791 ->(wq_completion)nfc2_nci_cmd_wq#792 ->(wq_completion)nfc3_nci_cmd_wq#374 ->(wq_completion)nfc4_nci_cmd_wq#141 ->(wq_completion)nfc2_nci_cmd_wq#793 ->(wq_completion)nfc2_nci_cmd_wq#794 ->(wq_completion)nfc3_nci_cmd_wq#375 ->(wq_completion)nfc2_nci_cmd_wq#795 ->(wq_completion)nfc4_nci_cmd_wq#142 ->(wq_completion)nfc3_nci_cmd_wq#376 ->(wq_completion)nfc2_nci_cmd_wq#796 ->(wq_completion)nfc4_nci_cmd_wq#143 ->(wq_completion)nfc2_nci_cmd_wq#797 ->(wq_completion)nfc3_nci_cmd_wq#377 ->(wq_completion)nfc4_nci_cmd_wq#144 ->(wq_completion)nfc3_nci_cmd_wq#378 ->(wq_completion)nfc2_nci_cmd_wq#798 ->(wq_completion)nfc2_nci_cmd_wq#799 ->(wq_completion)nfc2_nci_cmd_wq#800 ->(wq_completion)nfc3_nci_cmd_wq#379 ->(wq_completion)nfc6_nci_cmd_wq#19 ->(wq_completion)nfc12_nci_cmd_wq#13 ->(wq_completion)nfc20_nci_cmd_wq#13 ->(wq_completion)nfc27_nci_cmd_wq#13 ->(wq_completion)nfc33_nci_cmd_wq#12 ->(wq_completion)nfc2_nci_cmd_wq#801 ->(wq_completion)nfc42_nci_cmd_wq#5 ->(wq_completion)nfc41_nci_cmd_wq#5 ->(wq_completion)nfc40_nci_cmd_wq#6 ->(wq_completion)nfc39_nci_cmd_wq#5 ->(wq_completion)nfc38_nci_cmd_wq#5 ->(wq_completion)nfc37_nci_cmd_wq#6 ->(wq_completion)nfc33_nci_cmd_wq#13 ->(wq_completion)nfc36_nci_cmd_wq#6 ->(wq_completion)nfc35_nci_cmd_wq#11 ->(wq_completion)nfc34_nci_cmd_wq#11 ->(wq_completion)nfc32_nci_cmd_wq#11 ->(wq_completion)nfc31_nci_cmd_wq#11 ->(wq_completion)nfc30_nci_cmd_wq#12 ->(wq_completion)nfc29_nci_cmd_wq#13 ->(wq_completion)nfc27_nci_cmd_wq#14 ->(wq_completion)nfc28_nci_cmd_wq#13 ->(wq_completion)nfc26_nci_cmd_wq#17 ->(wq_completion)nfc25_nci_cmd_wq#12 ->(wq_completion)nfc24_nci_cmd_wq#12 ->(wq_completion)nfc23_nci_cmd_wq#14 ->(wq_completion)nfc20_nci_cmd_wq#14 ->(wq_completion)nfc22_nci_cmd_wq#13 ->(wq_completion)nfc21_nci_cmd_wq#12 ->(wq_completion)nfc19_nci_cmd_wq#13 ->(wq_completion)nfc18_nci_cmd_wq#14 ->(wq_completion)nfc17_nci_cmd_wq#12 ->(wq_completion)nfc16_nci_cmd_wq#15 ->(wq_completion)nfc15_nci_cmd_wq#13 ->(wq_completion)nfc12_nci_cmd_wq#14 ->(wq_completion)nfc14_nci_cmd_wq#13 ->(wq_completion)nfc13_nci_cmd_wq#12 ->(wq_completion)nfc11_nci_cmd_wq#17 ->(wq_completion)nfc10_nci_cmd_wq#13 ->(wq_completion)nfc9_nci_cmd_wq#16 ->(wq_completion)nfc8_nci_cmd_wq#16 ->(wq_completion)nfc7_nci_cmd_wq#14 ->(wq_completion)nfc6_nci_cmd_wq#20 ->(wq_completion)nfc3_nci_cmd_wq#380 ->(wq_completion)nfc5_nci_cmd_wq#50 ->(wq_completion)nfc4_nci_cmd_wq#145 ->(wq_completion)nfc2_nci_cmd_wq#802 ->(wq_completion)nfc2_nci_cmd_wq#803 ->(wq_completion)nfc2_nci_cmd_wq#804 ->(wq_completion)nfc2_nci_cmd_wq#805 ->(wq_completion)nfc2_nci_cmd_wq#806 ->(wq_completion)nfc2_nci_cmd_wq#807 ->(wq_completion)nfc3_nci_cmd_wq#381 ->(wq_completion)nfc2_nci_cmd_wq#808 ->(wq_completion)nfc2_nci_cmd_wq#809 ->(wq_completion)nfc2_nci_cmd_wq#810 ->(wq_completion)nfc3_nci_cmd_wq#382 ->(wq_completion)nfc2_nci_cmd_wq#811 ->(wq_completion)nfc2_nci_cmd_wq#812 ->(wq_completion)nfc2_nci_cmd_wq#813 ->(wq_completion)nfc2_nci_cmd_wq#814 ->(wq_completion)nfc2_nci_cmd_wq#815 ->(wq_completion)nfc2_nci_cmd_wq#816 ->(wq_completion)nfc2_nci_cmd_wq#817 ->(wq_completion)nfc2_nci_cmd_wq#818 ->(wq_completion)nfc2_nci_cmd_wq#819 ->(wq_completion)nfc2_nci_cmd_wq#820 ->(wq_completion)nfc2_nci_cmd_wq#821 ->(wq_completion)nfc2_nci_cmd_wq#822 ->(wq_completion)nfc2_nci_cmd_wq#823 ->(wq_completion)nfc3_nci_cmd_wq#383 ->(wq_completion)nfc2_nci_cmd_wq#824 ->(wq_completion)nfc2_nci_cmd_wq#825 ->(wq_completion)nfc2_nci_cmd_wq#826 ->(wq_completion)nfc2_nci_cmd_wq#827 ->(wq_completion)nfc2_nci_cmd_wq#828 ->(wq_completion)nfc2_nci_cmd_wq#829 ->(wq_completion)nfc2_nci_cmd_wq#830 ->(wq_completion)nfc2_nci_cmd_wq#831 ->(wq_completion)nfc2_nci_cmd_wq#832 ->(wq_completion)nfc2_nci_cmd_wq#833 ->(wq_completion)nfc3_nci_cmd_wq#384 ->(wq_completion)nfc2_nci_cmd_wq#835 ->(wq_completion)nfc4_nci_cmd_wq#146 ->(wq_completion)nfc2_nci_cmd_wq#836 ->(wq_completion)nfc2_nci_cmd_wq#837 ->(wq_completion)nfc4_nci_cmd_wq#147 ->(wq_completion)nfc3_nci_cmd_wq#385 ->(wq_completion)nfc2_nci_cmd_wq#838 ->(wq_completion)nfc2_nci_cmd_wq#839 ->(wq_completion)nfc2_nci_cmd_wq#840 ->(wq_completion)nfc2_nci_cmd_wq#841 ->(wq_completion)nfc3_nci_cmd_wq#386 ->(wq_completion)nfc2_nci_cmd_wq#842 ->(wq_completion)nfc3_nci_cmd_wq#387 ->(wq_completion)nfc2_nci_cmd_wq#843 ->(wq_completion)nfc3_nci_cmd_wq#388 ->(wq_completion)nfc2_nci_cmd_wq#844 ->(wq_completion)nfc2_nci_cmd_wq#845 ->(wq_completion)nfc3_nci_cmd_wq#389 ->(wq_completion)nfc4_nci_cmd_wq#148 ->(wq_completion)nfc2_nci_cmd_wq#846 ->(wq_completion)nfc3_nci_cmd_wq#390 ->(wq_completion)nfc2_nci_cmd_wq#847 ->(wq_completion)nfc2_nci_cmd_wq#848 ->(wq_completion)nfc2_nci_cmd_wq#849 ->(wq_completion)nfc2_nci_cmd_wq#850 ->(wq_completion)nfc4_nci_cmd_wq#149 ->(wq_completion)nfc3_nci_cmd_wq#391 ->(wq_completion)nfc3_nci_cmd_wq#392 ->(wq_completion)nfc2_nci_cmd_wq#851 ->(wq_completion)nfc3_nci_cmd_wq#393 ->(wq_completion)nfc4_nci_cmd_wq#150 ->(wq_completion)nfc2_nci_cmd_wq#852 ->(wq_completion)nfc3_nci_cmd_wq#394 ->(wq_completion)nfc2_nci_cmd_wq#853 ->(wq_completion)nfc2_nci_cmd_wq#854 ->(wq_completion)nfc2_nci_cmd_wq#855 ->(wq_completion)nfc3_nci_cmd_wq#395 ->(wq_completion)nfc4_nci_cmd_wq#151 ->(wq_completion)nfc2_nci_cmd_wq#856 ->(wq_completion)nfc3_nci_cmd_wq#396 ->(wq_completion)nfc2_nci_cmd_wq#857 ->(wq_completion)nfc7_nci_cmd_wq#15 ->(wq_completion)nfc7_nci_cmd_wq#16 ->(wq_completion)nfc14_nci_cmd_wq#14 ->(wq_completion)nfc14_nci_cmd_wq#15 ->(wq_completion)nfc23_nci_cmd_wq#15 ->(wq_completion)nfc29_nci_cmd_wq#14 ->(wq_completion)nfc35_nci_cmd_wq#12 ->(wq_completion)nfc2_nci_cmd_wq#858 ->(wq_completion)nfc3_nci_cmd_wq#397 ->(wq_completion)nfc4_nci_cmd_wq#152 ->(wq_completion)nfc40_nci_cmd_wq#7 ->(wq_completion)nfc5_nci_cmd_wq#51 ->(wq_completion)nfc6_nci_cmd_wq#21 ->(wq_completion)nfc8_nci_cmd_wq#17 ->(wq_completion)nfc9_nci_cmd_wq#17 ->(wq_completion)nfc10_nci_cmd_wq#14 ->(wq_completion)nfc4_nci_cmd_wq#153 ->(wq_completion)nfc11_nci_cmd_wq#18 ->(wq_completion)nfc7_nci_cmd_wq#17 ->(wq_completion)nfc12_nci_cmd_wq#15 ->(wq_completion)nfc13_nci_cmd_wq#13 ->(wq_completion)nfc15_nci_cmd_wq#14 ->(wq_completion)nfc16_nci_cmd_wq#16 ->(wq_completion)nfc17_nci_cmd_wq#13 ->(wq_completion)nfc18_nci_cmd_wq#15 ->(wq_completion)nfc19_nci_cmd_wq#14 ->(wq_completion)nfc14_nci_cmd_wq#16 ->(wq_completion)nfc20_nci_cmd_wq#15 ->(wq_completion)nfc21_nci_cmd_wq#13 ->(wq_completion)nfc22_nci_cmd_wq#14 ->(wq_completion)nfc24_nci_cmd_wq#13 ->(wq_completion)nfc25_nci_cmd_wq#13 ->(wq_completion)nfc39_nci_cmd_wq#6 ->(wq_completion)nfc2_nci_cmd_wq#859 ->(wq_completion)nfc3_nci_cmd_wq#398 ->(wq_completion)nfc5_nci_cmd_wq#52 ->(wq_completion)nfc4_nci_cmd_wq#154 ->(wq_completion)nfc6_nci_cmd_wq#22 ->(wq_completion)nfc7_nci_cmd_wq#18 ->(wq_completion)nfc8_nci_cmd_wq#18 ->(wq_completion)nfc9_nci_cmd_wq#18 ->(wq_completion)nfc10_nci_cmd_wq#15 ->(wq_completion)nfc11_nci_cmd_wq#19 ->(wq_completion)nfc12_nci_cmd_wq#16 ->(wq_completion)nfc38_nci_cmd_wq#6 ->(wq_completion)nfc37_nci_cmd_wq#7 ->(wq_completion)nfc35_nci_cmd_wq#13 ->(wq_completion)nfc36_nci_cmd_wq#8 ->(wq_completion)nfc34_nci_cmd_wq#12 ->(wq_completion)nfc33_nci_cmd_wq#14 ->(wq_completion)nfc32_nci_cmd_wq#12 ->(wq_completion)nfc29_nci_cmd_wq#15 ->(wq_completion)nfc31_nci_cmd_wq#12 ->(wq_completion)nfc30_nci_cmd_wq#13 ->(wq_completion)nfc28_nci_cmd_wq#14 ->(wq_completion)nfc27_nci_cmd_wq#15 ->(wq_completion)nfc26_nci_cmd_wq#18 ->(wq_completion)nfc23_nci_cmd_wq#16 ->(wq_completion)nfc8_nci_cmd_wq#19 ->(wq_completion)nfc7_nci_cmd_wq#19 ->(wq_completion)nfc6_nci_cmd_wq#23 ->(wq_completion)nfc5_nci_cmd_wq#53 ->(wq_completion)nfc4_nci_cmd_wq#155 ->(wq_completion)nfc3_nci_cmd_wq#399 ->(wq_completion)nfc2_nci_cmd_wq#860 ->(wq_completion)nfc2_nci_cmd_wq#862 ->(wq_completion)nfc2_nci_cmd_wq#863 ->(wq_completion)nfc2_nci_cmd_wq#864 ->(wq_completion)nfc2_nci_cmd_wq#865 ->(wq_completion)nfc3_nci_cmd_wq#400 ->(wq_completion)nfc2_nci_cmd_wq#866 ->(wq_completion)nfc3_nci_cmd_wq#401 ->(wq_completion)nfc2_nci_cmd_wq#867 ->(wq_completion)nfc4_nci_cmd_wq#156 ->(wq_completion)nfc3_nci_cmd_wq#402 ->(wq_completion)nfc2_nci_cmd_wq#868 ->(wq_completion)nfc3_nci_cmd_wq#403 ->(wq_completion)nfc2_nci_cmd_wq#869 ->(wq_completion)nfc3_nci_cmd_wq#404 ->(wq_completion)nfc5_nci_cmd_wq#54 ->(wq_completion)nfc4_nci_cmd_wq#157 ->(wq_completion)nfc2_nci_cmd_wq#870 ->(wq_completion)nfc3_nci_cmd_wq#405 ->(wq_completion)nfc4_nci_cmd_wq#158 ->(wq_completion)nfc2_nci_cmd_wq#871 ->(wq_completion)nfc2_nci_cmd_wq#875 ->(wq_completion)nfc3_nci_cmd_wq#406 ->(wq_completion)nfc2_nci_cmd_wq#876 ->(wq_completion)nfc3_nci_cmd_wq#407 ->(wq_completion)nfc2_nci_cmd_wq#877 ->(wq_completion)nfc4_nci_cmd_wq#159 ->(wq_completion)nfc2_nci_cmd_wq#878 ->(wq_completion)nfc2_nci_cmd_wq#879 ->(wq_completion)nfc2_nci_cmd_wq#880 ->(wq_completion)nfc3_nci_cmd_wq#408 ->(wq_completion)nfc2_nci_cmd_wq#881 ->(wq_completion)nfc3_nci_cmd_wq#409 ->(wq_completion)nfc2_nci_cmd_wq#882 ->(wq_completion)nfc3_nci_cmd_wq#410 ->(wq_completion)nfc2_nci_cmd_wq#883 ->(wq_completion)nfc2_nci_cmd_wq#884 ->(wq_completion)nfc2_nci_cmd_wq#885 ->(wq_completion)nfc2_nci_cmd_wq#886 ->(wq_completion)nfc3_nci_cmd_wq#411 ->(wq_completion)nfc2_nci_cmd_wq#887 ->(wq_completion)nfc3_nci_cmd_wq#412 ->(wq_completion)nfc2_nci_cmd_wq#888 ->(wq_completion)nfc2_nci_cmd_wq#889 ->(wq_completion)nfc2_nci_cmd_wq#890 ->(wq_completion)nfc2_nci_cmd_wq#891 ->(wq_completion)nfc3_nci_cmd_wq#413 ->(wq_completion)nfc3_nci_cmd_wq#414 ->(wq_completion)nfc2_nci_cmd_wq#892 ->(wq_completion)nfc2_nci_cmd_wq#894 ->(wq_completion)nfc2_nci_cmd_wq#895 ->(wq_completion)nfc2_nci_cmd_wq#896 ->(wq_completion)nfc3_nci_cmd_wq#415 ->(wq_completion)nfc2_nci_cmd_wq#897 ->(wq_completion)nfc4_nci_cmd_wq#160 ->(wq_completion)nfc2_nci_cmd_wq#899 ->(wq_completion)nfc2_nci_cmd_wq#900 ->(wq_completion)nfc3_nci_cmd_wq#416 ->(wq_completion)nfc2_nci_cmd_wq#901 ->(wq_completion)nfc4_nci_cmd_wq#161 ->(wq_completion)nfc2_nci_cmd_wq#902 ->(wq_completion)nfc3_nci_cmd_wq#417 ->(wq_completion)nfc2_nci_cmd_wq#903 ->(wq_completion)nfc2_nci_cmd_wq#904 ->(wq_completion)nfc2_nci_cmd_wq#905 ->(wq_completion)nfc2_nci_cmd_wq#906 ->(wq_completion)nfc2_nci_cmd_wq#907 ->(wq_completion)nfc3_nci_cmd_wq#418 ->(wq_completion)nfc2_nci_cmd_wq#908 ->(wq_completion)nfc4_nci_cmd_wq#163 ->(wq_completion)nfc3_nci_cmd_wq#419 ->(wq_completion)nfc3_nci_cmd_wq#420 ->(wq_completion)nfc3_nci_cmd_wq#421 ->(wq_completion)nfc2_nci_cmd_wq#909 ->(wq_completion)nfc3_nci_cmd_wq#422 ->(wq_completion)nfc4_nci_cmd_wq#164 ->(wq_completion)nfc2_nci_cmd_wq#910 ->(wq_completion)nfc3_nci_cmd_wq#423 ->(wq_completion)nfc4_nci_cmd_wq#165 ->(wq_completion)nfc5_nci_cmd_wq#55 ->(wq_completion)nfc2_nci_cmd_wq#911 ->(wq_completion)nfc3_nci_cmd_wq#425 ->(wq_completion)nfc2_nci_cmd_wq#912 ->(wq_completion)nfc2_nci_cmd_wq#913 ->(wq_completion)nfc3_nci_cmd_wq#426 ->(wq_completion)nfc4_nci_cmd_wq#166 ->(wq_completion)nfc2_nci_cmd_wq#914 ->(wq_completion)nfc3_nci_cmd_wq#427 ->(wq_completion)nfc2_nci_cmd_wq#916 ->(wq_completion)nfc3_nci_cmd_wq#428 ->(wq_completion)nfc2_nci_cmd_wq#917 ->(wq_completion)nfc4_nci_cmd_wq#167 ->(wq_completion)nfc3_nci_cmd_wq#429 ->(wq_completion)nfc5_nci_cmd_wq#56 ->(wq_completion)nfc2_nci_cmd_wq#918 ->(wq_completion)nfc3_nci_cmd_wq#430 ->(wq_completion)nfc5_nci_cmd_wq#57 ->(wq_completion)nfc4_nci_cmd_wq#168 ->(wq_completion)nfc6_nci_cmd_wq#25 ->(wq_completion)nfc2_nci_cmd_wq#919 ->(wq_completion)nfc2_nci_cmd_wq#920 ->(wq_completion)nfc2_nci_cmd_wq#921 ->(wq_completion)nfc2_nci_cmd_wq#922 ->(wq_completion)nfc2_nci_cmd_wq#923 ->(wq_completion)nfc2_nci_cmd_wq#924 ->(wq_completion)nfc3_nci_cmd_wq#431 ->(wq_completion)nfc2_nci_cmd_wq#925 ->(wq_completion)nfc4_nci_cmd_wq#169 ->(wq_completion)nfc2_nci_cmd_wq#926 ->(wq_completion)nfc5_nci_cmd_wq#58 ->(wq_completion)nfc3_nci_cmd_wq#432 ->(wq_completion)nfc2_nci_cmd_wq#927 ->(wq_completion)nfc3_nci_cmd_wq#433 ->(wq_completion)nfc4_nci_cmd_wq#170 ->(wq_completion)nfc2_nci_cmd_wq#928 ->(wq_completion)nfc2_nci_cmd_wq#929 ->(wq_completion)nfc3_nci_cmd_wq#434 ->(wq_completion)nfc2_nci_cmd_wq#930 ->(wq_completion)nfc2_nci_cmd_wq#931 ->(wq_completion)nfc3_nci_cmd_wq#435 ->(wq_completion)nfc2_nci_cmd_wq#932 ->(wq_completion)nfc2_nci_cmd_wq#933 ->(wq_completion)nfc3_nci_cmd_wq#436 ->(wq_completion)nfc4_nci_cmd_wq#171 ->(wq_completion)nfc2_nci_cmd_wq#934 ->(wq_completion)nfc2_nci_cmd_wq#935 ->(wq_completion)nfc3_nci_cmd_wq#437 ->(wq_completion)nfc2_nci_cmd_wq#936 ->(wq_completion)nfc2_nci_cmd_wq#937 ->(wq_completion)nfc2_nci_cmd_wq#939 ->(wq_completion)nfc2_nci_cmd_wq#940 ->(wq_completion)nfc2_nci_cmd_wq#941 ->(wq_completion)nfc3_nci_cmd_wq#438 ->(wq_completion)nfc3_nci_cmd_wq#439 ->(wq_completion)nfc2_nci_cmd_wq#942 ->(wq_completion)nfc2_nci_cmd_wq#943 ->(wq_completion)nfc2_nci_cmd_wq#944 ->(wq_completion)nfc2_nci_cmd_wq#945 ->(wq_completion)nfc3_nci_cmd_wq#440 ->(wq_completion)nfc2_nci_cmd_wq#946 ->(wq_completion)nfc4_nci_cmd_wq#172 ->(wq_completion)nfc2_nci_cmd_wq#947 ->(wq_completion)nfc2_nci_cmd_wq#948 ->(wq_completion)nfc3_nci_cmd_wq#441 ->(wq_completion)nfc2_nci_cmd_wq#949 ->(wq_completion)nfc2_nci_cmd_wq#950 ->(wq_completion)nfc2_nci_cmd_wq#951 ->(wq_completion)nfc2_nci_cmd_wq#952 ->(wq_completion)nfc2_nci_cmd_wq#953 ->(wq_completion)nfc3_nci_cmd_wq#442 ->(wq_completion)nfc2_nci_cmd_wq#954 ->(wq_completion)nfc2_nci_cmd_wq#955 ->(wq_completion)nfc3_nci_cmd_wq#443 ->(wq_completion)nfc4_nci_cmd_wq#173 ->(wq_completion)nfc26_nci_cmd_wq#19 ->(wq_completion)nfc35_nci_cmd_wq#14 ->(wq_completion)nfc3_nci_cmd_wq#444 ->(wq_completion)nfc5_nci_cmd_wq#59 ->(wq_completion)nfc2_nci_cmd_wq#956 ->(wq_completion)nfc6_nci_cmd_wq#27 ->(wq_completion)nfc7_nci_cmd_wq#20 ->(wq_completion)nfc8_nci_cmd_wq#20 ->(wq_completion)nfc5_nci_cmd_wq#60 ->(wq_completion)nfc4_nci_cmd_wq#174 ->(wq_completion)nfc9_nci_cmd_wq#19 ->(wq_completion)nfc10_nci_cmd_wq#16 ->(wq_completion)nfc11_nci_cmd_wq#20 ->(wq_completion)nfc12_nci_cmd_wq#17 ->(wq_completion)nfc13_nci_cmd_wq#14 ->(wq_completion)nfc14_nci_cmd_wq#17 ->(wq_completion)nfc15_nci_cmd_wq#15 ->(wq_completion)nfc16_nci_cmd_wq#17 ->(wq_completion)nfc17_nci_cmd_wq#14 ->(wq_completion)nfc18_nci_cmd_wq#16 ->(wq_completion)nfc14_nci_cmd_wq#18 ->(wq_completion)nfc19_nci_cmd_wq#15 ->(wq_completion)nfc20_nci_cmd_wq#16 ->(wq_completion)nfc21_nci_cmd_wq#14 ->(wq_completion)nfc22_nci_cmd_wq#15 ->(wq_completion)nfc23_nci_cmd_wq#17 ->(wq_completion)nfc24_nci_cmd_wq#14 ->(wq_completion)nfc25_nci_cmd_wq#14 ->(wq_completion)nfc27_nci_cmd_wq#16 ->(wq_completion)nfc28_nci_cmd_wq#15 ->(wq_completion)nfc29_nci_cmd_wq#16 ->(wq_completion)nfc42_nci_cmd_wq#6 ->(wq_completion)nfc3_nci_cmd_wq#445 ->(wq_completion)nfc2_nci_cmd_wq#957 ->(wq_completion)nfc6_nci_cmd_wq#28 ->(wq_completion)nfc7_nci_cmd_wq#21 ->(wq_completion)nfc44_nci_cmd_wq#4 ->(wq_completion)nfc4_nci_cmd_wq#175 ->(wq_completion)nfc5_nci_cmd_wq#61 ->(wq_completion)nfc8_nci_cmd_wq#21 ->(wq_completion)nfc9_nci_cmd_wq#20 ->(wq_completion)nfc10_nci_cmd_wq#17 ->(wq_completion)nfc11_nci_cmd_wq#21 ->(wq_completion)nfc12_nci_cmd_wq#18 ->(wq_completion)nfc13_nci_cmd_wq#15 ->(wq_completion)nfc15_nci_cmd_wq#16 ->(wq_completion)nfc16_nci_cmd_wq#18 ->(wq_completion)nfc17_nci_cmd_wq#15 ->(wq_completion)nfc14_nci_cmd_wq#19 ->(wq_completion)nfc18_nci_cmd_wq#17 ->(wq_completion)nfc19_nci_cmd_wq#16 ->(wq_completion)nfc20_nci_cmd_wq#17 ->(wq_completion)nfc21_nci_cmd_wq#15 ->(wq_completion)nfc45_nci_cmd_wq#4 ->(wq_completion)nfc43_nci_cmd_wq#5 ->(wq_completion)nfc41_nci_cmd_wq#6 ->(wq_completion)nfc40_nci_cmd_wq#8 ->(wq_completion)nfc35_nci_cmd_wq#15 ->(wq_completion)nfc39_nci_cmd_wq#7 ->(wq_completion)nfc38_nci_cmd_wq#7 ->(wq_completion)nfc37_nci_cmd_wq#8 ->(wq_completion)nfc36_nci_cmd_wq#9 ->(wq_completion)nfc34_nci_cmd_wq#13 ->(wq_completion)nfc33_nci_cmd_wq#15 ->(wq_completion)nfc32_nci_cmd_wq#13 ->(wq_completion)nfc31_nci_cmd_wq#13 ->(wq_completion)nfc30_nci_cmd_wq#14 ->(wq_completion)nfc26_nci_cmd_wq#20 ->(wq_completion)nfc22_nci_cmd_wq#16 ->(wq_completion)nfc2_nci_cmd_wq#958 ->(wq_completion)nfc2_nci_cmd_wq#959 ->(wq_completion)nfc2_nci_cmd_wq#960 ->(wq_completion)nfc2_nci_cmd_wq#961 ->(wq_completion)nfc2_nci_cmd_wq#962 ->(wq_completion)nfc2_nci_cmd_wq#963 ->(wq_completion)nfc2_nci_cmd_wq#964 ->(wq_completion)nfc2_nci_cmd_wq#965 ->(wq_completion)nfc2_nci_cmd_wq#966 ->(wq_completion)nfc2_nci_cmd_wq#967 ->(wq_completion)nfc2_nci_cmd_wq#968 ->(wq_completion)nfc2_nci_cmd_wq#969 ->(wq_completion)nfc2_nci_cmd_wq#970 ->(wq_completion)nfc2_nci_cmd_wq#971 ->(wq_completion)nfc2_nci_cmd_wq#972 ->(wq_completion)nfc2_nci_cmd_wq#973 ->(wq_completion)nfc2_nci_cmd_wq#974 ->(wq_completion)nfc2_nci_cmd_wq#975 ->(wq_completion)nfc3_nci_cmd_wq#446 ->(wq_completion)nfc4_nci_cmd_wq#176 ->(wq_completion)nfc2_nci_cmd_wq#976 ->(wq_completion)nfc3_nci_cmd_wq#447 ->(wq_completion)nfc4_nci_cmd_wq#177 ->(wq_completion)nfc5_nci_cmd_wq#62 ->(wq_completion)nfc2_nci_cmd_wq#977 ->(wq_completion)nfc2_nci_cmd_wq#978 ->(wq_completion)nfc3_nci_cmd_wq#448 ->(wq_completion)nfc2_nci_cmd_wq#979 ->(wq_completion)nfc3_nci_cmd_wq#449 ->(wq_completion)nfc2_nci_cmd_wq#980 ->(wq_completion)nfc2_nci_cmd_wq#981 ->(wq_completion)nfc3_nci_cmd_wq#450 ->(wq_completion)nfc4_nci_cmd_wq#178 ->(wq_completion)nfc4_nci_cmd_wq#179 ->(wq_completion)nfc2_nci_cmd_wq#982 ->(wq_completion)nfc3_nci_cmd_wq#451 ->(wq_completion)nfc2_nci_cmd_wq#983 ->(wq_completion)nfc2_nci_cmd_wq#984 ->(wq_completion)nfc2_nci_cmd_wq#985 ->(wq_completion)nfc3_nci_cmd_wq#452 ->(wq_completion)nfc2_nci_cmd_wq#986 ->(wq_completion)nfc2_nci_cmd_wq#987 ->(wq_completion)nfc2_nci_cmd_wq#988 ->(wq_completion)nfc3_nci_cmd_wq#453 ->(wq_completion)nfc2_nci_cmd_wq#989 ->(wq_completion)nfc2_nci_cmd_wq#990 ->(wq_completion)nfc3_nci_cmd_wq#454 ->(wq_completion)nfc2_nci_cmd_wq#991 ->(wq_completion)nfc3_nci_cmd_wq#455 ->(wq_completion)nfc2_nci_cmd_wq#992 ->(wq_completion)nfc2_nci_cmd_wq#993 ->(wq_completion)nfc2_nci_cmd_wq#994 ->(wq_completion)nfc3_nci_cmd_wq#456 ->(wq_completion)nfc2_nci_cmd_wq#995 ->(wq_completion)nfc2_nci_cmd_wq#996 ->(wq_completion)nfc2_nci_cmd_wq#997 ->(wq_completion)nfc2_nci_cmd_wq#998 ->(wq_completion)nfc2_nci_cmd_wq#999 ->(wq_completion)nfc2_nci_cmd_wq#1000 ->(wq_completion)nfc3_nci_cmd_wq#457 ->(wq_completion)nfc4_nci_cmd_wq#180 ->(wq_completion)nfc2_nci_cmd_wq#1001 ->(wq_completion)nfc3_nci_cmd_wq#458 ->(wq_completion)nfc4_nci_cmd_wq#181 ->(wq_completion)nfc2_nci_cmd_wq#1002 ->(wq_completion)nfc2_nci_cmd_wq#1004 ->(wq_completion)nfc4_nci_cmd_wq#182 ->(wq_completion)nfc3_nci_cmd_wq#459 ->(wq_completion)nfc2_nci_cmd_wq#1005 ->(wq_completion)nfc2_nci_cmd_wq#1006 ->(wq_completion)nfc2_nci_cmd_wq#1007 ->(wq_completion)nfc3_nci_cmd_wq#460 ->(wq_completion)nfc2_nci_cmd_wq#1008 ->(wq_completion)nfc2_nci_cmd_wq#1009 ->(wq_completion)nfc3_nci_cmd_wq#461 ->(wq_completion)nfc3_nci_cmd_wq#462 ->(wq_completion)nfc3_nci_cmd_wq#463 ->(wq_completion)nfc3_nci_cmd_wq#465 ->(wq_completion)nfc3_nci_cmd_wq#466 ->(wq_completion)nfc3_nci_cmd_wq#467 ->(wq_completion)nfc4_nci_cmd_wq#183 ->(wq_completion)nfc2_nci_cmd_wq#1010 ->(wq_completion)nfc3_nci_cmd_wq#468 ->(wq_completion)nfc2_nci_cmd_wq#1011 ->(wq_completion)nfc3_nci_cmd_wq#469 ->(wq_completion)nfc2_nci_cmd_wq#1012 ->(wq_completion)nfc3_nci_cmd_wq#470 ->(wq_completion)nfc2_nci_cmd_wq#1013 ->(wq_completion)nfc3_nci_cmd_wq#471 ->(wq_completion)nfc4_nci_cmd_wq#184 ->(wq_completion)nfc2_nci_cmd_wq#1014 ->(wq_completion)nfc3_nci_cmd_wq#472 ->(wq_completion)nfc4_nci_cmd_wq#185 ->(wq_completion)nfc6_nci_cmd_wq#29 ->(wq_completion)nfc5_nci_cmd_wq#63 ->(wq_completion)nfc2_nci_cmd_wq#1015 ->(wq_completion)nfc2_nci_cmd_wq#1017 ->(wq_completion)nfc2_nci_cmd_wq#1018 ->(wq_completion)nfc2_nci_cmd_wq#1019 ->(wq_completion)nfc2_nci_cmd_wq#1020 ->(wq_completion)nfc3_nci_cmd_wq#473 ->(wq_completion)nfc4_nci_cmd_wq#186 ->(wq_completion)nfc5_nci_cmd_wq#64 ->(wq_completion)nfc4_nci_cmd_wq#187 ->(wq_completion)nfc5_nci_cmd_wq#65 ->(wq_completion)nfc4_nci_cmd_wq#188 ->(wq_completion)nfc2_nci_cmd_wq#1021 ->(wq_completion)nfc3_nci_cmd_wq#474 ->(wq_completion)nfc2_nci_cmd_wq#1022 ->(wq_completion)nfc2_nci_cmd_wq#1023 ->(wq_completion)nfc3_nci_cmd_wq#475 ->(wq_completion)nfc4_nci_cmd_wq#189 ->(wq_completion)nfc2_nci_cmd_wq#1024 ->(wq_completion)nfc2_nci_cmd_wq#1026 ->(wq_completion)nfc3_nci_cmd_wq#476 ->(wq_completion)nfc4_nci_cmd_wq#190 ->(wq_completion)nfc2_nci_cmd_wq#1027 ->(wq_completion)nfc3_nci_cmd_wq#477 ->(wq_completion)nfc2_nci_cmd_wq#1028 ->(wq_completion)nfc3_nci_cmd_wq#478 ->(wq_completion)nfc2_nci_cmd_wq#1029 ->(wq_completion)nfc2_nci_cmd_wq#1030 ->(wq_completion)nfc2_nci_cmd_wq#1031 ->(wq_completion)nfc2_nci_cmd_wq#1032 ->(wq_completion)nfc3_nci_cmd_wq#479 ->(wq_completion)nfc2_nci_cmd_wq#1033 ->(wq_completion)nfc3_nci_cmd_wq#480 ->(wq_completion)nfc2_nci_cmd_wq#1034 ->(wq_completion)nfc2_nci_cmd_wq#1035 ->(wq_completion)nfc3_nci_cmd_wq#481 ->(wq_completion)nfc2_nci_cmd_wq#1036 ->(wq_completion)nfc3_nci_cmd_wq#482 ->(wq_completion)nfc2_nci_cmd_wq#1038 ->(wq_completion)nfc2_nci_cmd_wq#1039 ->(wq_completion)nfc2_nci_cmd_wq#1040 ->(wq_completion)nfc2_nci_cmd_wq#1041 ->(wq_completion)nfc3_nci_cmd_wq#483 ->(wq_completion)nfc2_nci_cmd_wq#1042 ->(wq_completion)nfc3_nci_cmd_wq#484 ->(wq_completion)nfc2_nci_cmd_wq#1043 ->(wq_completion)nfc4_nci_cmd_wq#191 ->(wq_completion)nfc3_nci_cmd_wq#485 ->(wq_completion)nfc2_nci_cmd_wq#1044 ->(wq_completion)nfc5_nci_cmd_wq#66 ->(wq_completion)nfc4_nci_cmd_wq#192 ->(wq_completion)nfc4_nci_cmd_wq#193 ->(wq_completion)nfc3_nci_cmd_wq#486 ->(wq_completion)nfc2_nci_cmd_wq#1045 ->(wq_completion)nfc5_nci_cmd_wq#67 ->(wq_completion)nfc6_nci_cmd_wq#31 ->(wq_completion)nfc2_nci_cmd_wq#1046 ->(wq_completion)nfc2_nci_cmd_wq#1047 ->(wq_completion)nfc3_nci_cmd_wq#487 ->(wq_completion)nfc2_nci_cmd_wq#1048 ->(wq_completion)nfc3_nci_cmd_wq#488 ->(wq_completion)nfc4_nci_cmd_wq#194 ->(wq_completion)nfc2_nci_cmd_wq#1049 ->(wq_completion)nfc3_nci_cmd_wq#489 ->(wq_completion)nfc2_nci_cmd_wq#1050 ->(wq_completion)nfc2_nci_cmd_wq#1051 ->(wq_completion)nfc3_nci_cmd_wq#490 ->(wq_completion)nfc4_nci_cmd_wq#195 ->(wq_completion)nfc2_nci_cmd_wq#1052 ->(wq_completion)nfc3_nci_cmd_wq#491 ->(wq_completion)nfc4_nci_cmd_wq#196 ->(wq_completion)nfc2_nci_cmd_wq#1053 ->(wq_completion)nfc3_nci_cmd_wq#492 ->(wq_completion)nfc2_nci_cmd_wq#1054 ->(wq_completion)nfc4_nci_cmd_wq#197 ->(wq_completion)nfc2_nci_cmd_wq#1055 ->(wq_completion)nfc3_nci_cmd_wq#493 ->(wq_completion)nfc5_nci_cmd_wq#68 ->(wq_completion)nfc7_nci_cmd_wq#22 ->(wq_completion)nfc6_nci_cmd_wq#32 ->(wq_completion)nfc4_nci_cmd_wq#198 ->(wq_completion)nfc2_nci_cmd_wq#1056 ->(wq_completion)nfc2_nci_cmd_wq#1057 ->(wq_completion)nfc2_nci_cmd_wq#1058 ->(wq_completion)nfc4_nci_cmd_wq#199 ->(wq_completion)nfc3_nci_cmd_wq#494 ->(wq_completion)nfc5_nci_cmd_wq#69 ->(wq_completion)nfc4_nci_cmd_wq#200 ->(wq_completion)nfc2_nci_cmd_wq#1059 ->(wq_completion)nfc3_nci_cmd_wq#495 ->(wq_completion)nfc5_nci_cmd_wq#70 ->(wq_completion)nfc6_nci_cmd_wq#33 ->(wq_completion)nfc3_nci_cmd_wq#496 ->(wq_completion)nfc4_nci_cmd_wq#201 ->(wq_completion)nfc2_nci_cmd_wq#1060 ->(wq_completion)nfc4_nci_cmd_wq#202 ->(wq_completion)nfc3_nci_cmd_wq#497 ->(wq_completion)nfc2_nci_cmd_wq#1061 ->(wq_completion)nfc3_nci_cmd_wq#498 ->(wq_completion)nfc5_nci_cmd_wq#71 ->(wq_completion)nfc2_nci_cmd_wq#1062 ->(wq_completion)nfc4_nci_cmd_wq#203 ->(wq_completion)nfc3_nci_cmd_wq#499 ->(wq_completion)nfc5_nci_cmd_wq#72 ->(wq_completion)nfc3_nci_cmd_wq#500 ->(wq_completion)nfc2_nci_cmd_wq#1063 ->(wq_completion)nfc4_nci_cmd_wq#204 ->(wq_completion)nfc2_nci_cmd_wq#1064 ->(wq_completion)nfc3_nci_cmd_wq#501 ->(wq_completion)nfc4_nci_cmd_wq#205 ->(wq_completion)nfc2_nci_cmd_wq#1065 ->(wq_completion)nfc3_nci_cmd_wq#502 ->(wq_completion)nfc2_nci_cmd_wq#1066 ->(wq_completion)nfc2_nci_cmd_wq#1067 ->(wq_completion)nfc3_nci_cmd_wq#503 ->(wq_completion)nfc2_nci_cmd_wq#1068 ->(wq_completion)nfc3_nci_cmd_wq#504 ->(wq_completion)nfc2_nci_cmd_wq#1069 ->(wq_completion)nfc2_nci_cmd_wq#1070 ->(wq_completion)nfc3_nci_cmd_wq#505 ->(wq_completion)nfc4_nci_cmd_wq#206 ->(wq_completion)nfc2_nci_cmd_wq#1071 ->(wq_completion)nfc3_nci_cmd_wq#506 ->(wq_completion)nfc2_nci_cmd_wq#1072 ->(wq_completion)nfc4_nci_cmd_wq#207 ->(wq_completion)nfc2_nci_cmd_wq#1073 ->(wq_completion)nfc3_nci_cmd_wq#507 ->(wq_completion)nfc2_nci_cmd_wq#1074 ->(wq_completion)nfc3_nci_cmd_wq#508 ->(wq_completion)nfc3_nci_cmd_wq#509 ->(wq_completion)nfc2_nci_cmd_wq#1075 ->(wq_completion)nfc3_nci_cmd_wq#511 ->(wq_completion)nfc2_nci_cmd_wq#1076 ->(wq_completion)nfc2_nci_cmd_wq#1077 ->(wq_completion)nfc2_nci_cmd_wq#1078 ->(wq_completion)nfc3_nci_cmd_wq#512 ->(wq_completion)nfc3_nci_cmd_wq#513 ->(wq_completion)nfc2_nci_cmd_wq#1079 ->(wq_completion)nfc4_nci_cmd_wq#208 ->(wq_completion)nfc2_nci_cmd_wq#1080 ->(wq_completion)nfc3_nci_cmd_wq#514 ->(wq_completion)nfc4_nci_cmd_wq#209 ->(wq_completion)nfc3_nci_cmd_wq#515 ->(wq_completion)nfc2_nci_cmd_wq#1082 ->(wq_completion)nfc4_nci_cmd_wq#210 ->(wq_completion)nfc5_nci_cmd_wq#73 ->(wq_completion)nfc2_nci_cmd_wq#1083 ->(wq_completion)nfc2_nci_cmd_wq#1084 ->(wq_completion)nfc3_nci_cmd_wq#516 ->(wq_completion)nfc2_nci_cmd_wq#1085 ->(wq_completion)nfc4_nci_cmd_wq#211 ->(wq_completion)nfc2_nci_cmd_wq#1087 ->(wq_completion)nfc3_nci_cmd_wq#517 ->(wq_completion)nfc2_nci_cmd_wq#1088 ->(wq_completion)nfc3_nci_cmd_wq#518 ->(wq_completion)nfc4_nci_cmd_wq#212 ->(wq_completion)nfc2_nci_cmd_wq#1089 ->(wq_completion)nfc3_nci_cmd_wq#519 ->(wq_completion)nfc2_nci_cmd_wq#1090 ->(wq_completion)nfc2_nci_cmd_wq#1091 ->(wq_completion)nfc3_nci_cmd_wq#520 ->(wq_completion)nfc2_nci_cmd_wq#1092 ->(wq_completion)nfc3_nci_cmd_wq#521 ->(wq_completion)nfc2_nci_cmd_wq#1093 ->(wq_completion)nfc2_nci_cmd_wq#1094 ->(wq_completion)nfc5_nci_cmd_wq#74 ->(wq_completion)nfc3_nci_cmd_wq#522 ->(wq_completion)nfc4_nci_cmd_wq#213 ->(wq_completion)nfc3_nci_cmd_wq#523 ->(wq_completion)nfc4_nci_cmd_wq#214 ->(wq_completion)nfc2_nci_cmd_wq#1095 ->(wq_completion)nfc3_nci_cmd_wq#524 ->(wq_completion)nfc2_nci_cmd_wq#1096 ->(wq_completion)nfc4_nci_cmd_wq#215 ->(wq_completion)nfc5_nci_cmd_wq#75 ->(wq_completion)nfc2_nci_cmd_wq#1097 ->(wq_completion)nfc3_nci_cmd_wq#525 ->(wq_completion)nfc2_nci_cmd_wq#1098 ->(wq_completion)nfc3_nci_cmd_wq#526 ->(wq_completion)nfc4_nci_cmd_wq#216 ->(wq_completion)nfc2_nci_cmd_wq#1099 ->(wq_completion)nfc2_nci_cmd_wq#1100 ->(wq_completion)nfc3_nci_cmd_wq#527 ->(wq_completion)nfc2_nci_cmd_wq#1101 ->(wq_completion)nfc3_nci_cmd_wq#528 ->(wq_completion)nfc2_nci_cmd_wq#1102 ->(wq_completion)nfc3_nci_cmd_wq#529 ->(wq_completion)nfc4_nci_cmd_wq#217 ->(wq_completion)nfc2_nci_cmd_wq#1103 ->(wq_completion)nfc2_nci_cmd_wq#1105 ->(wq_completion)nfc3_nci_cmd_wq#530 ->(wq_completion)nfc2_nci_cmd_wq#1106 ->(wq_completion)nfc3_nci_cmd_wq#531 ->(wq_completion)nfc4_nci_cmd_wq#218 ->(wq_completion)nfc5_nci_cmd_wq#76 ->(wq_completion)nfc2_nci_cmd_wq#1107 ->(wq_completion)nfc3_nci_cmd_wq#532 ->(wq_completion)nfc4_nci_cmd_wq#219 ->(wq_completion)nfc2_nci_cmd_wq#1108 ->(wq_completion)nfc2_nci_cmd_wq#1109 ->(wq_completion)nfc2_nci_cmd_wq#1110 ->(wq_completion)nfc3_nci_cmd_wq#533 ->(wq_completion)nfc2_nci_cmd_wq#1111 ->(wq_completion)nfc4_nci_cmd_wq#220 ->(wq_completion)nfc3_nci_cmd_wq#534 ->(wq_completion)nfc4_nci_cmd_wq#221 ->(wq_completion)nfc2_nci_cmd_wq#1112 ->(wq_completion)nfc3_nci_cmd_wq#535 ->(wq_completion)nfc2_nci_cmd_wq#1113 ->(wq_completion)nfc2_nci_cmd_wq#1114 ->(wq_completion)nfc3_nci_cmd_wq#536 ->(wq_completion)nfc4_nci_cmd_wq#222 ->(wq_completion)nfc3_nci_cmd_wq#537 ->(wq_completion)nfc2_nci_cmd_wq#1115 ->(wq_completion)nfc5_nci_cmd_wq#77 ->(wq_completion)nfc2_nci_cmd_wq#1116 ->(wq_completion)nfc6_nci_cmd_wq#34 ->(wq_completion)nfc2_nci_cmd_wq#1117 ->(wq_completion)nfc4_nci_cmd_wq#223 ->(wq_completion)nfc3_nci_cmd_wq#538 ->(wq_completion)nfc2_nci_cmd_wq#1118 ->(wq_completion)nfc4_nci_cmd_wq#224 ->(wq_completion)nfc2_nci_cmd_wq#1119 ->(wq_completion)nfc3_nci_cmd_wq#539 ->(wq_completion)nfc4_nci_cmd_wq#225 ->(wq_completion)nfc2_nci_cmd_wq#1120 ->(wq_completion)nfc3_nci_cmd_wq#541 ->(wq_completion)nfc4_nci_cmd_wq#226 ->(wq_completion)nfc5_nci_cmd_wq#78 ->(wq_completion)nfc2_nci_cmd_wq#1121 ->(wq_completion)nfc3_nci_cmd_wq#542 ->(wq_completion)nfc3_nci_cmd_wq#543 ->(wq_completion)nfc4_nci_cmd_wq#227 ->(wq_completion)nfc2_nci_cmd_wq#1122 ->(wq_completion)nfc5_nci_cmd_wq#79 ->(wq_completion)nfc3_nci_cmd_wq#544 ->(wq_completion)nfc6_nci_cmd_wq#35 ->(wq_completion)nfc2_nci_cmd_wq#1123 ->(wq_completion)nfc2_nci_cmd_wq#1124 ->(wq_completion)nfc3_nci_cmd_wq#545 ->(wq_completion)nfc2_nci_cmd_wq#1126 ->(wq_completion)nfc2_nci_cmd_wq#1127 ->(wq_completion)nfc2_nci_cmd_wq#1128 ->(wq_completion)nfc3_nci_cmd_wq#546 ->(wq_completion)nfc4_nci_cmd_wq#228 ->(wq_completion)nfc2_nci_cmd_wq#1129 ->(wq_completion)nfc3_nci_cmd_wq#547 ->(wq_completion)nfc5_nci_cmd_wq#80 ->(wq_completion)nfc2_nci_cmd_wq#1130 ->(wq_completion)nfc3_nci_cmd_wq#548 ->(wq_completion)nfc3_nci_cmd_wq#549 ->(wq_completion)nfc2_nci_cmd_wq#1131 ->(wq_completion)nfc2_nci_cmd_wq#1132 ->(wq_completion)nfc2_nci_cmd_wq#1133 ->(wq_completion)nfc3_nci_cmd_wq#550 ->(wq_completion)nfc2_nci_cmd_wq#1134 ->(wq_completion)nfc3_nci_cmd_wq#551 ->(wq_completion)nfc2_nci_cmd_wq#1135 ->(wq_completion)nfc2_nci_cmd_wq#1136 ->(wq_completion)nfc2_nci_cmd_wq#1137 ->(wq_completion)nfc3_nci_cmd_wq#552 ->(wq_completion)nfc4_nci_cmd_wq#229 ->(wq_completion)nfc2_nci_cmd_wq#1138 ->(wq_completion)nfc3_nci_cmd_wq#553 ->(wq_completion)nfc4_nci_cmd_wq#230 ->(wq_completion)nfc2_nci_cmd_wq#1139 ->(wq_completion)nfc2_nci_cmd_wq#1140 ->(wq_completion)nfc2_nci_cmd_wq#1141 ->(wq_completion)nfc3_nci_cmd_wq#554 ->(wq_completion)nfc2_nci_cmd_wq#1143 ->(wq_completion)nfc2_nci_cmd_wq#1144 ->(wq_completion)nfc2_nci_cmd_wq#1145 ->(wq_completion)nfc2_nci_cmd_wq#1146 ->(wq_completion)nfc4_nci_cmd_wq#231 ->(wq_completion)nfc5_nci_cmd_wq#82 ->(wq_completion)nfc3_nci_cmd_wq#555 ->(wq_completion)nfc6_nci_cmd_wq#36 ->(wq_completion)nfc2_nci_cmd_wq#1147 ->(wq_completion)nfc3_nci_cmd_wq#556 ->(wq_completion)nfc2_nci_cmd_wq#1148 ->(wq_completion)nfc3_nci_cmd_wq#557 ->(wq_completion)nfc2_nci_cmd_wq#1149 ->(wq_completion)nfc3_nci_cmd_wq#558 ->(wq_completion)nfc2_nci_cmd_wq#1150 ->(wq_completion)nfc2_nci_cmd_wq#1151 ->(wq_completion)nfc3_nci_cmd_wq#559 ->(wq_completion)nfc4_nci_cmd_wq#233 ->(wq_completion)nfc2_nci_cmd_wq#1152 ->(wq_completion)nfc3_nci_cmd_wq#560 ->(wq_completion)nfc5_nci_cmd_wq#83 ->(wq_completion)nfc2_nci_cmd_wq#1153 ->(wq_completion)nfc4_nci_cmd_wq#234 ->(wq_completion)nfc3_nci_cmd_wq#561 ->(wq_completion)nfc2_nci_cmd_wq#1154 ->(wq_completion)nfc5_nci_cmd_wq#84 ->(wq_completion)nfc3_nci_cmd_wq#562 ->(wq_completion)nfc4_nci_cmd_wq#235 ->(wq_completion)nfc2_nci_cmd_wq#1155 ->(wq_completion)nfc3_nci_cmd_wq#563 ->(wq_completion)nfc4_nci_cmd_wq#236 ->(wq_completion)nfc2_nci_cmd_wq#1156 ->(wq_completion)nfc3_nci_cmd_wq#565 ->(wq_completion)nfc2_nci_cmd_wq#1157 ->(wq_completion)nfc4_nci_cmd_wq#237 ->(wq_completion)nfc2_nci_cmd_wq#1158 ->(wq_completion)nfc3_nci_cmd_wq#566 ->(wq_completion)nfc6_nci_cmd_wq#37 ->(wq_completion)nfc4_nci_cmd_wq#238 ->(wq_completion)nfc5_nci_cmd_wq#86 ->(wq_completion)nfc3_nci_cmd_wq#567 ->(wq_completion)nfc4_nci_cmd_wq#239 ->(wq_completion)nfc2_nci_cmd_wq#1160 ->(wq_completion)nfc5_nci_cmd_wq#87 ->(wq_completion)nfc3_nci_cmd_wq#568 ->(wq_completion)nfc2_nci_cmd_wq#1161 ->(wq_completion)nfc4_nci_cmd_wq#240 ->(wq_completion)nfc2_nci_cmd_wq#1162 ->(wq_completion)nfc2_nci_cmd_wq#1163 ->(wq_completion)nfc3_nci_cmd_wq#569 ->(wq_completion)nfc3_nci_cmd_wq#570 ->(wq_completion)nfc2_nci_cmd_wq#1164 ->(wq_completion)nfc5_nci_cmd_wq#88 ->(wq_completion)nfc4_nci_cmd_wq#241 ->(wq_completion)nfc2_nci_cmd_wq#1165 ->(wq_completion)nfc3_nci_cmd_wq#571 ->(wq_completion)nfc2_nci_cmd_wq#1166 ->(wq_completion)nfc3_nci_cmd_wq#573 ->(wq_completion)nfc4_nci_cmd_wq#242 ->(wq_completion)nfc6_nci_cmd_wq#38 ->(wq_completion)nfc5_nci_cmd_wq#89 ->(wq_completion)nfc2_nci_cmd_wq#1167 ->(wq_completion)nfc3_nci_cmd_wq#574 ->(wq_completion)nfc2_nci_cmd_wq#1168 ->(wq_completion)nfc2_nci_cmd_wq#1170 ->(wq_completion)nfc3_nci_cmd_wq#577 ->(wq_completion)nfc2_nci_cmd_wq#1171 ->(wq_completion)nfc2_nci_cmd_wq#1172 ->(wq_completion)nfc2_nci_cmd_wq#1173 ->(wq_completion)nfc3_nci_cmd_wq#578 ->(wq_completion)nfc2_nci_cmd_wq#1174 ->(wq_completion)nfc4_nci_cmd_wq#243 ->(wq_completion)nfc3_nci_cmd_wq#579 ->(wq_completion)nfc2_nci_cmd_wq#1175 ->(wq_completion)nfc3_nci_cmd_wq#580 ->(wq_completion)nfc4_nci_cmd_wq#244 ->(wq_completion)nfc2_nci_cmd_wq#1176 ->(wq_completion)nfc2_nci_cmd_wq#1177 ->(wq_completion)nfc3_nci_cmd_wq#581 ->(wq_completion)nfc2_nci_cmd_wq#1178 ->(wq_completion)nfc3_nci_cmd_wq#582 ->(wq_completion)nfc4_nci_cmd_wq#245 ->(wq_completion)nfc4_nci_cmd_wq#246 ->(wq_completion)nfc3_nci_cmd_wq#583 ->(wq_completion)nfc2_nci_cmd_wq#1179 ->(wq_completion)nfc5_nci_cmd_wq#90 ->(wq_completion)nfc2_nci_cmd_wq#1180 ->(wq_completion)nfc3_nci_cmd_wq#584 ->(wq_completion)nfc2_nci_cmd_wq#1181 ->(wq_completion)nfc2_nci_cmd_wq#1182 ->(wq_completion)nfc4_nci_cmd_wq#247 ->(wq_completion)nfc3_nci_cmd_wq#585 ->(wq_completion)nfc2_nci_cmd_wq#1183 ->(wq_completion)nfc4_nci_cmd_wq#248 ->(wq_completion)nfc3_nci_cmd_wq#586 ->(wq_completion)nfc2_nci_cmd_wq#1184 ->(wq_completion)nfc3_nci_cmd_wq#587 ->(wq_completion)nfc2_nci_cmd_wq#1185 ->(wq_completion)nfc4_nci_cmd_wq#249 ->(wq_completion)nfc2_nci_cmd_wq#1186 ->(wq_completion)nfc3_nci_cmd_wq#588 ->(wq_completion)nfc2_nci_cmd_wq#1187 ->(wq_completion)nfc2_nci_cmd_wq#1188 ->(wq_completion)nfc3_nci_cmd_wq#589 ->(wq_completion)nfc2_nci_cmd_wq#1189 ->(wq_completion)nfc2_nci_cmd_wq#1190 ->(wq_completion)nfc3_nci_cmd_wq#590 ->(wq_completion)nfc2_nci_cmd_wq#1191 ->(wq_completion)nfc3_nci_cmd_wq#591 ->(wq_completion)nfc2_nci_cmd_wq#1192 ->(wq_completion)nfc3_nci_cmd_wq#592 ->(wq_completion)nfc2_nci_cmd_wq#1193 ->(wq_completion)nfc2_nci_cmd_wq#1194 ->(wq_completion)nfc2_nci_cmd_wq#1195 ->(wq_completion)nfc2_nci_cmd_wq#1196 ->(wq_completion)nfc3_nci_cmd_wq#593 ->(wq_completion)nfc2_nci_cmd_wq#1197 ->(wq_completion)nfc3_nci_cmd_wq#594 ->(wq_completion)nfc4_nci_cmd_wq#250 ->(wq_completion)nfc2_nci_cmd_wq#1198 ->(wq_completion)nfc3_nci_cmd_wq#595 ->(wq_completion)nfc2_nci_cmd_wq#1199 ->(wq_completion)nfc3_nci_cmd_wq#596 ->(wq_completion)nfc3_nci_cmd_wq#597 ->(wq_completion)nfc4_nci_cmd_wq#251 ->(wq_completion)nfc3_nci_cmd_wq#599 ->(wq_completion)nfc2_nci_cmd_wq#1200 ->(wq_completion)nfc2_nci_cmd_wq#1201 ->(wq_completion)nfc3_nci_cmd_wq#600 ->(wq_completion)nfc2_nci_cmd_wq#1202 ->(wq_completion)nfc3_nci_cmd_wq#601 ->(wq_completion)nfc2_nci_cmd_wq#1203 ->(wq_completion)nfc3_nci_cmd_wq#602 ->(wq_completion)nfc2_nci_cmd_wq#1204 ->(wq_completion)nfc4_nci_cmd_wq#252 ->(wq_completion)nfc2_nci_cmd_wq#1205 ->(wq_completion)nfc2_nci_cmd_wq#1206 ->(wq_completion)nfc3_nci_cmd_wq#603 ->(wq_completion)nfc8_nci_cmd_wq#23 ->(wq_completion)nfc9_nci_cmd_wq#21 ->(wq_completion)nfc21_nci_cmd_wq#16 ->(wq_completion)nfc19_nci_cmd_wq#17 ->(wq_completion)nfc19_nci_cmd_wq#18 ->(wq_completion)nfc34_nci_cmd_wq#14 ->(wq_completion)nfc19_nci_cmd_wq#19 ->(wq_completion)nfc33_nci_cmd_wq#16 ->(wq_completion)nfc32_nci_cmd_wq#14 ->(wq_completion)nfc31_nci_cmd_wq#14 ->(wq_completion)nfc35_nci_cmd_wq#16 ->(wq_completion)nfc30_nci_cmd_wq#15 ->(wq_completion)nfc19_nci_cmd_wq#20 ->(wq_completion)nfc29_nci_cmd_wq#17 ->(wq_completion)nfc28_nci_cmd_wq#16 ->(wq_completion)nfc27_nci_cmd_wq#17 ->(wq_completion)nfc34_nci_cmd_wq#15 ->(wq_completion)nfc26_nci_cmd_wq#21 ->(wq_completion)nfc25_nci_cmd_wq#15 ->(wq_completion)nfc24_nci_cmd_wq#15 ->(wq_completion)nfc23_nci_cmd_wq#18 ->(wq_completion)nfc21_nci_cmd_wq#17 ->(wq_completion)nfc22_nci_cmd_wq#17 ->(wq_completion)nfc20_nci_cmd_wq#18 ->(wq_completion)nfc18_nci_cmd_wq#18 ->(wq_completion)nfc17_nci_cmd_wq#16 ->(wq_completion)nfc19_nci_cmd_wq#21 ->(wq_completion)nfc16_nci_cmd_wq#19 ->(wq_completion)nfc15_nci_cmd_wq#17 ->(wq_completion)nfc14_nci_cmd_wq#20 ->(wq_completion)nfc15_nci_cmd_wq#18 ->(wq_completion)nfc9_nci_cmd_wq#22 ->(wq_completion)nfc13_nci_cmd_wq#16 ->(wq_completion)nfc12_nci_cmd_wq#19 ->(wq_completion)nfc8_nci_cmd_wq#24 ->(wq_completion)nfc11_nci_cmd_wq#22 ->(wq_completion)nfc10_nci_cmd_wq#18 ->(wq_completion)nfc8_nci_cmd_wq#25 ->(wq_completion)nfc7_nci_cmd_wq#23 ->(wq_completion)nfc6_nci_cmd_wq#39 ->(wq_completion)nfc5_nci_cmd_wq#91 ->(wq_completion)nfc4_nci_cmd_wq#253 ->(wq_completion)nfc3_nci_cmd_wq#604 ->(wq_completion)nfc2_nci_cmd_wq#1207 ->(wq_completion)nfc2_nci_cmd_wq#1208 ->(wq_completion)nfc5_nci_cmd_wq#92 ->(wq_completion)nfc2_nci_cmd_wq#1209 ->(wq_completion)nfc4_nci_cmd_wq#254 ->(wq_completion)nfc3_nci_cmd_wq#605 ->(wq_completion)nfc2_nci_cmd_wq#1210 ->(wq_completion)nfc3_nci_cmd_wq#606 ->(wq_completion)nfc4_nci_cmd_wq#255 ->(wq_completion)nfc2_nci_cmd_wq#1212 ->(wq_completion)nfc2_nci_cmd_wq#1213 ->(wq_completion)nfc2_nci_cmd_wq#1214 ->(wq_completion)nfc3_nci_cmd_wq#607 ->(wq_completion)nfc2_nci_cmd_wq#1215 ->(wq_completion)nfc4_nci_cmd_wq#256 ->(wq_completion)nfc3_nci_cmd_wq#608 ->(wq_completion)nfc2_nci_cmd_wq#1216 ->(wq_completion)nfc4_nci_cmd_wq#257 ->(wq_completion)nfc2_nci_cmd_wq#1217 ->(wq_completion)nfc3_nci_cmd_wq#609 ->(wq_completion)nfc3_nci_cmd_wq#610 ->(wq_completion)nfc2_nci_cmd_wq#1218 ->(wq_completion)nfc6_nci_cmd_wq#40 ->(wq_completion)nfc4_nci_cmd_wq#258 ->(wq_completion)nfc3_nci_cmd_wq#611 ->(wq_completion)nfc2_nci_cmd_wq#1219 ->(wq_completion)nfc5_nci_cmd_wq#93 ->(wq_completion)nfc3_nci_cmd_wq#613 ->(wq_completion)nfc3_nci_cmd_wq#614 ->(wq_completion)nfc2_nci_cmd_wq#1220 ->(wq_completion)nfc4_nci_cmd_wq#261 ->(wq_completion)nfc2_nci_cmd_wq#1221 ->(wq_completion)nfc3_nci_cmd_wq#615 ->(wq_completion)nfc4_nci_cmd_wq#262 ->(wq_completion)nfc2_nci_cmd_wq#1222 ->(wq_completion)nfc2_nci_cmd_wq#1223 ->(wq_completion)nfc3_nci_cmd_wq#616 ->(wq_completion)nfc4_nci_cmd_wq#263 ->(wq_completion)nfc3_nci_cmd_wq#617 ->(wq_completion)nfc2_nci_cmd_wq#1224 ->(wq_completion)nfc4_nci_cmd_wq#264 ->(wq_completion)nfc5_nci_cmd_wq#94 ->(wq_completion)nfc2_nci_cmd_wq#1225 ->(wq_completion)nfc3_nci_cmd_wq#618 ->(wq_completion)nfc2_nci_cmd_wq#1226 ->(wq_completion)nfc2_nci_cmd_wq#1227 ->(wq_completion)nfc2_nci_cmd_wq#1228 ->(wq_completion)nfc3_nci_cmd_wq#619 ->(wq_completion)nfc4_nci_cmd_wq#267 ->(wq_completion)nfc3_nci_cmd_wq#620 ->(wq_completion)nfc2_nci_cmd_wq#1229 ->(wq_completion)nfc2_nci_cmd_wq#1230 ->(wq_completion)nfc2_nci_cmd_wq#1231 ->(wq_completion)nfc2_nci_cmd_wq#1232 ->(wq_completion)nfc4_nci_cmd_wq#268 ->(wq_completion)nfc3_nci_cmd_wq#621 ->(wq_completion)nfc5_nci_cmd_wq#95 ->(wq_completion)nfc2_nci_cmd_wq#1233 FD: 1 BD: 61 +.-.: &r->producer_lock#3 FD: 73 BD: 1 ++++: &type->s_umount_key#47 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key ->&s->s_inode_list_lock ->&sbinfo->stat_lock ->&xa->xa_lock#8 ->pool_lock#2 ->&fsnotify_mark_srcu ->&simple_offset_xa_lock ->&rq->__lock ->&dentry->d_lock/1 ->percpu_counters_lock ->pcpu_lock FD: 1 BD: 1 .+.+: &type->s_umount_key#48 FD: 25 BD: 1 .+.+: &type->s_umount_key#49 ->&rq->__lock FD: 25 BD: 1 ++++: &type->s_umount_key#50 ->&rq->__lock FD: 25 BD: 1 .+.+: &type->s_umount_key#51 ->&rq->__lock FD: 1 BD: 1 .+.+: &type->s_umount_key#52 FD: 1 BD: 1 .+.+: &type->s_umount_key#53 FD: 25 BD: 1 .+.+: &type->s_umount_key#54 ->&rq->__lock FD: 25 BD: 1 .+.+: &type->s_umount_key#55 ->&rq->__lock FD: 25 BD: 1 .+.+: &type->s_umount_key#56 ->&rq->__lock FD: 70 BD: 1 ++++: &type->s_umount_key#57 ->&rq->__lock ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#16 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 ->pool_lock#2 FD: 1 BD: 1 .+.+: &type->s_umount_key#58 FD: 69 BD: 3 ++++: &type->s_umount_key#59 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->pool_lock#2 ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#19 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 FD: 25 BD: 1 .+.+: &type->s_umount_key#60 ->&rq->__lock FD: 1 BD: 2 ....: (&ndev->cmd_timer) FD: 1 BD: 2 ....: (&ndev->data_timer) FD: 33 BD: 50 +.-.: (&peer->timer_persistent_keepalive) ->pool_lock#2 ->&list->lock#14 ->tk_core.seq.seqcount ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->init_task.mems_allowed_seq.seqcount ->batched_entropy_u8.lock ->kfence_freelist_lock FD: 84 BD: 1 +.-.: (&ndev->rs_timer) ->&ndev->lock ->&zone->lock ->&____s->seqcount ->pool_lock#2 ->&c->lock ->&____s->seqcount#2 ->&dir->lock#2 ->&ul->lock#2 ->&n->list_lock ->batched_entropy_u8.lock ->kfence_freelist_lock ->init_task.mems_allowed_seq.seqcount FD: 25 BD: 5 +.+.: (work_completion)(&rfkill->uevent_work) ->&rq->__lock FD: 1 BD: 4 ++++: &local->sockets.lock FD: 1 BD: 3 +.+.: &local->raw_sockets.lock FD: 1 BD: 1 ....: (&local->link_timer) FD: 25 BD: 1 +.+.: (work_completion)(&local->tx_work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 25 BD: 1 +.+.: (work_completion)(&local->rx_work) ->&rq->__lock FD: 25 BD: 1 +.+.: (work_completion)(&local->timeout_work) ->&rq->__lock FD: 1 BD: 1 ....: (&local->sdreq_timer) FD: 25 BD: 1 +.+.: (work_completion)(&local->sdreq_timeout_work) ->&rq->__lock FD: 26 BD: 5 -.-.: &rq_wait->wait ->&p->pi_lock FD: 77 BD: 1 ++++: &type->s_umount_key#61 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->&root->kernfs_supers_rwsem ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#24 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->inode_hash_lock ->pool_lock#2 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 1 .+.+: &type->s_umount_key#62 FD: 63 BD: 2 ++++: &type->s_umount_key#63 ->sb_lock ->fs_reclaim ->stock_lock ->pool_lock#2 ->&rq->__lock ->&dentry->d_lock ->&lru->node[i].lock ->&obj_hash[i].lock ->&c->lock FD: 69 BD: 1 ++++: &type->s_umount_key#64 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#29 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 ->&rq->__lock ->pool_lock#2 FD: 735 BD: 2 ++++: &type->s_umount_key#65 ->sb_lock ->&dentry->d_lock ->cgroup_mutex ->&rq->__lock FD: 71 BD: 1 ++++: &type->s_umount_key#66 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->pool_lock#2 ->rename_lock.seqcount ->&dentry->d_lock ->&dentry->d_lock/1 ->&sb->s_type->i_lock_key#32 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->binderfs_minors_mutex ->&rq->__lock FD: 1 BD: 2 +...: clock-AF_NFC FD: 30 BD: 1 ..-.: &(&bat_priv->tt.work)->timer FD: 34 BD: 6 +.+.: (work_completion)(&(&bat_priv->tt.work)->work) ->key#16 ->&rq->__lock ->key#21 ->&bat_priv->tt.req_list_lock ->&bat_priv->tt.roam_list_lock ->&obj_hash[i].lock ->&base->lock ->&cfs_rq->removed.lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 7 +...: key#21 FD: 1 BD: 7 +...: &bat_priv->tt.req_list_lock FD: 1 BD: 7 +...: &bat_priv->tt.roam_list_lock FD: 1 BD: 6 +.+.: ebt_mutex.wait_lock FD: 1 BD: 85 ....: key#22 FD: 10 BD: 198 +...: map_idr_lock ->&c->lock ->pool_lock#2 ->&obj_hash[i].lock ->&n->list_lock FD: 39 BD: 2 +.+.: (work_completion)(&map->work) ->stock_lock ->&obj_hash[i].lock ->pool_lock#2 ->vmap_area_lock ->purge_vmap_area_lock ->rcu_node_0 ->&rq->__lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&cfs_rq->removed.lock ->quarantine_lock ->pcpu_lock ->&x->wait#2 ->pool_lock FD: 1 BD: 103 +.+.: smack_known_lock.wait_lock FD: 57 BD: 50 +.-.: (&peer->timer_retransmit_handshake) ->&peer->endpoint_lock FD: 1 BD: 11 ....: key#23 FD: 1 BD: 2 +.+.: nf_sockopt_mutex.wait_lock FD: 1 BD: 1 +.+.: &mq_lock FD: 76 BD: 2 +.+.: free_ipc_work ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->mount_lock ->&fsnotify_mark_srcu ->&dentry->d_lock ->&type->s_umount_key#59 ->sb_lock ->unnamed_dev_ida.xa_lock ->list_lrus_mutex ->&xa->xa_lock#15 ->pool_lock#2 ->mnt_id_ida.xa_lock ->&ids->rwsem ->(work_completion)(&ht->run_work) ->&ht->mutex ->percpu_counters_lock ->pcpu_lock ->sysctl_lock ->&sb->s_type->i_lock_key#23 ->rename_lock.seqcount ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->proc_inum_ida.xa_lock ->quarantine_lock ->pool_lock FD: 1 BD: 3 +.+.: &ids->rwsem FD: 1 BD: 1 +.+.: (work_completion)(&data->suspend_work) FD: 1 BD: 1 +.+.: (work_completion)(&hdev->reenable_adv_work) FD: 1 BD: 14 +.+.: (work_completion)(&(&hdev->interleave_scan)->work) FD: 1 BD: 27 +.+.: (work_completion)(&(&conn->id_addr_timer)->work) FD: 53 BD: 26 +.+.: (work_completion)(&(&conn->disc_work)->work) ->&hdev->unregister_lock FD: 1 BD: 26 +.+.: (work_completion)(&(&conn->auto_accept_work)->work) FD: 1 BD: 26 +.+.: (work_completion)(&(&conn->idle_work)->work) FD: 1 BD: 4 +.+.: oom_adj_mutex.wait_lock FD: 39 BD: 52 +...: _xmit_NETROM ->(console_sem).lock ->console_owner_lock ->console_owner ->&obj_hash[i].lock ->pool_lock#2 FD: 26 BD: 49 +.+.: __ip_vs_mutex ->&ipvs->dest_trash_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 50 +...: &ipvs->dest_trash_lock FD: 25 BD: 49 +.+.: flowtable_lock ->&rq->__lock FD: 77 BD: 2 +.+.: &type->s_umount_key#72/1 ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->&____s->seqcount ->list_lrus_mutex ->&c->lock ->sb_lock ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#35 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock ->&sb->s_type->i_mutex_key#19 ->&n->list_lock ->rcu_node_0 ->&____s->seqcount#2 ->remove_cache_srcu FD: 35 BD: 7 +.+.: nf_conntrack_mutex ->&nf_conntrack_locks[i] ->&rq->__lock ->&____s->seqcount#7 ->&obj_hash[i].lock ->pool_lock#2 ->&nf_conntrack_locks[i]/1 ->&cfs_rq->removed.lock ->nf_conntrack_mutex.wait_lock FD: 25 BD: 53 +.+.: (work_completion)(&(&idev->mc_report_work)->work) ->&rq->__lock FD: 25 BD: 53 +.+.: &net->xdp.lock ->&rq->__lock FD: 1 BD: 53 +.+.: mirred_list_lock FD: 1 BD: 53 +...: &idev->mc_report_lock FD: 26 BD: 53 +.+.: &pnn->pndevs.lock ->&rq->__lock FD: 25 BD: 53 +.+.: &pnn->routes.lock ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 6 ....: netdev_unregistering_wq.lock FD: 910 BD: 1 +.+.: (wq_completion)netns ->net_cleanup_work FD: 909 BD: 2 +.+.: net_cleanup_work ->pernet_ops_rwsem ->rcu_state.barrier_mutex ->&obj_hash[i].lock ->pool_lock#2 ->krc.lock ->&dir->lock ->rcu_state.barrier_mutex.wait_lock ->&p->pi_lock ->quarantine_lock ->&rq->__lock FD: 1 BD: 5 +...: &net->nsid_lock FD: 30 BD: 1 ..-.: &(&bat_priv->dat.work)->timer FD: 30 BD: 1 ..-.: &(&bat_priv->bla.work)->timer FD: 29 BD: 6 +.+.: (work_completion)(&(&bat_priv->dat.work)->work) ->&hash->list_locks[i] ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->&cfs_rq->removed.lock ->pool_lock#2 ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 7 +...: &hash->list_locks[i] FD: 32 BD: 6 +.+.: (work_completion)(&(&bat_priv->bla.work)->work) ->key#20 ->pool_lock#2 ->&obj_hash[i].lock ->&base->lock ->rcu_node_0 ->&rq->__lock ->crngs.lock ->&cfs_rq->removed.lock FD: 1 BD: 5 +...: &tn->node_list_lock FD: 1 BD: 49 +...: &net->xfrm.xfrm_state_lock FD: 1 BD: 49 +...: &net->xfrm.xfrm_policy_lock FD: 1 BD: 1 +.+.: (work_completion)(&data->fib_flush_work) FD: 1 BD: 1 +.+.: ®ion->snapshot_lock FD: 14 BD: 51 +.-.: _xmit_NONE#2 FD: 1 BD: 171 +.+.: pcpu_alloc_mutex.wait_lock FD: 1 BD: 5 +.+.: netns_bpf_mutex FD: 1 BD: 5 ....: (&net->fs_probe_timer) FD: 1 BD: 7 ++++: &net->cells_lock FD: 1 BD: 5 ....: (&net->cells_timer) FD: 31 BD: 1 +.+.: (wq_completion)afs ->(work_completion)(&net->cells_manager) ->(work_completion)(&net->fs_manager) FD: 28 BD: 2 +.+.: (work_completion)(&net->cells_manager) ->&net->cells_lock ->bit_wait_table + i ->&rq->__lock FD: 1 BD: 5 ....: (&net->fs_timer) FD: 28 BD: 2 +.+.: (work_completion)(&net->fs_manager) ->&(&net->fs_lock)->lock ->bit_wait_table + i ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 3 +.+.: &(&net->fs_lock)->lock FD: 1 BD: 6 +.+.: &rx->incoming_lock FD: 1 BD: 6 +.+.: &call->notify_lock FD: 1 BD: 6 ....: (rxrpc_call_limiter).lock FD: 1 BD: 6 +.+.: &rx->recvmsg_lock FD: 1 BD: 6 ....: (&call->timer) FD: 1 BD: 6 ....: &list->lock#17 FD: 1 BD: 5 +.+.: (wq_completion)kafsd FD: 1 BD: 5 +...: k-clock-AF_RXRPC FD: 1 BD: 5 ..-.: rlock-AF_RXRPC FD: 1 BD: 1 ....: (&local->client_conn_reap_timer) FD: 1 BD: 1 ....: &list->lock#18 FD: 1 BD: 8 +.+.: (work_completion)(&data->gc_work) FD: 1 BD: 5 +.+.: (work_completion)(&ovs_net->dp_notify_work) FD: 1 BD: 5 +...: &srv->idr_lock FD: 1 BD: 5 ....: (&rxnet->service_conn_reap_timer) FD: 2 BD: 6 +.+.: (work_completion)(&rxnet->service_conn_reaper) ->&rxnet->conn_lock FD: 1 BD: 7 +...: &nt->cluster_scope_lock FD: 1 BD: 5 +.+.: (work_completion)(&tn->work) FD: 1 BD: 5 +.+.: (work_completion)(&(&c->work)->work) FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#151 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)krdsd FD: 1 BD: 5 +.+.: (work_completion)(&rtn->rds_tcp_accept_w) FD: 1 BD: 54 +...: &icsk->icsk_accept_queue.rskq_lock#2 FD: 1 BD: 5 ....: rds_tcp_conn_lock FD: 1 BD: 5 ....: loop_conns_lock FD: 1 BD: 5 +.+.: (wq_completion)l2tp FD: 1 BD: 51 +...: _xmit_SIT#2 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#54 ->&rq->__lock FD: 1 BD: 51 +...: _xmit_TUNNEL#2 FD: 1 BD: 51 +...: _xmit_IPGRE#2 FD: 1 BD: 5 +.+.: &fn->fou_lock FD: 1 BD: 5 +.+.: ipvs->sync_mutex FD: 1 BD: 53 ....: (&ifibss->timer) FD: 1 BD: 49 +.+.: (work_completion)(&wdev->disconnect_wk) FD: 1 BD: 49 +.+.: (work_completion)(&wdev->pmsr_free_wk) FD: 724 BD: 6 +.+.: (work_completion)(&(&rdev->dfs_update_channels_wk)->work) ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 53 ....: (&local->dynamic_ps_timer) FD: 1 BD: 53 ....: (&dwork->timer)#3 FD: 1 BD: 50 +...: nr_list_lock FD: 1 BD: 49 ....: &rdev->dev_wait FD: 1 BD: 53 ....: (&dwork->timer)#4 FD: 1 BD: 52 +.+.: rcu_state.barrier_mutex.wait_lock FD: 1 BD: 5 +.+.: (work_completion)(&local->restart_work) FD: 1 BD: 5 +.+.: (work_completion)(&rdev->conn_work) FD: 1 BD: 5 +.+.: (work_completion)(&(&rdev->background_cac_done_wk)->work) FD: 1 BD: 5 +.+.: (work_completion)(&rdev->destroy_work) FD: 1 BD: 5 +.+.: (work_completion)(&rdev->propagate_radar_detect_wk) FD: 1 BD: 5 +.+.: (work_completion)(&rdev->propagate_cac_done_wk) FD: 1 BD: 5 +.+.: (work_completion)(&rdev->mgmt_registrations_update_wk) FD: 1 BD: 5 +.+.: (work_completion)(&rdev->background_cac_abort_wk) FD: 1 BD: 5 ....: (&local->sta_cleanup) FD: 3 BD: 5 +.+.: rdma_nets_rwsem ->rdma_nets.xa_lock FD: 1 BD: 5 +...: k-clock-AF_NETLINK FD: 33 BD: 50 +.-.: (&peer->timer_send_keepalive) ->pool_lock#2 ->&list->lock#14 ->tk_core.seq.seqcount ->&c->lock ->&n->list_lock ->&____s->seqcount#2 ->&____s->seqcount ->init_task.mems_allowed_seq.seqcount FD: 1 BD: 49 +...: nr_neigh_list_lock FD: 1 BD: 49 +...: &bpq_netdev_addr_lock_key FD: 1 BD: 49 +.+.: (work_completion)(&(&priv->scan_result)->work) FD: 25 BD: 49 +.+.: (work_completion)(&(&priv->connect)->work) ->&rq->__lock FD: 1 BD: 56 +.+.: rcu_state.exp_mutex.wait_lock FD: 1 BD: 49 +...: &bat_priv->forw_bcast_list_lock FD: 43 BD: 50 +.+.: (work_completion)(&br->mcast_gc_work) ->&br->multicast_lock ->(&p->rexmit_timer) ->&obj_hash[i].lock ->&base->lock ->(&p->timer) ->pool_lock#2 ->krc.lock ->(&mp->timer) FD: 1 BD: 51 ....: (&p->rexmit_timer) FD: 39 BD: 51 +.-.: (&p->timer) ->&br->multicast_lock FD: 39 BD: 51 +.-.: (&mp->timer) ->&br->multicast_lock FD: 1 BD: 49 ....: (&pmctx->ip6_mc_router_timer) FD: 1 BD: 49 ....: (&pmctx->ip4_mc_router_timer) FD: 1 BD: 51 +...: &qdisc_xmit_lock_key FD: 1 BD: 51 +...: &qdisc_xmit_lock_key#2 FD: 1 BD: 51 +...: &vlan_netdev_xmit_lock_key FD: 1 BD: 51 +...: &batadv_netdev_xmit_lock_key FD: 1 BD: 51 +...: &qdisc_xmit_lock_key#3 FD: 1 BD: 51 +...: &qdisc_xmit_lock_key#4 FD: 1 BD: 51 +...: _xmit_LOOPBACK#2 FD: 1 BD: 50 ....: (&peer->timer_new_handshake) FD: 1 BD: 50 ....: (&peer->timer_zero_key_material) FD: 1 BD: 50 +.+.: (work_completion)(&peer->clear_peer_work) FD: 1 BD: 49 +.+.: (work_completion)(&(&bond->mii_work)->work) FD: 1 BD: 49 +.+.: (work_completion)(&(&bond->arp_work)->work) FD: 1 BD: 49 +.+.: (work_completion)(&(&bond->alb_work)->work) FD: 1 BD: 49 +.+.: (work_completion)(&(&bond->ad_work)->work) FD: 1 BD: 49 +.+.: (work_completion)(&(&bond->mcast_work)->work) FD: 1 BD: 49 +.+.: (work_completion)(&(&bond->slave_arr_work)->work) FD: 1 BD: 49 ....: (&br->hello_timer) FD: 1 BD: 49 ....: (&br->topology_change_timer) FD: 1 BD: 49 ....: (&br->tcn_timer) FD: 1 BD: 49 ....: (&brmctx->ip4_mc_router_timer) FD: 1 BD: 49 ....: (&brmctx->ip4_other_query.timer) FD: 1 BD: 49 ....: (&brmctx->ip4_other_query.delay_timer) FD: 1 BD: 49 ....: (&brmctx->ip6_mc_router_timer) FD: 1 BD: 49 ....: (&brmctx->ip6_other_query.timer) FD: 1 BD: 49 ....: (&brmctx->ip6_other_query.delay_timer) FD: 1 BD: 49 +.+.: raw_notifier_lock FD: 1 BD: 49 +.+.: bcm_notifier_lock FD: 1 BD: 49 +.+.: isotp_notifier_lock FD: 1 BD: 49 +.+.: (work_completion)(&port->wq) FD: 1 BD: 49 +...: &bond->ipsec_lock FD: 1 BD: 49 +...: _xmit_NETROM#2 FD: 25 BD: 1 +.+.: (work_completion)(&(&team->mcast_rejoin.dw)->work) ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: (work_completion)(&(&team->notify_peers.dw)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#63 FD: 31 BD: 2 +.+.: (work_completion)(&(&devlink->rwork)->work) ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->pool_lock#2 ->&cfs_rq->removed.lock ->&base->lock FD: 62 BD: 7 +.+.: k-sk_lock-AF_INET6/1 ->k-slock-AF_INET6 ->rlock-AF_INET6 ->&list->lock#19 ->pool_lock#2 ->&dir->lock ->fs_reclaim ->&c->lock ->k-clock-AF_INET6 ->&n->list_lock ->&rq->__lock ->rcu_node_0 ->remove_cache_srcu FD: 1 BD: 9 ....: rlock-AF_INET6 FD: 1 BD: 10 ....: &list->lock#19 FD: 6 BD: 53 +...: k-slock-AF_INET6/1 ->&sctp_ep_hashtable[i].lock ->&obj_hash[i].lock ->pool_lock#2 ->k-clock-AF_INET6 FD: 1 BD: 56 +...: &sctp_ep_hashtable[i].lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#461 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#62 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#189 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#536 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#453 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#518 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#215 ->&rq->__lock FD: 1 BD: 1 +...: link_idr_lock FD: 55 BD: 1 +.-.: (&p->forward_delay_timer) ->&br->lock FD: 1 BD: 5 +.+.: &sn->gssp_lock FD: 1 BD: 8 +.+.: &cd->hash_lock FD: 1 BD: 5 +.+.: xfrm_state_gc_work FD: 1 BD: 5 +...: ip6_fl_lock FD: 1 BD: 5 ....: (&net->ipv6.ip6_fib_timer) FD: 1 BD: 49 ....: (&mrt->ipmr_expire_timer) FD: 1 BD: 5 ....: (&ipvs->dest_trash_timer) FD: 1 BD: 5 +.+.: (work_completion)(&(&ipvs->expire_nodest_conn_work)->work) FD: 1 BD: 5 +.+.: (work_completion)(&(&ipvs->est_reload_work)->work) FD: 1 BD: 5 +...: recent_lock FD: 1 BD: 5 +.+.: hashlimit_mutex FD: 1 BD: 5 +.+.: trans_gc_work FD: 1 BD: 5 +.+.: (work_completion)(&(&cnet->ecache.dwork)->work) FD: 1 BD: 5 +.+.: (work_completion)(&net->xfrm.policy_hash_work) FD: 1 BD: 5 +.+.: (work_completion)(&net->xfrm.state_hash_work) FD: 1 BD: 2 +.+.: misc_mtx.wait_lock FD: 1 BD: 5 +.+.: nfc_devlist_mutex.wait_lock FD: 25 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#22 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#433 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#510 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#220 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#154 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#939 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc22_nci_tx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#448 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#257 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#51 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#466 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#177 FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg1#216 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#171 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#354 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#443 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#512 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#974 ->&rq->__lock FD: 26 BD: 16 ....: &root->deactivate_waitq ->&p->pi_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#204 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#188 ->&rq->__lock FD: 1 BD: 2 +...: slock-AF_ALG FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#418 FD: 1 BD: 2 +.+.: (wq_completion)nfc20_nci_cmd_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#535 FD: 1 BD: 12 +.+.: (work_completion)(&(&pipe->queue_timeout)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#82 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#208 FD: 15 BD: 1 +.-.: (&n->timer) ->&n->lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#650 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#166 FD: 1 BD: 1 +.+.: (wq_completion)nfc40_nci_tx_wq#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#146 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#392 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#531 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#750 ->&rq->__lock FD: 31 BD: 1 +.-.: (&timer) ->&obj_hash[i].lock ->&base->lock ->&txlock ->&txwq FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#494 FD: 1 BD: 1 ....: _rs.lock#2 FD: 1 BD: 2 +.+.: &knet->mutex FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#694 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#447 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1088 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_rx_wq#75 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#910 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#452 FD: 166 BD: 14 +.+.: (wq_completion)hci3#6 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#190 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#507 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1027 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#792 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#423 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#880 ->&rq->__lock FD: 1 BD: 57 ..-.: key#24 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#69 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#206 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#431 ->&rq->__lock FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg0#216 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#391 FD: 1 BD: 2 +.+.: vlan_ioctl_mutex.wait_lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#961 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#288 FD: 1 BD: 2 +...: slock-AF_CAIF FD: 1 BD: 97 +.+.: dev_pm_qos_sysfs_mtx.wait_lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#888 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#893 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1024 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#790 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#252 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#455 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#499 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1215 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#515 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#670 ->&rq->__lock FD: 4 BD: 1 +.+.: put_task_map-wait-type-override#2 ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#484 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#965 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#203 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1016 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#353 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#591 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#181 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1028 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1063 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#12 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#822 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#205 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#142 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#610 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#217 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#253 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#521 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#175 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#199 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#571 ->&rq->__lock FD: 5 BD: 2 +.+.: (ima_keys_delayed_work).work ->ima_keys_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 25 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#22 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#31 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#779 ->&rq->__lock FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg0#215 ->&rq->__lock ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1041 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#192 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#446 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#941 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#598 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#578 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#958 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#431 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1000 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#511 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#434 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#150 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#148 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#49 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#981 ->&rq->__lock FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#106 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#981 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#48 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1015 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#730 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1022 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#449 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#259 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1073 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#881 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#207 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#517 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#507 ->&rq->__lock FD: 1 BD: 2 +.+.: &net->ipv4.ra_mutex FD: 5 BD: 50 +...: _xmit_SLIP#2 ->&eql->queue.lock FD: 5 BD: 1 +...: &nr_netdev_xmit_lock_key ->nr_node_list_lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 2 +...: nr_node_list_lock FD: 5 BD: 1 +...: _xmit_X25#2 ->&lapbeth->up_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#986 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#77 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#513 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#599 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#216 FD: 1 BD: 1 +...: btf_idr_lock FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#105 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 724 BD: 2 +.+.: (crda_timeout).work ->rtnl_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#829 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#954 FD: 30 BD: 1 ..-.: net/wireless/reg.c:533 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#169 FD: 179 BD: 1 +.+.: (wq_completion)hci2 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1131 FD: 1 BD: 2 +...: clock-AF_ISDN FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#771 FD: 25 BD: 5 +.+.: (wq_completion)tipc_crypto#9 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#151 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#942 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#539 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#180 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#828 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1003 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#203 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1217 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#483 FD: 1 BD: 94 +.+.: freezer_mutex.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#861 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#951 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#924 ->&rq->__lock FD: 35 BD: 3 +.+.: &q->blkcg_mutex ->(&sq->pending_timer) ->&obj_hash[i].lock ->&base->lock ->percpu_counters_lock ->pcpu_lock ->pool_lock#2 ->&q->queue_lock ->pool_lock ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#24 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#417 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#568 FD: 166 BD: 14 +.+.: (wq_completion)hci5#2 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#216 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#895 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#516 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#406 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#893 FD: 1 BD: 27 +.+.: hci_cb_list_lock.wait_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#885 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#74 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#739 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#730 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#498 FD: 1 BD: 14 ....: &data->read_wait#2 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#964 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#220 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#475 ->&rq->__lock FD: 43 BD: 1 +.+.: (wq_completion)bond0#116 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#422 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#995 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#215 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#202 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#231 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#202 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#505 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#517 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#352 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#442 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc16_nci_cmd_wq#17 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#411 ->&rq->__lock FD: 25 BD: 3 +.+.: tcpv6_prot_mutex ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#444 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#701 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#785 FD: 161 BD: 1 +.+.: sk_lock-AF_ALG ->slock-AF_ALG ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&obj_hash[i].lock ->&c->lock ->&dir->lock ->remove_cache_srcu ->&n->list_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc31_nci_cmd_wq#12 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#479 FD: 1 BD: 2 +.+.: (work_completion)(&kcm->tx_work) FD: 1 BD: 3 +.+.: chan_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#180 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#947 FD: 1 BD: 51 ..-.: rlock-AF_CAN FD: 1 BD: 1 +.+.: (wq_completion)nfc25_nci_rx_wq#14 FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg1#199 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#171 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#436 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#162 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#163 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#174 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#953 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#979 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#948 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#959 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#954 FD: 30 BD: 1 ..-.: &(&net->ipv6.addr_chk_work)->timer FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#185 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#488 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1093 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1101 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#561 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#444 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#434 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#465 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1084 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1092 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#706 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#329 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#499 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#450 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#806 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1017 ->&rq->__lock FD: 30 BD: 1 +.-.: security/keys/gc.c:28 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1027 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#930 ->&rq->__lock FD: 1 BD: 8 +.+.: nf_conntrack_mutex.wait_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#478 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1012 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1026 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#184 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#474 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1006 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#882 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#472 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#639 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#930 FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#106 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#526 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#519 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1011 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1017 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1015 ->&rq->__lock FD: 1 BD: 3 ....: rlock-AF_INET FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#187 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1016 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1003 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#467 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1004 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1008 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#465 ->&rq->__lock FD: 1 BD: 1 ....: _rs.lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#463 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#872 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#928 FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg2#205 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#873 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#453 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#977 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#435 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#967 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1095 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#971 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#973 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#78 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1126 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#543 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1133 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#794 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#934 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#172 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#427 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#429 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#52 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#150 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#427 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#53 FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg2#192 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#148 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#976 ->&rq->__lock FD: 43 BD: 1 +.+.: (wq_completion)bond0#118 ->(work_completion)(&(&slave->notify_work)->work) FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#520 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#918 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#198 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#185 FD: 1 BD: 1 +.+.: (wq_completion)nfc44_nci_tx_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#438 FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg0#212 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#966 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#973 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#147 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#595 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#636 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#143 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#249 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#159 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#563 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#984 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#748 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#450 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#530 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#985 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#158 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#166 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#439 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#55 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#440 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#458 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#451 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc23_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#948 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#474 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#946 ->&rq->__lock FD: 1 BD: 3 ....: unix_gc_wait.lock FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg2#210 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#473 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#69 FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#108 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1188 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#67 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1197 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#448 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#311 FD: 1 BD: 4 +.+.: tty_mutex.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#493 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#200 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#343 ->&rq->__lock ->&cfs_rq->removed.lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#331 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#496 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#205 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#496 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#344 FD: 25 BD: 1 +.+.: (wq_completion)nfc15_nci_tx_wq#17 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#251 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#499 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#692 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#977 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#79 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#520 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#211 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#82 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#523 FD: 1 BD: 51 ..-.: &list->lock#20 FD: 1 BD: 1 +.-.: x25_forward_list_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1009 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#751 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1020 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1008 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#446 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#504 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#465 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1063 FD: 1 BD: 29 +.+.: &conn->lock#2 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1089 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#176 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#31 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#525 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1018 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#83 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#427 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#775 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#952 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#131 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#391 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#18 FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#27 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#922 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1031 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#596 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#181 FD: 1 BD: 1 +.+.: (wq_completion)nfc27_nci_rx_wq#13 FD: 64 BD: 2 +.+.: root_key_user.cons_lock ->key_user_lock ->root_key_user.lock ->fs_reclaim ->pool_lock#2 ->crngs.lock ->key_serial_lock ->&type->lock_class ->&rq->__lock ->root_key_user.cons_lock.wait_lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#461 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#999 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#218 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#440 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#168 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1023 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1117 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1124 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#58 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1114 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1132 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#962 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#966 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#993 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1009 ->&rq->__lock FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg1#215 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#9 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#464 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#728 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#866 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#821 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 2 +.+.: (wq_completion)nfc27_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#376 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#564 FD: 1 BD: 3102 ....: cid_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#201 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#968 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#196 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#946 FD: 28 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#989 ->rcu_node_0 ->&rcu_state.expedited_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#971 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#708 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#970 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#409 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#985 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#411 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#415 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#412 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#144 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#450 FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_rx_wq#9 FD: 25 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#19 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#417 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#625 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#264 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#152 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#413 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#146 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#414 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#147 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#420 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#416 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#983 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#416 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#422 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#419 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#422 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#422 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#426 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#425 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#425 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#153 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#50 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#424 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#149 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#25 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#428 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#25 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#24 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#430 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#156 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#431 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#429 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#917 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#152 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#179 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#553 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#437 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#167 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#49 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#455 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#746 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#445 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#519 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#509 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#60 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#469 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#217 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#877 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#508 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#341 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#506 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#211 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#528 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#461 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#500 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#514 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#478 FD: 4 BD: 1 +.+.: sk_lock-AF_SMC ->slock-AF_SMC ->smc_v4_hashinfo.lock ->clock-AF_SMC FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1190 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#484 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#827 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_tx_wq#7 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#891 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#519 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#888 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1034 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#997 FD: 1 BD: 6 +.+.: nf_nat_proto_mutex.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#440 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#864 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#901 FD: 1 BD: 1 +.+.: s_shared_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_rx_wq#13 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#432 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#419 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#918 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#54 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#24 ->&rq->__lock FD: 25 BD: 1 +.+.: (work_completion)(&tty->SAK_work) ->&rq->__lock FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg0#205 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#943 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#810 FD: 25 BD: 2 +.+.: (wq_completion)nfc18_nci_cmd_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#442 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#420 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#796 ->&rq->__lock FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg2#102 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#546 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#654 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#171 ->&rq->__lock FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg2#209 ->&rq->__lock ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#448 FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg1#210 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#47 ->&rq->__lock FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg2#191 ->&rq->__lock ->(work_completion)(&peer->transmit_handshake_work) FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#236 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#798 FD: 26 BD: 1 ....: &x->wait#27 ->&p->pi_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#905 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#410 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#512 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#421 ->&rq->__lock FD: 1 BD: 1 ....: (&lo->timer) FD: 1 BD: 4 ....: (&sq->pending_timer) FD: 38 BD: 2 +.+.: (work_completion)(&blkg->free_work) ->&q->blkcg_mutex ->&obj_hash[i].lock ->pool_lock#2 ->&xa->xa_lock#10 ->pcpu_lock ->blk_queue_ida.xa_lock ->percpu_ref_switch_lock ->&base->lock ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#874 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#33 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#462 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#974 FD: 1 BD: 2 +...: slock-AF_SMC FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#404 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#26 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#155 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#23 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#202 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#162 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#163 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#162 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#159 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#163 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#62 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#450 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#452 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#455 FD: 1 BD: 1 ....: &port->buf.lock/1 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#458 FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg1#191 ->(work_completion)(&peer->transmit_handshake_work) FD: 25 BD: 2 +.+.: (wq_completion)nfc35_nci_cmd_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#153 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#727 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#416 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#432 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#579 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#179 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#68 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#957 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#189 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#475 FD: 1 BD: 2 +...: smc_v4_hashinfo.lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#996 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#475 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#187 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#194 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#188 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#189 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#191 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#193 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#481 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#73 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#482 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#71 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#776 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#485 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#81 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1007 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1005 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#486 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1010 FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg2#215 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1007 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#467 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1013 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#458 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1011 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#609 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1040 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1162 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1167 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#622 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#606 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#587 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#542 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1014 ->&rq->__lock FD: 1 BD: 1 +...: &xs->map_list_lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#71 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#488 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#493 FD: 25 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#17 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#588 ->&rq->__lock FD: 1 BD: 65 ....: &new->fa_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#490 FD: 1 BD: 95 +.+.: wq_pool_mutex.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#72 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#75 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#490 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#78 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#197 FD: 30 BD: 1 +.-.: (&pool->idle_timer) ->&pool->lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1198 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#76 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#493 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#707 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#589 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#317 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1206 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#691 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_rx_wq#16 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#123 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#492 FD: 1 BD: 2 +.+.: nfsd_mutex FD: 64 BD: 3 +.+.: &sb->s_type->i_mutex_key#19 ->fs_reclaim ->stock_lock ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#35 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&n->list_lock ->&rq->__lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#80 FD: 73 BD: 1 +.+.: &type->s_umount_key#73 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->nfsd_mutex ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#35 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->pool_lock#2 ->&fsnotify_mark_srcu ->&rq->__lock ->&dentry->d_lock/1 ->&cfs_rq->removed.lock FD: 25 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#20 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#98 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#735 ->&rq->__lock FD: 166 BD: 14 +.+.: (wq_completion)hci4#4 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#502 FD: 179 BD: 1 +.+.: (wq_completion)hci4#3 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#252 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1201 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#504 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1213 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#513 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#501 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#516 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#212 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#512 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#520 FD: 26 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#214 ->rcu_node_0 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#218 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#81 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#524 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#756 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#755 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#524 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#222 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#218 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#257 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#526 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#527 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#86 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#529 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#218 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#529 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1224 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#609 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#260 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#136 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#393 FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg2#186 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#739 ->&rq->__lock FD: 6 BD: 2 +.+.: sk_lock-AF_NETROM ->slock-AF_NETROM ->&obj_hash[i].lock ->nr_list_lock ->rlock-AF_NETROM FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#441 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#773 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#498 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#503 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#509 ->&rq->__lock FD: 1 BD: 3 ....: rlock-AF_NETROM FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#441 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#184 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#186 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#81 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#505 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#518 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#187 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#476 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#213 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#875 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#925 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc35_nci_rx_wq#7 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#922 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#814 FD: 729 BD: 13 +.+.: &devlink->lock_key#101 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&c->lock ->&xa->xa_lock#19 ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->&rq->__lock ->rtnl_mutex ->rcu_node_0 ->&(&fn_net->fib_chain)->lock ->stack_depot_init_mutex ->&n->list_lock ->rtnl_mutex.wait_lock ->&p->pi_lock ->&devlink_port->type_lock ->&nsim_trap_data->trap_lock ->&rcu_state.expedited_wq ->&cfs_rq->removed.lock FD: 1 BD: 2 +.+.: (wq_completion)nfc16_nci_cmd_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#498 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#857 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#413 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#141 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#445 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#438 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#179 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#491 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#432 ->&rq->__lock FD: 1 BD: 2 +...: clock-AF_ROSE FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#28 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#155 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#428 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#419 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#414 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#572 FD: 30 BD: 1 ..-.: &(&conn->disc_work)->timer FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#605 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#351 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#751 FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg1#96 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#638 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#297 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#455 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#710 ->&rq->__lock FD: 166 BD: 14 +.+.: (wq_completion)hci0#6 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#128 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#720 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#832 FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg1#102 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#833 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#392 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#847 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#849 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#388 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#397 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#881 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#414 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#720 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#140 ->&rq->__lock FD: 1 BD: 3 +...: clock-AF_KCM FD: 1 BD: 1 +.+.: (work_completion)(&smc->connect_work) FD: 1 BD: 3 +...: slock-AF_KCM FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq#15 FD: 1 BD: 1 +...: rlock-AF_CAIF FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#894 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#872 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_tx_wq#885 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#867 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#886 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#793 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#443 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#945 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#935 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#19 FD: 1 BD: 1 ....: _rs.lock#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#212 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#897 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#905 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#961 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#962 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#960 FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg0#103 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#964 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#968 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#969 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1055 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#197 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#468 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#62 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#477 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1092 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1079 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1077 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#521 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#441 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#183 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1012 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1002 FD: 1 BD: 159 ...-: init_task.mems_allowed_seq.seqcount FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#514 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#400 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#496 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#216 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#842 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1195 ->&rq->__lock FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg2#214 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#823 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#178 FD: 70 BD: 1 .+.+: sb_writers#13 ->mount_lock ->&sb->s_type->i_mutex_key#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#442 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#100 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#957 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#961 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#741 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#416 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#451 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#170 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#589 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#174 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#722 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#915 ->&rq->__lock FD: 1 BD: 53 +.+.: (wq_completion)phy222 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#167 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#27 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#909 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#937 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#915 FD: 1 BD: 3 ....: key#26 FD: 1 BD: 4 ....: &port->open_wait FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#215 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#807 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#462 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#458 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#454 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#451 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#949 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#942 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#936 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#417 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#944 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#925 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#911 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#916 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#174 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#813 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#205 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#423 FD: 1 BD: 102 +.+.: &pa->pa_lock#2 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#333 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#908 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#950 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#430 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#593 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#524 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#497 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#418 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#724 ->&rq->__lock FD: 7 BD: 53 +...: slock-AF_INET6/1 ->&sctp_ep_hashtable[i].lock ->&obj_hash[i].lock ->pool_lock#2 ->clock-AF_INET6 ->&____s->seqcount FD: 43 BD: 1 +.+.: (wq_completion)bond0#101 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#218 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#917 ->&rq->__lock ->&cfs_rq->removed.lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#818 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#886 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#91 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#972 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#987 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#777 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#114 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#45 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#667 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#155 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#576 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#534 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#61 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#223 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#530 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#456 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#168 FD: 1 BD: 97 +.+.: wq_pool_attach_mutex.wait_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#695 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#413 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#219 FD: 25 BD: 5 +.+.: (wq_completion)tipc_rcv#6 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#681 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#439 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_tx_wq#925 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#407 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#702 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#666 FD: 38 BD: 4 +.+.: &sb->s_type->i_lock_key#34 ->&dentry->d_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#330 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#940 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#876 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#812 ->&rq->__lock FD: 1 BD: 3 +...: clock-AF_PPPOX FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#495 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#910 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#99 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#941 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#428 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#419 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#909 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#900 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#429 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#171 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#635 ->&rq->__lock FD: 1 BD: 3 +...: rose_list_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc33_nci_cmd_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#183 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#531 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1098 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#500 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#744 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#745 FD: 1 BD: 1 ....: (&bdi->laptop_mode_wb_timer) FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#57 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#498 FD: 26 BD: 1 +.+.: &bdi->cgwb_release_mutex ->cgwb_lock ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1021 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1015 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#466 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#619 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#718 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#969 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1020 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#457 FD: 32 BD: 2 +.+.: (work_completion)(&aux->work)#2 ->&aux->poke_mutex ->map_idr_lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#471 FD: 78 BD: 2 +.+.: sk_lock-AF_INET/1 ->slock-AF_INET#2 ->rlock-AF_INET ->&list->lock#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#67 FD: 1 BD: 27 +.+.: (work_completion)(&(&conn->timeout_work)->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#551 FD: 1 BD: 2 +.+.: (wq_completion)nfc45_nci_cmd_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#470 FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#108 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#29 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#473 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#467 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1005 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#978 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1049 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#523 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#210 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#510 FD: 1 BD: 3 +...: slock-AF_NETROM FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#695 FD: 1 BD: 1 ....: _rs.lock#4 FD: 25 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#14 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#525 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#718 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_rx_wq#86 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#153 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#155 FD: 25 BD: 1 +.+.: (wq_completion)nfc33_nci_rx_wq#12 ->&rq->__lock FD: 1 BD: 50 ..-.: &list->lock#21 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#54 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#156 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#325 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#725 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#406 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#791 FD: 25 BD: 2 +.+.: (wq_completion)nfc34_nci_cmd_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#103 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#541 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#54 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#157 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#56 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#59 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#158 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#437 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#57 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#164 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#158 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#443 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#446 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#166 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#160 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#442 ->&rq->__lock FD: 1 BD: 113 ....: key#27 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#57 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#59 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#165 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#164 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#168 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#170 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#169 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#59 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#448 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#450 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#457 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#452 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#454 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#171 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#172 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#459 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#60 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#172 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#63 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#173 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#174 FD: 165 BD: 6 +.+.: &tty->ldisc_sem/1 ->&tty->termios_rwsem ->tty_ldiscs_lock ->&obj_hash[i].lock ->pool_lock#2 ->&____s->seqcount ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#457 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#61 FD: 262 BD: 4 +.+.: &tty->legacy_mutex/1 ->&tty->files_lock ->tasklist_lock ->&tty->write_wait ->&tty->read_wait ->&tty->ldisc_sem ->&tty->ctrl.lock ->&obj_hash[i].lock ->&rq->__lock ->&f->f_lock FD: 31 BD: 6 +.+.: &o_tty->termios_rwsem/1 ->vmap_area_lock ->&obj_hash[i].lock ->purge_vmap_area_lock ->pool_lock#2 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#62 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#455 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#63 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#175 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#174 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#463 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#603 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#176 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#457 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#955 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#459 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#435 FD: 166 BD: 1 +.+.: (wq_completion)hci0#8 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#66 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#177 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#173 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq#19 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#462 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#64 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#65 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#936 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#465 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#178 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#468 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#705 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#463 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#435 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#934 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#414 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#238 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#726 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#559 ->&rq->__lock FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg0#202 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#638 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#243 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#148 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#430 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#327 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#810 FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_tx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#651 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#904 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#920 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#672 ->&rq->__lock FD: 1 BD: 3 +...: ax25_list_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#460 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#729 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#904 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#178 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#21 FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_tx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#180 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#66 FD: 89 BD: 2 +.+.: &type->s_umount_key#67/1 ->fs_reclaim ->&c->lock ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->sb_mutex ->&dentry->d_lock ->&n->list_lock ->bit_wait_table + i ->&rq->__lock ->&cfs_rq->removed.lock FD: 76 BD: 4 +.+.: sb_mutex ->udc_lock ->fs_reclaim ->pool_lock#2 ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#33 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&____s->seqcount ->&rq->__lock ->&dentry->d_lock ->rename_lock.seqcount ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 ->sb_lock ->unnamed_dev_ida.xa_lock ->&c->lock ->&n->list_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#252 ->&rq->__lock FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg1#209 ->(work_completion)(&peer->transmit_handshake_work) ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#185 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#470 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#186 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#472 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#474 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#185 ->&rq->__lock FD: 1 BD: 3 +.+.: sco_sk_list.lock FD: 10 BD: 197 +...: prog_idr_lock ->pool_lock#2 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 729 BD: 13 +.+.: &devlink->lock_key#116 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&c->lock ->&xa->xa_lock#19 ->&n->list_lock ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->&rq->__lock ->&(&fn_net->fib_chain)->lock ->stack_depot_init_mutex ->&____s->seqcount#2 ->&____s->seqcount ->&devlink_port->type_lock ->&nsim_trap_data->trap_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#60 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#59 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#191 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#70 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#72 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#269 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#62 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#966 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#978 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#476 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#993 ->&rq->__lock FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg2#206 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 50 ....: (inetaddr_chain).rwsem.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#822 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#455 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#985 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#474 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#915 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#922 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#929 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#924 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#481 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#450 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#478 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#476 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#477 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#449 FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg0#193 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#189 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#479 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#481 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#191 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#190 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#480 FD: 1 BD: 1 +.+.: &resv_map->lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#197 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#483 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#195 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#70 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#199 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1001 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#451 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#356 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#615 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#173 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#552 ->&rq->__lock FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg1#214 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#433 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#997 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#181 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#453 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#491 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#488 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#461 FD: 1 BD: 53 +.+.: (wq_completion)phy221 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#457 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#702 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#102 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#364 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#591 FD: 69 BD: 1 +.+.: &type->s_umount_key#75 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->pool_lock#2 ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#36 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1006 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1014 ->&rq->__lock FD: 136 BD: 50 +.+.: team->team_lock_key#101 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&c->lock ->&n->list_lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&rq->__lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->lweventlist_lock ->(console_sem).lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1050 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1063 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#490 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#195 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#567 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1165 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#606 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1164 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1227 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#482 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#849 FD: 1 BD: 1 +...: clock-AF_XDP FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#108 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#487 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#489 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#197 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#77 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#73 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#30 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#247 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#496 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#199 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1196 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#597 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1201 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1189 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc21_nci_cmd_wq#16 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#21 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#23 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc25_nci_rx_wq#15 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#335 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#728 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#207 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#23 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#501 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1208 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#504 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#502 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#204 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#509 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#506 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#510 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#511 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#206 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1191 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#207 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#208 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#210 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#209 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#211 FD: 30 BD: 1 ..-.: security/integrity/ima/ima_queue_keys.c:35 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#594 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1001 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#751 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#212 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#6 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#521 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#83 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#131 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#746 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#527 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#532 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#85 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#528 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#224 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#532 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1221 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1223 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#396 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#151 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#65 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#940 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#499 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#783 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#784 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#480 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#76 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#515 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#67 FD: 1 BD: 2 +.+.: (wq_completion)nfc17_nci_cmd_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#345 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#525 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#803 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#494 FD: 25 BD: 1 +.+.: (wq_completion)nfc34_nci_rx_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#227 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#596 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#206 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#711 FD: 1 BD: 3 +...: slock-AF_QIPCRTR FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#927 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#213 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#77 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1067 FD: 179 BD: 1 +.+.: (wq_completion)hci0#5 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#732 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#731 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#437 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#522 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#907 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#534 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#529 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#811 FD: 1 BD: 1 +.+.: (wq_completion)nfc36_nci_tx_wq#6 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#447 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#733 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#797 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#56 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#796 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#537 ->&rq->__lock FD: 1 BD: 3 +...: slock-AF_ROSE FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#547 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#539 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#228 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#549 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#429 ->&rq->__lock FD: 1 BD: 13 ....: &tpk_port.spinlock FD: 1 BD: 2 +.+.: (wq_completion)nfc20_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#424 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#559 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#557 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#554 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#423 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#418 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#150 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#88 FD: 1 BD: 50 +.-.: &jsk->sk_session_queue_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#147 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#792 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#799 FD: 1 BD: 3 ....: rlock-AF_ROSE FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg0#213 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#146 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#412 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#408 FD: 1 BD: 3 +...: device_spinlock FD: 25 BD: 2 +.+.: (wq_completion)nfc16_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#743 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#757 ->&rq->__lock FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg2#96 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#943 FD: 25 BD: 1 +.+.: (wq_completion)nfc18_nci_rx_wq#16 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc17_nci_tx_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#139 FD: 1 BD: 2 +.+.: (wq_completion)nfc21_nci_cmd_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_rx_wq#16 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#408 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#494 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#201 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#609 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#611 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#606 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#629 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq#16 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#630 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#280 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#631 FD: 1 BD: 1 +.+.: (wq_completion)nfc22_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#645 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#652 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#645 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#106 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#180 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#658 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#884 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#668 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#294 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#304 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#301 FD: 1 BD: 50 +.-.: &priv->active_session_list_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#675 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#111 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#676 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#679 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#693 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#686 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#725 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#921 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#928 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#729 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#339 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#953 ->&rq->__lock FD: 136 BD: 50 +.+.: team->team_lock_key#115 ->fs_reclaim ->&c->lock ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->&n->list_lock ->lweventlist_lock ->(console_sem).lock ->&rq->__lock ->&____s->seqcount#2 ->&____s->seqcount FD: 25 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq#19 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#747 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#733 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#522 FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg2#199 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#444 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#801 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#814 FD: 179 BD: 1 +.+.: (wq_completion)hci4#5 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 25 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#13 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#817 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_tx_wq#826 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#835 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#837 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#388 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#839 ->&rq->__lock FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#107 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc28_nci_cmd_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#855 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#844 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc35_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#17 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#853 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#862 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#856 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#868 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#865 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#859 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#401 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#858 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#866 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#156 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#868 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#878 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#876 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#892 ->&rq->__lock FD: 2 BD: 2 +...: &mux->rx_lock ->rlock-AF_KCM FD: 27 BD: 2 +.+.: sk_lock-AF_KCM ->slock-AF_KCM ->&rq->__lock ->clock-AF_KCM FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#401 FD: 160 BD: 2 .+.+: sb_writers#14 ->&rq->__lock ->fs_reclaim ->pool_lock#2 ->&mm->mmap_lock ->&obj_hash[i].lock ->&c->lock ->remove_cache_srcu ->mmu_notifier_invalidate_range_start ->smack_known_lock ->smack_known_lock.wait_lock ->&p->pi_lock ->smk_net6addr_lock ->&n->list_lock ->quarantine_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#409 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#408 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#406 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#399 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#442 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#480 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#409 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#417 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#891 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#897 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#412 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#477 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#483 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#916 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#942 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#444 FD: 50 BD: 1 +.+.: sk_lock-AF_CAIF ->slock-AF_CAIF ->&obj_hash[i].lock ->&this->info_list_lock ->(console_sem).lock ->&x->wait#2 ->&rq->__lock ->&ei->socket.wq.wait ->clock-AF_CAIF ->stock_lock ->key ->pcpu_lock ->percpu_counters_lock ->pool_lock#2 ->&cfs_rq->removed.lock FD: 25 BD: 2 +.+.: (wq_completion)nfc35_nci_cmd_wq#7 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#54 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#140 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#936 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#906 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#423 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#412 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#906 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#864 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq#11 FD: 135 BD: 50 +.+.: team->team_lock_key#116 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->input_pool.lock ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->quarantine_lock ->nl_table_wait.lock ->remove_cache_srcu ->&rq->__lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->&n->list_lock ->lweventlist_lock ->(console_sem).lock FD: 2 BD: 3 +.+.: &local->sdp_lock ->&local->sockets.lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#963 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#949 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#954 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#864 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#390 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#959 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#967 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#964 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#956 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#960 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#969 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#545 FD: 50 BD: 7 +.+.: &type->lock_class/1 ->fs_reclaim ->&c->lock ->pool_lock#2 ->&obj_hash[i].lock ->&rq->__lock ->&n->list_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc29_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#956 FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#960 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#439 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#177 ->&rq->__lock FD: 2 BD: 92 +.+.: (work_completion)(flush) ->&list->lock#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#436 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#789 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#607 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#979 FD: 136 BD: 50 +.+.: team->team_lock_key#117 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&dir->lock#2 ->&c->lock ->input_pool.lock ->&n->list_lock ->&rq->__lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->lweventlist_lock ->(console_sem).lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#991 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#486 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#460 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#464 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#987 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#999 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#992 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#996 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1004 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#203 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1049 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#497 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1060 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#508 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1074 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#894 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1100 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1091 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#515 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#510 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1088 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1076 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1097 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#515 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#523 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#410 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1099 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#529 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#523 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1103 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_tx_wq#1085 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1110 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1109 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1102 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#519 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1084 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1094 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1080 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#70 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#516 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1070 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#209 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#989 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1078 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#496 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#509 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#200 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#71 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#187 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#192 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#193 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1045 ->&rq->__lock FD: 1 BD: 3 +.+.: root_key_user.cons_lock.wait_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#237 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1057 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1061 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1039 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#190 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1043 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1029 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#479 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1023 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1032 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1024 FD: 25 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#20 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#923 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1010 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1021 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1013 FD: 77 BD: 2 +.+.: &type->s_umount_key#69/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->stock_lock ->&c->lock ->&____s->seqcount ->(console_sem).lock ->&obj_hash[i].lock ->&rq->__lock ->&x->wait#23 ->&n->list_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#466 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1008 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#183 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#63 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#877 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#473 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#184 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#186 FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_rx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#624 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#612 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#183 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#875 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1006 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#889 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#994 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#887 ->&rq->__lock FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#110 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#990 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#176 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#983 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#994 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#30 ->&rq->__lock FD: 26 BD: 2 +.+.: &type->lock_class#2 ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->root_key_user.lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#76 ->&rq->__lock FD: 43 BD: 1 +.+.: (wq_completion)bond0#115 ->(work_completion)(&(&slave->notify_work)->work) FD: 7 BD: 2 +.+.: sk_lock-AF_AX25 ->slock-AF_AX25 ->clock-AF_AX25 ->ax25_list_lock ->&obj_hash[i].lock ->rlock-AF_AX25 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#879 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#884 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#891 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#828 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#879 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#883 ->&rq->__lock FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg1#103 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#995 FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg2#103 ->&rq->__lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg0#208 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg1#205 ->(work_completion)(&peer->transmit_handshake_work) FD: 38 BD: 1 +.+.: sk_lock-AF_INET6/1 ->&rq->__lock ->slock-AF_INET6 ->rlock-AF_INET6 ->&list->lock#19 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#987 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#975 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#440 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#983 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#439 FD: 42 BD: 1 .+.+: sb_writers#12 ->mount_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#980 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#965 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#439 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#57 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#540 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#531 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#525 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#549 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#228 ->&rq->__lock FD: 29 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#17 ->&rq->__lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1136 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#465 FD: 25 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#17 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#19 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#240 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#433 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#935 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#539 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#558 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#552 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#560 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#59 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1038 ->&rq->__lock FD: 179 BD: 1 +.+.: (wq_completion)hci0#7 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#887 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#408 FD: 1 BD: 2 +...: &mux->lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#881 ->&rq->__lock FD: 1 BD: 3 ....: rlock-AF_KCM FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#165 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#435 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#445 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#407 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#896 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1016 FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg1#186 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#785 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1232 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#427 FD: 1 BD: 12 +.+.: &pipe->lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#145 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#432 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#165 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#183 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#184 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#182 FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#105 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#968 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#70 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#937 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#12 FD: 6 BD: 1 +.+.: put_task_map-wait-type-override#3 ->&obj_hash[i].lock ->pool_lock#2 ->stock_lock FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#35 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1214 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#590 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#30 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1174 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#260 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#267 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#323 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#201 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#200 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#327 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#503 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#546 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#430 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#753 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#214 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#84 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#530 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#527 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#219 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#134 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#497 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#479 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#172 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#736 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#517 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1026 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#65 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#140 ->&rq->__lock FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg2#185 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#637 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#561 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#560 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#424 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#453 FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg0#99 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#173 FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg0#207 ->(work_completion)(&peer->transmit_handshake_work) FD: 729 BD: 13 +.+.: &devlink->lock_key#117 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&c->lock ->&xa->xa_lock#19 ->pcpu_alloc_mutex ->&n->list_lock ->&obj_hash[i].lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->&(&fn_net->fib_chain)->lock ->stack_depot_init_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&devlink_port->type_lock ->&nsim_trap_data->trap_lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#975 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#411 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#52 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#50 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1030 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#681 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#415 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#145 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#417 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#420 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#421 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#780 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#420 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#421 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#423 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#149 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#426 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#24 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#23 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#426 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#51 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#428 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#831 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#97 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_tx_wq#74 FD: 1 BD: 2 +...: clock-AF_NETROM FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#195 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#219 ->&rq->__lock FD: 3 BD: 2 +.+.: unix_gc_lock ->rlock-AF_UNIX ->unix_gc_wait.lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#505 FD: 64 BD: 1 +.+.: &smc->clcsock_release_lock ->k-sk_lock-AF_INET ->k-slock-AF_INET#2 ->pool_lock#2 ->&dir->lock ->&obj_hash[i].lock ->stock_lock ->&sb->s_type->i_lock_key#8 ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->quarantine_lock FD: 1 BD: 2 +...: clock-AF_SMC FD: 1 BD: 4 +.+.: bpf_preload_lock.wait_lock FD: 1 BD: 15 ....: triggers_list_lock.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#453 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#178 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#52 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#363 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#398 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#991 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#990 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#814 FD: 179 BD: 1 +.+.: (wq_completion)hci3#7 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#181 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#178 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#384 FD: 1 BD: 98 +.+.: rcu_state.exp_wake_mutex.wait_lock FD: 1 BD: 53 +.+.: (wq_completion)phy229 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#893 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#55 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#913 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#421 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#919 FD: 25 BD: 1 +.+.: (work_completion)(&tty->hangup_work) ->&rq->__lock FD: 33 BD: 2 +.+.: (work_completion)(&tty->hangup_work)#2 ->&tty->files_lock ->stock_lock ->&obj_hash[i].lock ->pool_lock#2 ->&rq->__lock ->(work_completion)(&buf->work) ->quarantine_lock ->&meta->lock ->kfence_freelist_lock ->per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) ->&base->lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#465 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#429 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#920 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#169 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#531 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#491 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#229 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#160 FD: 1 BD: 3 +...: slock-AF_AX25 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#743 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#375 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#354 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#432 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#420 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#927 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#914 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#795 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#803 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#902 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#694 FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg0#106 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg1#213 ->(work_completion)(&peer->transmit_handshake_work) FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#494 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)kintegrityd FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1011 FD: 1 BD: 2 +.+.: (wq_completion)nfc18_nci_cmd_wq#12 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1012 ->&rq->__lock FD: 1 BD: 3 +.+.: &aux->poke_mutex FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#566 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#154 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#440 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#436 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#83 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#186 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1177 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#572 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1019 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#479 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#484 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#467 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#533 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#33 FD: 25 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#447 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#358 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#373 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#25 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#154 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#433 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#434 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#477 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#189 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#161 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#156 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#437 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#435 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#58 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#56 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#438 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#157 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#29 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#443 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#160 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#161 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#162 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#58 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#164 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#163 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#963 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#58 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#167 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#173 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#456 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#62 FD: 1 BD: 50 ..-.: &jsk->waitq FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#460 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#944 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#606 FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg0#185 ->(work_completion)(&peer->transmit_handshake_work) FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#950 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#459 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#64 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#181 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#460 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#643 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#648 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#106 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#291 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#378 ->&rq->__lock FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg0#201 ->(work_completion)(&peer->transmit_handshake_work) FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#157 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#41 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#727 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#715 FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg2#100 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#757 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#816 ->&rq->__lock FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg2#200 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc29_nci_rx_wq#14 FD: 179 BD: 1 +.+.: (wq_completion)hci1#7 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#952 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#958 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#951 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc16_nci_rx_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#179 FD: 79 BD: 2 +.+.: &type->s_umount_key#68 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->pool_lock#2 ->sb_mutex ->&rq->__lock ->sb_lock ->&dentry->d_lock FD: 38 BD: 5 +.+.: &sb->s_type->i_lock_key#33 ->&dentry->d_lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#466 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#458 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#871 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#447 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#182 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#816 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#824 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#383 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#882 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#962 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#471 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#617 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#469 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#437 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#475 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#184 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#472 FD: 1 BD: 4 +...: &msk->pm.lock FD: 1 BD: 2 +.+.: (work_completion)(&msk->work) FD: 38 BD: 28 +.+.: sk_lock-AF_BLUETOOTH-BTPROTO_SCO ->slock-AF_BLUETOOTH-BTPROTO_SCO ->&rq->__lock ->&conn->lock#2 ->&obj_hash[i].lock ->&base->lock ->&ei->socket.wq.wait ->rcu_node_0 ->&cfs_rq->removed.lock ->pool_lock#2 FD: 1 BD: 29 +...: slock-AF_BLUETOOTH-BTPROTO_SCO FD: 11 BD: 50 ++.-: &priv->j1939_socks_lock ->&jsk->filters_lock ->pool_lock#2 ->rlock-AF_CAN ->&c->lock ->&____s->seqcount#2 ->&____s->seqcount FD: 1 BD: 9 +.+.: key_construction_mutex.wait_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#68 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#473 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#28 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#973 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#562 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#485 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#192 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#486 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1002 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#994 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#581 FD: 28 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#180 ->rcu_node_0 ->&rcu_state.expedited_wq ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#192 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#198 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#193 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#194 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#489 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#486 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#484 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#196 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#195 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#238 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#26 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#913 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#417 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#926 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#604 ->&rq->__lock FD: 1 BD: 3 +...: clock-AF_AX25 FD: 729 BD: 13 +.+.: &devlink->lock_key#118 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&xa->xa_lock#19 ->&c->lock ->pcpu_alloc_mutex ->&n->list_lock ->&obj_hash[i].lock ->&base->lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->&rq->__lock ->&(&fn_net->fib_chain)->lock ->&____s->seqcount#2 ->&____s->seqcount ->stack_depot_init_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&devlink_port->type_lock ->&nsim_trap_data->trap_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_rx_wq#456 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#459 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#586 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1182 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1172 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1152 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1185 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1187 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1200 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1176 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#516 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#182 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#598 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1218 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#612 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#703 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#72 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#17 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#596 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1205 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#196 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1204 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#319 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#703 FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg0#210 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#21 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#678 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#204 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#78 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#497 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#500 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#504 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#502 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#503 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#506 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#507 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#508 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#508 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#82 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1215 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#80 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#79 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#208 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#513 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#209 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#519 FD: 43 BD: 5 +.+.: (wq_completion)bond0#114 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 1 ....: &f->f_owner.lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#518 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#522 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#216 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#6 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#523 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#128 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#747 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#80 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#32 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#528 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#84 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#535 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1210 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#7 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#434 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#369 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#786 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#500 FD: 25 BD: 1 +.+.: (wq_completion)nfc16_nci_tx_wq#17 ->&rq->__lock FD: 52 BD: 1 .+.+: kn->active#53 ->fs_reclaim ->&kernfs_locks->open_file_mutex[count] ->&c->lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#47 FD: 43 BD: 5 +.+.: (wq_completion)bond0#112 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#623 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#931 FD: 1 BD: 2 +.+.: (wq_completion)nfc40_nci_cmd_wq#6 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#707 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#734 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#933 FD: 25 BD: 2 +.+.: (wq_completion)nfc17_nci_cmd_wq#14 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#809 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#930 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#696 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#61 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#952 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#427 FD: 1 BD: 2 +.+.: br_ioctl_mutex.wait_lock FD: 1 BD: 4 +...: &pernet->lock FD: 31 BD: 2 +.+.: sk_lock-AF_PPPOX ->slock-AF_PPPOX ->&pn->hash_lock ->clock-AF_PPPOX ->chan_lock ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->pool_lock FD: 1 BD: 3 +...: slock-AF_PPPOX FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#446 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#276 FD: 1 BD: 1 +...: bpf_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#439 FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg1#200 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#368 FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg1#100 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg1#192 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#90 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#543 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#425 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#556 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc20_nci_rx_wq#13 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#419 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#413 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#239 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#143 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#787 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 6 BD: 2 +.+.: sk_lock-AF_ROSE ->slock-AF_ROSE ->rose_list_lock ->&obj_hash[i].lock ->rlock-AF_ROSE FD: 66 BD: 2 +.+.: key_gc_work ->key_serial_lock ->&obj_hash[i].lock ->&x->wait#2 ->&rq->__lock ->keyring_name_lock ->pool_lock#2 ->root_key_user.lock ->pool_lock ->&base->lock ->&cfs_rq->removed.lock ->rcu_node_0 ->quarantine_lock ->&type->lock_class ->&meta->lock ->kfence_freelist_lock ->krc.lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#50 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#90 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#564 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#821 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_rx_wq#47 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#332 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#639 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#644 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#651 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#300 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#664 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#302 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#671 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#296 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#721 ->&rq->__lock FD: 166 BD: 1 +.+.: (wq_completion)hci1#8 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#752 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#348 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#786 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#177 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#812 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#820 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#501 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#851 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq#15 FD: 25 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#12 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc20_nci_tx_wq#16 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#853 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#857 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#386 FD: 25 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#15 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#870 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#416 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#900 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#899 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#902 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#892 ->&rq->__lock FD: 27 BD: 2 +.+.: sk_lock-AF_QIPCRTR ->&rq->__lock ->slock-AF_QIPCRTR ->clock-AF_QIPCRTR FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#899 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#658 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_rx_wq#160 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#420 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#428 ->&rq->__lock FD: 166 BD: 1 +.+.: (wq_completion)hci2#2 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#193 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#214 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#48 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#482 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#478 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#521 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#933 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#910 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_tx_wq#902 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#909 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#903 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#167 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#165 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#914 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#7 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#946 FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq#19 FD: 25 BD: 1 +.+.: (wq_completion)nfc35_nci_tx_wq#13 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#203 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#221 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#262 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#615 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1214 FD: 25 BD: 1 +.+.: (wq_completion)nfc17_nci_rx_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#867 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#495 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#56 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#431 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#907 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#56 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#418 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#903 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#413 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#418 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#866 FD: 166 BD: 1 +.+.: (wq_completion)hci4#6 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#451 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#442 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#992 FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg1#206 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1071 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1081 FD: 1 BD: 51 +.-.: &jsk->filters_lock FD: 33 BD: 2 +.+.: (work_completion)(&pool->idle_cull_work) ->wq_pool_attach_mutex ->wq_pool_attach_mutex.wait_lock ->&p->pi_lock ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#941 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#533 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1091 FD: 27 BD: 1 +.+.: sk_lock-AF_ISDN ->&rq->__lock ->slock-AF_ISDN ->clock-AF_ISDN FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#501 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1075 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_rx_wq#502 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1049 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#501 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1064 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#69 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#468 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#144 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1041 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1023 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1036 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1035 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#471 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1014 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1022 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#462 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#188 ->&rq->__lock FD: 179 BD: 1 +.+.: (wq_completion)hci5 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#187 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#64 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1005 ->&rq->__lock FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg0#217 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#337 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#928 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#466 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#588 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1001 FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg0#218 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1002 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#878 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#898 FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg2#216 ->(work_completion)(&peer->transmit_handshake_work) FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1018 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#61 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#463 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1009 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#185 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1000 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#454 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#458 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#462 ->&rq->__lock FD: 1 BD: 53 +.+.: (wq_completion)phy213 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#883 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#890 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#565 ->&rq->__lock FD: 67 BD: 11 +.+.: &sb->s_type->i_mutex_key#20 ->&rq->__lock ->rename_lock.seqcount ->fs_reclaim ->stock_lock ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#37 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&c->lock ->&n->list_lock ->&pipe->lock ->&obj_hash[i].lock ->(work_completion)(&(&pipe->queue_timeout)->work) ->&____s->seqcount#2 ->&____s->seqcount FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#992 ->&rq->__lock FD: 25 BD: 53 +.+.: (wq_completion)phy214 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#460 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#998 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#449 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#880 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#879 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#991 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#452 ->&rq->__lock FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#107 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#9 FD: 1 BD: 5 +.+.: (wq_completion)tipc_rcv#10 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#278 ->&rq->__lock FD: 1 BD: 5 +.+.: (wq_completion)tipc_send#10 FD: 1 BD: 5 +.+.: (wq_completion)tipc_crypto#10 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#459 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1004 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#996 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#179 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#988 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1000 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#993 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#982 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#176 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#972 FD: 1 BD: 53 +.+.: (wq_completion)phy230 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#977 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#175 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#438 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#965 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#971 FD: 179 BD: 1 +.+.: (wq_completion)hci1#5 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#548 FD: 25 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#14 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#19 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#805 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#945 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#170 FD: 25 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#20 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#20 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1194 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#492 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#12 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#48 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#490 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#456 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#453 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#178 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#436 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#970 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1003 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#778 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#804 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#233 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#217 FD: 25 BD: 2 +.+.: (wq_completion)nfc22_nci_cmd_wq#15 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#984 FD: 77 BD: 2 +.+.: &type->s_umount_key#74/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->&c->lock ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#36 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&dentry->d_lock ->&n->list_lock ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#415 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#152 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#824 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#825 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#955 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#53 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#55 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#431 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#159 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#53 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#27 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#28 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#26 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#55 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#436 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#441 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#27 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#161 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#441 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#447 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#449 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#176 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#449 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#454 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#64 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#967 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#179 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#25 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#953 FD: 25 BD: 5 +.+.: (wq_completion)tipc_send#7 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#464 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#176 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#464 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#177 FD: 25 BD: 2 +.+.: (wq_completion)nfc42_nci_cmd_wq#6 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#449 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#469 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#65 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#69 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#464 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#472 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#181 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#959 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#468 FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg0#211 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#972 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#182 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#467 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#66 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#187 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#177 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#452 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#488 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1046 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1010 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1175 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#32 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#75 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#198 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#699 ->&rq->__lock FD: 38 BD: 5 +.+.: &sb->s_type->i_lock_key#35 ->&dentry->d_lock ->&p->pi_lock FD: 30 BD: 9 +.+.: &type->lock_class#3/1 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#14 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#79 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc20_nci_cmd_wq#18 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#512 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#209 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#78 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#511 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#515 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#526 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#213 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#737 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#766 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#394 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#956 FD: 1 BD: 51 +...: &net->can.rcvlists_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#433 FD: 38 BD: 12 +.+.: &sb->s_type->i_lock_key#37 ->&dentry->d_lock FD: 1 BD: 3 ....: init_user_ns.keyring_sem.wait_lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#700 ->&rq->__lock FD: 38 BD: 4 +.+.: &sb->s_type->i_lock_key#36 ->&dentry->d_lock FD: 1 BD: 2 +...: slock-AF_ISDN FD: 78 BD: 1 +.+.: &type->s_umount_key#77 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->&sn->pipefs_sb_lock ->rename_lock.seqcount ->&rq->__lock ->&dentry->d_lock ->&dentry->d_lock/1 ->&sb->s_type->i_lock_key#37 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->pool_lock#2 ->&fsnotify_mark_srcu FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#421 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#409 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#135 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#392 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#395 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_rx_wq#256 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1210 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#10 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#92 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#244 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#986 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_rx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc21_nci_cmd_wq#10 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#103 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1053 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#487 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#99 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1047 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#471 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#263 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1033 FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg1#93 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg0#186 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg1#185 ->(work_completion)(&peer->transmit_handshake_work) FD: 1 BD: 1 +.+.: &map->freeze_mutex FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg2#93 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#692 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#689 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#123 ->&rq->__lock FD: 43 BD: 5 +.+.: (wq_completion)bond0#106 ->(work_completion)(&(&slave->notify_work)->work) FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#766 ->&rq->__lock FD: 77 BD: 10 +.+.: &sb->s_type->i_mutex_key#20/1 ->rename_lock.seqcount ->fs_reclaim ->stock_lock ->&rq->__lock ->pool_lock#2 ->&dentry->d_lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#37 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&____s->seqcount ->&sb->s_type->i_mutex_key#20 ->&fsnotify_mark_srcu ->&xa->xa_lock#8 ->&obj_hash[i].lock ->&c->lock ->&n->list_lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#438 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#60 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#167 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#170 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#446 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#444 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#482 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#182 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#463 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#461 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#188 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#470 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#471 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#71 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#190 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#487 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#487 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#448 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#488 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#485 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#31 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#29 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#492 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#200 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#169 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#49 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#408 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#51 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#410 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#142 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#407 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#405 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#407 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#410 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#140 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#141 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#145 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#49 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#48 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#46 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#46 ->&rq->__lock FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#111 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#281 FD: 1 BD: 1 +.+.: (wq_completion)nfc29_nci_rx_wq#10 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#246 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#11 FD: 43 BD: 5 +.+.: (wq_completion)bond0#110 ->(work_completion)(&(&slave->notify_work)->work) ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#47 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#404 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#406 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#409 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#139 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#138 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#139 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#143 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#403 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#405 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#45 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#45 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#46 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#137 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#138 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#402 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#142 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#404 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#407 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#403 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#402 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#405 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#136 FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg0#194 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#313 ->&rq->__lock FD: 179 BD: 1 +.+.: (wq_completion)hci3#5 ->(work_completion)(&hdev->power_on) ->(work_completion)(&hdev->cmd_sync_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#587 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#580 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#576 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#253 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#990 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#246 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#73 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#575 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#586 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#137 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#395 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#134 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#135 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#139 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#393 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#395 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#133 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#132 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#133 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#137 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#44 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#44 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#45 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#132 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#136 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#390 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#130 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#131 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#135 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#389 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1233 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1231 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1218 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1193 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#591 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1189 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1188 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1184 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1171 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#573 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#247 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1180 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#566 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#246 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1155 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1178 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#581 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1163 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1174 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#578 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1172 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1159 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1171 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1148 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1169 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#943 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#951 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#768 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#998 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#191 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#938 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_tx_wq#559 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1168 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#557 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1166 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#561 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1165 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1154 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#235 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#84 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#555 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#570 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#560 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#569 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1163 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1162 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#87 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1160 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#567 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#86 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#238 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1158 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1146 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#565 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#564 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1156 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1155 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1144 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#235 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#938 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#940 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#947 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#25 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#948 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#949 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#561 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#560 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#551 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#550 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1151 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1140 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1130 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#559 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1149 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#557 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1147 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1136 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#226 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1145 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1134 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1144 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1122 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1142 FD: 25 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#20 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#932 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#432 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#440 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#939 FD: 43 BD: 5 +.+.: (wq_completion)bond0#113 ->(work_completion)(&(&slave->notify_work)->work) FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#172 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#933 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#430 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#60 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#174 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#20 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1141 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1140 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1138 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_tx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#20 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#186 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#16 FD: 25 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#20 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#720 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#15 FD: 94 BD: 2 +.+.: &type->s_umount_key#76/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->&xa->xa_lock#15 ->&rq->__lock ->&obj_hash[i].lock ->stock_lock ->&____s->seqcount#2 ->&____s->seqcount ->&c->lock ->mmu_notifier_invalidate_range_start ->&sb->s_type->i_lock_key#37 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&sb->s_type->i_mutex_key#20 ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_mutex_key#20/1 ->&sn->pipefs_sb_lock ->&n->list_lock FD: 1 BD: 10 +.+.: &sn->rpc_client_lock FD: 25 BD: 1 +.+.: (wq_completion)nfc16_nci_rx_wq#17 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#17 FD: 1 BD: 1 +...: data_sockets.lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#366 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 729 BD: 13 +.+.: &devlink->lock_key#115 ->crngs.lock ->fs_reclaim ->devlinks.xa_lock ->&c->lock ->&xa->xa_lock#19 ->pcpu_alloc_mutex ->&obj_hash[i].lock ->&base->lock ->&rq->__lock ->pin_fs_lock ->&sb->s_type->i_mutex_key#3 ->batched_entropy_u32.lock ->rtnl_mutex ->&(&fn_net->fib_chain)->lock ->&n->list_lock ->stack_depot_init_mutex ->rtnl_mutex.wait_lock ->&p->pi_lock ->&devlink_port->type_lock ->&nsim_trap_data->trap_lock FD: 25 BD: 1 +.+.: (wq_completion)nfc15_nci_rx_wq#15 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#227 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#543 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1135 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1129 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1128 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1127 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#536 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#225 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#534 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#528 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1121 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1111 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#224 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#17 FD: 25 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#14 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq#15 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1098 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#20 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1107 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#222 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#536 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc20_nci_cmd_wq#16 FD: 166 BD: 14 +.+.: (wq_completion)hci1#6 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#937 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#825 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#802 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#522 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#970 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#963 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#958 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#955 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#945 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#863 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#443 FD: 1 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc39_nci_cmd_wq#6 FD: 1 BD: 2 +.+.: (wq_completion)nfc20_nci_cmd_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#51 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#570 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#858 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc35_nci_cmd_wq#12 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#869 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc36_nci_cmd_wq#7 FD: 25 BD: 1 +.+.: (wq_completion)nfc29_nci_tx_wq#14 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc29_nci_cmd_wq#14 FD: 25 BD: 2 +.+.: (wq_completion)nfc23_nci_cmd_wq#15 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#15 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#16 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#849 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#396 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#856 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#851 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#149 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#854 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#846 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#852 ->&rq->__lock FD: 1 BD: 2 +.+.: uuid_mutex FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#628 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#266 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#626 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc36_nci_rx_wq#5 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#270 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#847 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#393 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#850 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#843 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#848 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#842 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#390 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#846 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#845 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#840 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#843 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#838 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#837 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#378 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#841 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#840 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#833 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#837 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#827 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#384 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#376 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#832 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#827 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#830 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#826 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#819 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#815 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#812 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#816 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#818 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#819 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#813 FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg2#204 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg0#105 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg1#204 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg2#203 ->(work_completion)(&peer->transmit_handshake_work) FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg0#206 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg1#203 ->(work_completion)(&peer->transmit_handshake_work) FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#817 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#815 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#808 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#813 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#811 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#806 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#810 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#382 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc23_nci_cmd_wq#14 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#452 FD: 69 BD: 1 +.+.: &type->s_umount_key#71 ->&x->wait#23 ->shrinker_mutex ->&obj_hash[i].lock ->rename_lock.seqcount ->&dentry->d_lock ->&sb->s_type->i_lock_key#34 ->&s->s_inode_list_lock ->&xa->xa_lock#8 ->&fsnotify_mark_srcu ->&dentry->d_lock/1 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#13 FD: 77 BD: 2 +.+.: &type->s_umount_key#70/1 ->fs_reclaim ->pool_lock#2 ->pcpu_alloc_mutex ->&c->lock ->shrinker_mutex ->list_lrus_mutex ->sb_lock ->mmu_notifier_invalidate_range_start ->&xa->xa_lock#15 ->&obj_hash[i].lock ->stock_lock ->&____s->seqcount#2 ->&____s->seqcount ->&sb->s_type->i_lock_key#34 ->&s->s_inode_list_lock ->tk_core.seq.seqcount ->&rq->__lock ->&dentry->d_lock ->&n->list_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#369 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#379 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#371 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#800 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#365 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#372 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#781 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#780 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#355 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#761 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#754 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#736 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#726 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#716 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#13 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#276 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#713 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#697 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#690 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#312 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#688 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#685 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#116 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#686 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#683 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#684 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#683 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#680 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#679 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#676 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#673 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#295 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#660 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#665 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#662 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#659 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#293 ->&rq->__lock ->&cfs_rq->removed.lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#653 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#653 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#653 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#646 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#643 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#641 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#637 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#633 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#279 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#169 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#404 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#401 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#403 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#402 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#140 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#401 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#398 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#396 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#136 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#399 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#397 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#400 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#398 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#764 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#765 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#358 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#782 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#356 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#773 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#778 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#326 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#359 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#770 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#399 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#141 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#400 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#397 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#399 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#396 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#394 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#398 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#397 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#394 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#138 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#621 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#257 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#614 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#616 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#533 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#531 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#225 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#221 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#220 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#87 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#85 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#84 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#532 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#226 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#222 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#221 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#172 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#538 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#533 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#227 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#223 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#222 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#88 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#224 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#223 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#536 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#534 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#225 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#224 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#540 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#537 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#535 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#230 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#226 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#225 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#226 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#541 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#89 FD: 25 BD: 53 +.+.: (wq_completion)phy228 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#87 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#538 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#85 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#536 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#537 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#232 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#228 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#227 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#540 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#538 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#229 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#228 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#88 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#544 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#541 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#86 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#539 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#34 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#32 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#545 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#542 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#540 ->&rq->__lock FD: 1 BD: 53 +.+.: (wq_completion)phy227 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#543 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#541 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#234 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#230 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#229 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#235 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#231 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#544 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#230 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#542 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#89 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#87 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#548 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#543 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#546 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#544 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#232 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#550 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#231 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#547 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#545 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#548 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#546 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#233 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#232 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#549 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#547 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#550 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#548 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#234 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#551 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#549 FD: 28 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#555 ->&rq->__lock ->rcu_node_0 ->&rcu_state.expedited_wq FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#550 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#553 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#551 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#239 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#235 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#234 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#554 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#552 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#240 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#558 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#236 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#235 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#555 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#553 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#556 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#554 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#557 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#555 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#241 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#237 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#236 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#558 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#556 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#242 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#238 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#237 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#559 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#557 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#562 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#560 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#563 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#561 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#240 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#239 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#567 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#562 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#565 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#563 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#93 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#91 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#89 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#13 FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg0#209 ->(work_completion)(&peer->transmit_handshake_work) FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg1#207 ->(work_completion)(&peer->transmit_handshake_work) FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg2#207 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 87 BD: 50 +.+.: (wq_completion)wg-kex-wg1#208 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg1#104 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg2#104 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 53 BD: 50 +.+.: (wq_completion)wg-kex-wg2#208 ->(work_completion)(&peer->transmit_handshake_work) FD: 88 BD: 50 +.+.: (wq_completion)wg-crypt-wg0#107 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_tx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_tx_wq#11 FD: 25 BD: 2 +.+.: (wq_completion)nfc24_nci_cmd_wq#9 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_rx_wq#9 FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_tx_wq#9 FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq#11 FD: 25 BD: 2 +.+.: (wq_completion)nfc29_nci_cmd_wq#9 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc29_nci_rx_wq#9 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc29_nci_tx_wq#9 FD: 1 BD: 2 +.+.: (wq_completion)nfc30_nci_cmd_wq#9 FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_rx_wq#9 FD: 25 BD: 1 +.+.: (wq_completion)nfc30_nci_tx_wq#9 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#245 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#241 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#240 FD: 1 BD: 2 +.+.: (wq_completion)nfc35_nci_cmd_wq#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#242 FD: 25 BD: 2 +.+.: (wq_completion)nfc37_nci_cmd_wq#4 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#241 FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_rx_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_tx_wq#4 FD: 1 BD: 2 +.+.: (wq_completion)nfc36_nci_cmd_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc36_nci_rx_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc36_nci_tx_wq#4 FD: 1 BD: 2 +.+.: (wq_completion)nfc34_nci_cmd_wq#9 FD: 25 BD: 1 +.+.: (wq_completion)nfc34_nci_tx_wq#9 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc33_nci_cmd_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_rx_wq#10 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#247 FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_tx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#243 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#242 FD: 25 BD: 2 +.+.: (wq_completion)nfc32_nci_cmd_wq#9 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_rx_wq#9 FD: 25 BD: 1 +.+.: (wq_completion)nfc32_nci_tx_wq#9 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc30_nci_cmd_wq#10 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc35_nci_cmd_wq#9 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_rx_wq#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_tx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_tx_wq#8 FD: 25 BD: 2 +.+.: (wq_completion)nfc29_nci_cmd_wq#10 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc29_nci_tx_wq#10 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc31_nci_cmd_wq#9 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc31_nci_rx_wq#9 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc31_nci_tx_wq#9 FD: 25 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc28_nci_cmd_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_tx_wq#10 FD: 1 BD: 2 +.+.: (wq_completion)nfc27_nci_cmd_wq#11 FD: 25 BD: 1 +.+.: (wq_completion)nfc27_nci_rx_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc27_nci_tx_wq#11 FD: 25 BD: 2 +.+.: (wq_completion)nfc24_nci_cmd_wq#10 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc24_nci_rx_wq#10 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc29_nci_cmd_wq#11 FD: 25 BD: 1 +.+.: (wq_completion)nfc29_nci_rx_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_tx_wq#10 FD: 25 BD: 1 +.+.: (wq_completion)nfc29_nci_tx_wq#11 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#248 FD: 1 BD: 2 +.+.: (wq_completion)nfc23_nci_cmd_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#244 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_tx_wq#12 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#243 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc25_nci_cmd_wq#9 FD: 25 BD: 1 +.+.: (wq_completion)nfc25_nci_rx_wq#9 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc25_nci_tx_wq#9 FD: 1 BD: 2 +.+.: (wq_completion)nfc22_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc22_nci_rx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc20_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc21_nci_cmd_wq#9 FD: 25 BD: 1 +.+.: (wq_completion)nfc21_nci_rx_wq#9 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_tx_wq#9 FD: 1 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#11 FD: 25 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#11 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc18_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc17_nci_cmd_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc17_nci_rx_wq#10 FD: 25 BD: 1 +.+.: (wq_completion)nfc17_nci_tx_wq#10 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#14 FD: 25 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#10 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#10 FD: 1 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#10 FD: 25 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#10 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#11 FD: 25 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#10 FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#12 FD: 25 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#11 FD: 25 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#11 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#94 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#92 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#90 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#569 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#566 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#564 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#14 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#35 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#34 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#33 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#567 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#245 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#565 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#244 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#95 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#93 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#91 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#568 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#566 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#569 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#567 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#250 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#245 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#96 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#94 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#92 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#573 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#570 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#568 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#95 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#251 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#247 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#246 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#36 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#35 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#93 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#574 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#34 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#571 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#569 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#248 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#247 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#572 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#570 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#573 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#571 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#96 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#94 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#249 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#248 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#577 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#574 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#572 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#37 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#97 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#254 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#95 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#250 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#36 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#35 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#249 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#575 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#573 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#579 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#98 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#96 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#574 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#255 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#251 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#250 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#38 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#37 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#36 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#577 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#575 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#256 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#252 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#251 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#578 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#576 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#582 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#579 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#577 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#257 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#253 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#583 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#580 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#258 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#254 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#253 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#584 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#581 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#579 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#582 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#580 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#259 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#255 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#583 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#254 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#581 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#584 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#582 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#260 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#256 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#255 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#588 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#585 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#583 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#586 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#584 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#261 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#590 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#587 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#585 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#256 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#588 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#586 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#592 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#589 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#587 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#262 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#590 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#588 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#594 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#589 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#595 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#592 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#590 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#258 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#257 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#593 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#591 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#597 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#594 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#592 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#259 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#258 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#598 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#595 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#593 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#101 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#599 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#97 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#594 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#600 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#597 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#601 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#598 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#596 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#265 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#259 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#602 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#599 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#597 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#266 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#261 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#260 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#600 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#598 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#261 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#604 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#601 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#599 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#605 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#602 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#268 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#263 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#262 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#600 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#603 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#601 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#602 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#269 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#264 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#263 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#608 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#603 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#604 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#265 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#264 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#610 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#607 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#605 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#271 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#266 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#265 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#608 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#612 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#607 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#613 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#610 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#608 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#272 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#267 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#614 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#611 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#273 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#268 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#609 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#267 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#612 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#610 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#274 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#268 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#616 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#613 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#611 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#614 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#275 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#270 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#269 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#271 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#270 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#277 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#618 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#615 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#613 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#272 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#271 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#100 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#98 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#616 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#614 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#620 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#617 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#615 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#273 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#272 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#621 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#618 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#616 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#101 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#99 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#279 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#274 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#273 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#619 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#617 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#280 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#275 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#623 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#274 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#620 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#618 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#621 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#619 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#276 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#275 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#104 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#102 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#100 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#625 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#622 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#620 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#623 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#621 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#282 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#277 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#627 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#624 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#622 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#13 FD: 25 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#11 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#15 FD: 25 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#11 FD: 25 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#11 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc16_nci_rx_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_tx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_tx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_tx_wq#10 FD: 25 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc25_nci_cmd_wq#10 FD: 25 BD: 1 +.+.: (wq_completion)nfc25_nci_rx_wq#10 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc25_nci_tx_wq#10 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq#14 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq#14 FD: 25 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq#15 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc48_nci_tx_wq#3 FD: 1 BD: 2 +.+.: (wq_completion)nfc48_nci_rx_wq#3 FD: 1 BD: 2 +.+.: (wq_completion)nfc48_nci_cmd_wq#3 FD: 25 BD: 2 +.+.: (wq_completion)nfc47_nci_cmd_wq#3 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc47_nci_rx_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc47_nci_tx_wq#3 FD: 1 BD: 2 +.+.: (wq_completion)nfc46_nci_cmd_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc46_nci_rx_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc46_nci_tx_wq#3 FD: 1 BD: 2 +.+.: (wq_completion)nfc45_nci_cmd_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc45_nci_rx_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc45_nci_tx_wq#3 FD: 1 BD: 2 +.+.: (wq_completion)nfc44_nci_cmd_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc44_nci_rx_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc44_nci_tx_wq#3 FD: 1 BD: 2 +.+.: (wq_completion)nfc43_nci_cmd_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc43_nci_rx_wq#3 FD: 1 BD: 1 +.+.: (wq_completion)nfc43_nci_tx_wq#3 FD: 1 BD: 2 +.+.: (wq_completion)nfc42_nci_cmd_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc42_nci_rx_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc42_nci_tx_wq#4 FD: 1 BD: 2 +.+.: (wq_completion)nfc41_nci_cmd_wq#4 FD: 25 BD: 1 +.+.: (wq_completion)nfc41_nci_rx_wq#4 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc41_nci_tx_wq#4 FD: 1 BD: 2 +.+.: (wq_completion)nfc40_nci_cmd_wq#5 FD: 25 BD: 1 +.+.: (wq_completion)nfc40_nci_rx_wq#5 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc40_nci_tx_wq#5 FD: 1 BD: 2 +.+.: (wq_completion)nfc39_nci_cmd_wq#4 FD: 25 BD: 1 +.+.: (wq_completion)nfc39_nci_rx_wq#4 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc39_nci_tx_wq#4 FD: 1 BD: 2 +.+.: (wq_completion)nfc38_nci_cmd_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc38_nci_rx_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc38_nci_tx_wq#4 FD: 25 BD: 2 +.+.: (wq_completion)nfc37_nci_cmd_wq#5 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_rx_wq#5 FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_tx_wq#5 FD: 25 BD: 2 +.+.: (wq_completion)nfc36_nci_cmd_wq#5 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc36_nci_tx_wq#5 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc35_nci_cmd_wq#10 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc35_nci_rx_wq#9 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_tx_wq#9 FD: 1 BD: 2 +.+.: (wq_completion)nfc34_nci_cmd_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_tx_wq#10 FD: 25 BD: 2 +.+.: (wq_completion)nfc33_nci_cmd_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_rx_wq#11 FD: 25 BD: 1 +.+.: (wq_completion)nfc33_nci_tx_wq#11 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc32_nci_cmd_wq#10 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_tx_wq#10 FD: 1 BD: 2 +.+.: (wq_completion)nfc31_nci_cmd_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc31_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc31_nci_tx_wq#10 FD: 1 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq#16 FD: 1 BD: 2 +.+.: (wq_completion)nfc30_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_tx_wq#11 FD: 25 BD: 2 +.+.: (wq_completion)nfc29_nci_cmd_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc29_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc29_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc25_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc25_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc25_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc28_nci_cmd_wq#12 FD: 25 BD: 1 +.+.: (wq_completion)nfc28_nci_rx_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_tx_wq#12 FD: 25 BD: 2 +.+.: (wq_completion)nfc27_nci_cmd_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc27_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc27_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc24_nci_cmd_wq#11 FD: 25 BD: 1 +.+.: (wq_completion)nfc24_nci_rx_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc21_nci_cmd_wq#11 FD: 25 BD: 1 +.+.: (wq_completion)nfc21_nci_rx_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc23_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc22_nci_cmd_wq#12 FD: 25 BD: 1 +.+.: (wq_completion)nfc22_nci_rx_wq#12 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc20_nci_cmd_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#12 FD: 25 BD: 2 +.+.: (wq_completion)nfc18_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_rx_wq#13 FD: 25 BD: 1 +.+.: (wq_completion)nfc18_nci_tx_wq#13 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc16_nci_cmd_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_tx_wq#14 FD: 1 BD: 2 +.+.: (wq_completion)nfc17_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc17_nci_rx_wq#11 FD: 25 BD: 1 +.+.: (wq_completion)nfc17_nci_tx_wq#11 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#12 FD: 25 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#16 FD: 25 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#16 FD: 1 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#11 FD: 25 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#12 FD: 25 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#12 FD: 25 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#14 FD: 25 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#14 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#13 FD: 25 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#39 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#38 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#37 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#105 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#101 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#283 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#278 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#277 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#626 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#624 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#627 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#625 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#628 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#626 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#284 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#278 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#106 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#632 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#40 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#39 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#104 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#38 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#629 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#102 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#627 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#630 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#628 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#285 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#279 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#634 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#631 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#629 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#632 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#630 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#633 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#631 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#634 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#632 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#286 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#281 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#280 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#635 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#633 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#636 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#634 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#287 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#282 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#281 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#640 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#107 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#635 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#105 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#103 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#636 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#642 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#637 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#640 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#638 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#641 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#639 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#642 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#640 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#283 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#282 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#643 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#641 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#289 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#284 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#283 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#647 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#644 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#642 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#290 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#285 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#284 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#286 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#285 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#649 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#646 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#644 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#292 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#287 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#286 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#108 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#104 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#41 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#40 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#39 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#647 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#645 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#293 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#288 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#287 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#648 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#646 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_rx_wq#289 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#294 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#109 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#107 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#105 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#295 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#649 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#647 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#650 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#648 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#110 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#296 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#108 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#290 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#288 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#654 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#651 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#649 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#655 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#652 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#650 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#291 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#289 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#656 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#298 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#292 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#290 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#657 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#652 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#655 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#299 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#659 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#291 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#656 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#654 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#111 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#109 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#107 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#660 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#292 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#657 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#655 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#112 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#110 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#661 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#108 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#656 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#657 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#663 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#660 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#301 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#658 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#295 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#293 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#661 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#659 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#662 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#296 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#294 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#666 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#663 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#661 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#664 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#662 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#665 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#663 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#669 ->&rq->__lock ->&cfs_rq->removed.lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#664 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#303 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#297 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#670 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#667 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#665 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#668 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#666 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#669 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#667 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#670 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#668 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#298 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#674 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#671 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#669 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#113 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#111 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#109 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#675 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#672 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#673 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#671 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#112 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#110 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#305 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#299 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#297 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#677 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#674 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#672 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#673 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#306 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#300 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#298 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#676 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#674 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#677 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#675 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#307 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#299 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#678 FD: 29 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#682 ->&rq->__lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#677 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#680 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#678 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#308 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#302 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#300 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#679 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#115 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#113 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#309 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#303 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#301 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#685 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#682 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#680 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#681 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#310 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#304 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#302 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#687 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#684 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#114 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#112 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#682 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#683 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#311 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#305 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#303 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#684 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#306 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#304 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#687 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#685 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#307 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#117 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#305 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#115 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#691 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#688 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#686 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#113 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#689 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#687 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#314 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#308 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#306 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#690 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#688 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#118 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#116 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#315 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#114 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#309 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#307 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#689 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#690 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#316 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#310 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#308 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#119 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#117 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#115 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#317 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#309 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#318 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#312 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#310 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#693 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#120 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#118 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#691 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#116 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#692 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#121 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#119 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#117 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#313 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#311 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#698 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#693 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#696 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#694 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#320 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#314 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#312 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#13 FD: 25 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#13 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#16 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#42 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#40 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#122 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#120 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#118 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#321 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#315 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#313 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#43 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#700 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#42 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#697 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#695 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#41 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#698 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#696 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#322 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#316 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#314 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#315 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#121 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#119 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#699 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#697 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#324 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#698 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#318 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#316 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#704 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#701 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#699 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#700 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#44 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#706 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#701 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#124 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#122 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#120 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#325 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#319 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#317 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#704 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#702 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#326 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#705 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#320 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#703 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#318 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#125 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#123 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#121 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#43 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#42 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#709 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#706 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#704 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#17 FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#17 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#321 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#319 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#705 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#328 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#322 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#320 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#711 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#708 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#323 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#321 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#712 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#709 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#707 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#126 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#124 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#122 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#46 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#44 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#43 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#714 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#710 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#708 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#324 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#322 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#711 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#709 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#323 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#716 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#712 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#710 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#127 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#125 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#717 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#713 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#718 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#714 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#712 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#324 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#18 FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#18 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#126 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#124 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#719 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#715 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#334 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#713 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#45 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#44 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#714 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#721 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#717 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#715 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#335 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#325 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#716 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#336 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#328 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#326 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#723 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#719 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#717 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#329 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#327 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#338 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#330 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#328 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#719 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#726 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#722 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#723 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#721 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#724 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#722 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#723 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#331 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#329 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#724 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#725 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#332 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#330 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#333 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#331 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#727 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#734 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#728 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#731 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#729 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#342 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#334 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#332 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#732 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#730 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#737 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#731 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#333 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#738 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#732 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#129 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#127 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#125 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#735 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#733 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#336 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#334 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#740 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#734 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#130 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#128 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#126 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#735 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#337 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#335 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#742 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#738 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#736 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#737 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#740 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#738 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#346 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#338 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#336 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#741 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#739 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#742 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#740 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#741 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#744 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#742 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#347 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#745 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#749 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#339 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#337 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#348 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#340 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#338 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#744 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#752 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#748 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#745 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#349 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#341 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#339 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#129 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#127 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#749 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#746 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#750 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#747 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#350 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#342 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#340 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#748 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#343 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#341 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#132 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#130 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#752 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#749 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#133 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#131 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#129 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#344 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#342 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#753 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#750 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#758 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#754 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#345 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#343 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#759 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#134 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#132 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#130 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#755 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#346 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#344 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#760 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#756 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#753 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#754 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#762 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#347 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#345 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#356 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#346 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#357 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#763 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#758 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#349 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#755 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#347 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#759 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#756 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#350 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#348 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#351 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#349 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#135 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#765 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#133 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#760 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#131 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#757 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#360 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#352 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#134 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#350 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#132 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#761 ->&rq->__lock ->&cfs_rq->removed.lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#758 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#46 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#45 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#767 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#762 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#759 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#361 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#353 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#351 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#763 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#760 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#362 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#352 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#137 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#135 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#133 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#769 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#764 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#761 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#762 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#355 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#353 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#766 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#763 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#772 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#767 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#764 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#354 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#768 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#765 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#365 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#357 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#355 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#774 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#769 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#775 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#770 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#767 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#771 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#768 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#772 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#769 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#770 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#367 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#359 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#357 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#774 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#771 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#368 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#138 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#360 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#358 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#136 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#134 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#772 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#361 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#359 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#139 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#137 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#135 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#781 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#776 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#773 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#47 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#46 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#777 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#774 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#775 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#779 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#776 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#370 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#362 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#360 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#777 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#778 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#371 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#363 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#361 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#782 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#779 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#364 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#362 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#363 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#788 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#783 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#780 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#789 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#784 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#781 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#138 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#136 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#782 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#786 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#783 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#787 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#784 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#374 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#366 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#364 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#141 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#137 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#793 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#788 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#785 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#789 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#367 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#365 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#795 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#790 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#787 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#142 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#138 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#791 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#143 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#368 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#788 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#366 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#141 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#139 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#377 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#369 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#367 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#144 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#142 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#140 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#370 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#798 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#793 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#790 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#794 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#791 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#795 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#792 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_tx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc27_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc43_nci_tx_wq#4 FD: 1 BD: 2 +.+.: (wq_completion)nfc43_nci_rx_wq#4 FD: 1 BD: 2 +.+.: (wq_completion)nfc43_nci_cmd_wq#4 FD: 1 BD: 2 +.+.: (wq_completion)nfc42_nci_cmd_wq#5 FD: 1 BD: 1 +.+.: (wq_completion)nfc42_nci_rx_wq#5 FD: 1 BD: 1 +.+.: (wq_completion)nfc42_nci_tx_wq#5 FD: 1 BD: 2 +.+.: (wq_completion)nfc41_nci_cmd_wq#5 FD: 25 BD: 1 +.+.: (wq_completion)nfc41_nci_rx_wq#5 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc41_nci_tx_wq#5 FD: 25 BD: 1 +.+.: (wq_completion)nfc40_nci_rx_wq#6 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc40_nci_tx_wq#6 FD: 1 BD: 2 +.+.: (wq_completion)nfc39_nci_cmd_wq#5 FD: 25 BD: 1 +.+.: (wq_completion)nfc39_nci_rx_wq#5 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc39_nci_tx_wq#5 FD: 1 BD: 2 +.+.: (wq_completion)nfc38_nci_cmd_wq#5 FD: 1 BD: 1 +.+.: (wq_completion)nfc38_nci_rx_wq#5 FD: 1 BD: 1 +.+.: (wq_completion)nfc38_nci_tx_wq#5 FD: 25 BD: 2 +.+.: (wq_completion)nfc37_nci_cmd_wq#6 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_rx_wq#6 FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_tx_wq#6 FD: 25 BD: 2 +.+.: (wq_completion)nfc33_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc36_nci_cmd_wq#6 FD: 1 BD: 1 +.+.: (wq_completion)nfc36_nci_rx_wq#6 FD: 1 BD: 2 +.+.: (wq_completion)nfc35_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_rx_wq#10 FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_tx_wq#10 FD: 1 BD: 2 +.+.: (wq_completion)nfc34_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_tx_wq#11 FD: 25 BD: 2 +.+.: (wq_completion)nfc32_nci_cmd_wq#11 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_rx_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_tx_wq#11 FD: 1 BD: 2 +.+.: (wq_completion)nfc31_nci_cmd_wq#11 FD: 1 BD: 1 +.+.: (wq_completion)nfc31_nci_rx_wq#11 FD: 25 BD: 1 +.+.: (wq_completion)nfc31_nci_tx_wq#11 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc30_nci_cmd_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_tx_wq#12 FD: 25 BD: 1 +.+.: (wq_completion)nfc29_nci_rx_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc29_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc27_nci_cmd_wq#14 FD: 25 BD: 1 +.+.: (wq_completion)nfc27_nci_rx_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc27_nci_tx_wq#14 FD: 25 BD: 2 +.+.: (wq_completion)nfc28_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq#17 FD: 1 BD: 2 +.+.: (wq_completion)nfc25_nci_cmd_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc25_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc25_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc24_nci_cmd_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_tx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_tx_wq#14 FD: 1 BD: 2 +.+.: (wq_completion)nfc20_nci_cmd_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_tx_wq#14 FD: 1 BD: 2 +.+.: (wq_completion)nfc22_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc22_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc22_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc21_nci_cmd_wq#12 FD: 25 BD: 1 +.+.: (wq_completion)nfc21_nci_rx_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc18_nci_cmd_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_rx_wq#14 FD: 25 BD: 1 +.+.: (wq_completion)nfc18_nci_tx_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc17_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc17_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc16_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_tx_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#14 FD: 1 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#17 FD: 25 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#13 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#13 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#16 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#16 FD: 25 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#14 FD: 25 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#14 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#14 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#20 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#20 FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#20 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#380 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#372 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#370 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#50 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#48 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#47 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#143 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#141 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#797 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#794 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#799 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#796 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#800 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#797 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#801 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#798 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#799 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#381 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#373 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#371 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#808 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#803 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#800 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#804 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#801 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#374 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#805 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#372 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#802 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#807 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#804 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#805 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#809 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#806 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#807 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#808 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#809 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#811 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#375 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#815 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#373 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#820 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#817 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#826 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#821 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#818 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#822 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#819 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#823 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#820 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#823 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#824 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#825 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#829 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#834 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#374 ->&rq->__lock FD: 29 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#835 ->&rq->__lock ->stock_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#830 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#146 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#144 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#142 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#831 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#828 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#829 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#147 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#385 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#145 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#143 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#377 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#375 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#838 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#830 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#834 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#831 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#835 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#832 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#836 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#833 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#386 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#376 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#834 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#387 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#379 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#377 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#380 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#378 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#844 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#839 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#836 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#389 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#148 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#381 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#379 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#146 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#144 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#841 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#838 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#382 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#380 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#839 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#840 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#844 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#841 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#845 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#842 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#149 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#391 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#147 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#145 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#383 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#381 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#382 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#846 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#843 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#385 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#150 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#383 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#148 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#146 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#394 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#386 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#384 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#848 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#845 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#850 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#847 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#395 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#387 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#385 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#151 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#147 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#848 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#852 FD: 25 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#15 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#16 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#16 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#15 FD: 25 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#15 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc23_nci_rx_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_tx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#850 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#397 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#389 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#387 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#152 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc40_nci_cmd_wq#7 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc40_nci_rx_wq#7 FD: 25 BD: 1 +.+.: (wq_completion)nfc40_nci_tx_wq#7 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#150 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#148 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#49 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#48 FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#21 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#21 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#21 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#16 FD: 25 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#17 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#17 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#14 FD: 25 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#14 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#153 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#151 FD: 25 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#149 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#18 FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#17 FD: 1 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#13 FD: 25 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#13 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#13 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq#14 FD: 25 BD: 1 +.+.: (wq_completion)nfc15_nci_rx_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_tx_wq#14 FD: 25 BD: 2 +.+.: (wq_completion)nfc16_nci_cmd_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_rx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_tx_wq#16 FD: 1 BD: 2 +.+.: (wq_completion)nfc17_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc17_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc17_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc18_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_tx_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#14 FD: 1 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#16 FD: 25 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_tx_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc21_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc22_nci_cmd_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc22_nci_rx_wq#14 FD: 25 BD: 1 +.+.: (wq_completion)nfc22_nci_tx_wq#14 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc24_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_tx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc25_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc25_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc25_nci_tx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc39_nci_rx_wq#6 FD: 25 BD: 1 +.+.: (wq_completion)nfc39_nci_tx_wq#6 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#859 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#854 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#851 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#398 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#390 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#388 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#52 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#50 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#49 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#154 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#152 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#150 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#22 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#22 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#22 FD: 25 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#18 FD: 25 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#18 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#17 FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#19 FD: 25 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#19 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#19 FD: 1 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#16 FD: 25 BD: 2 +.+.: (wq_completion)nfc38_nci_cmd_wq#6 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc38_nci_rx_wq#6 FD: 25 BD: 1 +.+.: (wq_completion)nfc38_nci_tx_wq#6 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc37_nci_cmd_wq#7 FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_rx_wq#7 FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_tx_wq#7 FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_rx_wq#12 FD: 25 BD: 1 +.+.: (wq_completion)nfc35_nci_tx_wq#12 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc36_nci_cmd_wq#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc36_nci_rx_wq#7 FD: 1 BD: 1 +.+.: (wq_completion)nfc36_nci_tx_wq#7 FD: 1 BD: 2 +.+.: (wq_completion)nfc34_nci_cmd_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_rx_wq#12 FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_tx_wq#12 FD: 1 BD: 2 +.+.: (wq_completion)nfc33_nci_cmd_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_tx_wq#14 FD: 1 BD: 2 +.+.: (wq_completion)nfc32_nci_cmd_wq#12 FD: 25 BD: 1 +.+.: (wq_completion)nfc32_nci_rx_wq#12 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_tx_wq#12 FD: 25 BD: 2 +.+.: (wq_completion)nfc29_nci_cmd_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc29_nci_rx_wq#15 FD: 25 BD: 1 +.+.: (wq_completion)nfc29_nci_tx_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc31_nci_rx_wq#12 FD: 25 BD: 1 +.+.: (wq_completion)nfc31_nci_tx_wq#12 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc30_nci_cmd_wq#13 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_rx_wq#13 FD: 25 BD: 1 +.+.: (wq_completion)nfc30_nci_tx_wq#13 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc28_nci_cmd_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_rx_wq#14 FD: 25 BD: 1 +.+.: (wq_completion)nfc28_nci_tx_wq#14 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc27_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc27_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc27_nci_tx_wq#15 FD: 25 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq#18 FD: 1 BD: 2 +.+.: (wq_completion)nfc23_nci_cmd_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_rx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_tx_wq#16 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#19 FD: 25 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#18 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#18 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#19 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#23 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#23 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#23 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#53 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#51 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#50 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#153 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#151 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#399 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#391 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#389 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#860 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#855 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#852 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#861 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#853 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#858 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#854 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#859 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#855 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#860 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#856 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#400 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#392 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#861 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#857 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#393 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#391 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#867 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#156 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#862 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#402 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#154 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#152 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#394 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#392 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#403 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#395 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#393 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#863 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#860 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#404 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#396 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#394 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#52 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#157 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#155 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#51 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#153 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#865 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#405 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#397 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#395 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#158 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#154 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#871 ->&rq->__lock ->&cfs_rq->removed.lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#862 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#873 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#874 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#868 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#863 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#396 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#876 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#869 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#877 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#870 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#865 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#159 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#157 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#155 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#400 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#398 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#874 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#869 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#401 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#399 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#875 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#402 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#400 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#870 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#872 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#873 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#411 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#403 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#401 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#412 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#404 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#402 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#878 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#405 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#403 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#885 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#880 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#887 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#882 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#889 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#883 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#405 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#897 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#890 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#884 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#160 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#158 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#156 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#406 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#901 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#888 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#161 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#159 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#157 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#895 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#889 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#407 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#903 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#896 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#890 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#898 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#892 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#907 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#900 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#894 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#410 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#408 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#162 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#908 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#901 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#895 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#161 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#419 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#158 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#411 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#409 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#411 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#896 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#422 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#414 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#164 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#162 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#159 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#165 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#163 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#415 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#413 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#160 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#53 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#52 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_rx_wq#416 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#424 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#911 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#904 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#425 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#414 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#912 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#899 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#426 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#415 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#166 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#164 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#161 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#53 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#418 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#911 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#905 FD: 29 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#57 ->&rq->__lock ->stock_lock ->&obj_hash[i].lock ->key ->pcpu_lock ->percpu_counters_lock ->&cfs_rq->removed.lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#168 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#166 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#163 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#25 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#55 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#24 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#912 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#54 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#906 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#24 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#908 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#912 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#167 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#164 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#926 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#919 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#913 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#26 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#58 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#56 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#55 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#424 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#421 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#433 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#914 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#425 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#422 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#170 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#168 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#165 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#921 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#916 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#434 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#426 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#423 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#923 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#917 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#918 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#435 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#424 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#932 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#919 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#920 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#436 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#428 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#425 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#171 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#169 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#166 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#934 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#921 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#429 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#426 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#929 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#923 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#924 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#931 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#938 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#939 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#932 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#926 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#927 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#438 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#430 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#427 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#935 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#431 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#929 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#931 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc21_nci_tx_wq#14 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc22_nci_rx_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc22_nci_tx_wq#15 FD: 25 BD: 2 +.+.: (wq_completion)nfc23_nci_cmd_wq#17 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc23_nci_rx_wq#17 ->&rq->__lock FD: 1 BD: 53 +.+.: (wq_completion)phy226 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_tx_wq#17 FD: 25 BD: 2 +.+.: (wq_completion)nfc24_nci_cmd_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_rx_wq#14 FD: 25 BD: 1 +.+.: (wq_completion)nfc24_nci_tx_wq#14 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc25_nci_cmd_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc25_nci_tx_wq#14 FD: 1 BD: 2 +.+.: (wq_completion)nfc27_nci_cmd_wq#16 FD: 25 BD: 1 +.+.: (wq_completion)nfc27_nci_rx_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc27_nci_tx_wq#16 FD: 1 BD: 2 +.+.: (wq_completion)nfc28_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc28_nci_tx_wq#15 FD: 25 BD: 2 +.+.: (wq_completion)nfc29_nci_cmd_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc29_nci_rx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc29_nci_tx_wq#16 FD: 25 BD: 1 +.+.: (wq_completion)nfc42_nci_rx_wq#6 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc42_nci_tx_wq#6 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#445 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#437 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#434 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#957 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#950 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#944 FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#26 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#26 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#21 FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#21 FD: 25 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#21 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc44_nci_cmd_wq#4 FD: 25 BD: 1 +.+.: (wq_completion)nfc44_nci_rx_wq#4 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#175 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#173 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#170 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#61 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#59 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#58 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#21 FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#20 FD: 25 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#20 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#20 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#20 FD: 1 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#20 FD: 1 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#17 FD: 25 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#21 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#21 FD: 1 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#18 FD: 1 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#15 FD: 25 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_rx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_tx_wq#16 FD: 25 BD: 2 +.+.: (wq_completion)nfc16_nci_cmd_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_tx_wq#18 FD: 25 BD: 2 +.+.: (wq_completion)nfc17_nci_cmd_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc17_nci_rx_wq#15 FD: 25 BD: 1 +.+.: (wq_completion)nfc17_nci_tx_wq#15 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#19 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_tx_wq#19 FD: 25 BD: 2 +.+.: (wq_completion)nfc18_nci_cmd_wq#17 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_rx_wq#17 FD: 25 BD: 1 +.+.: (wq_completion)nfc18_nci_tx_wq#17 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#16 FD: 25 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#16 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc20_nci_cmd_wq#17 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_rx_wq#17 FD: 25 BD: 1 +.+.: (wq_completion)nfc20_nci_tx_wq#17 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc21_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_rx_wq#15 FD: 25 BD: 1 +.+.: (wq_completion)nfc21_nci_tx_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc45_nci_rx_wq#4 FD: 1 BD: 1 +.+.: (wq_completion)nfc45_nci_tx_wq#4 FD: 1 BD: 2 +.+.: (wq_completion)nfc43_nci_cmd_wq#5 FD: 1 BD: 1 +.+.: (wq_completion)nfc43_nci_rx_wq#5 FD: 1 BD: 1 +.+.: (wq_completion)nfc43_nci_tx_wq#5 FD: 25 BD: 53 +.+.: (wq_completion)phy225 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock ->pool_lock#2 FD: 1 BD: 2 +.+.: (wq_completion)nfc41_nci_cmd_wq#6 FD: 1 BD: 1 +.+.: (wq_completion)nfc41_nci_rx_wq#6 FD: 1 BD: 1 +.+.: (wq_completion)nfc41_nci_tx_wq#6 FD: 1 BD: 2 +.+.: (wq_completion)nfc40_nci_cmd_wq#8 FD: 25 BD: 1 +.+.: (wq_completion)nfc40_nci_rx_wq#8 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc35_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_rx_wq#14 FD: 25 BD: 1 +.+.: (wq_completion)nfc35_nci_tx_wq#14 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc39_nci_cmd_wq#7 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc39_nci_rx_wq#7 FD: 1 BD: 1 +.+.: (wq_completion)nfc39_nci_tx_wq#7 FD: 25 BD: 2 +.+.: (wq_completion)nfc38_nci_cmd_wq#7 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc38_nci_rx_wq#7 FD: 25 BD: 1 +.+.: (wq_completion)nfc38_nci_tx_wq#7 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc37_nci_cmd_wq#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_rx_wq#8 FD: 1 BD: 1 +.+.: (wq_completion)nfc37_nci_tx_wq#8 FD: 1 BD: 2 +.+.: (wq_completion)nfc36_nci_cmd_wq#9 FD: 25 BD: 1 +.+.: (wq_completion)nfc36_nci_rx_wq#8 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc36_nci_tx_wq#8 FD: 1 BD: 2 +.+.: (wq_completion)nfc34_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_rx_wq#13 FD: 25 BD: 1 +.+.: (wq_completion)nfc34_nci_tx_wq#13 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc33_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_tx_wq#15 FD: 1 BD: 2 +.+.: (wq_completion)nfc32_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_rx_wq#13 FD: 1 BD: 2 +.+.: (wq_completion)nfc31_nci_cmd_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc31_nci_rx_wq#13 FD: 1 BD: 1 +.+.: (wq_completion)nfc31_nci_tx_wq#13 FD: 25 BD: 2 +.+.: (wq_completion)nfc30_nci_cmd_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_tx_wq#14 FD: 1 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq#20 FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq#20 FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq#20 FD: 1 BD: 2 +.+.: (wq_completion)nfc22_nci_cmd_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc22_nci_rx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc22_nci_tx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#947 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#441 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#980 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#974 FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg1#211 ->(work_completion)(&peer->transmit_handshake_work) FD: 53 BD: 1 +.+.: (wq_completion)wg-kex-wg2#211 ->(work_completion)(&peer->transmit_handshake_work) FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg0#214 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg1#212 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 87 BD: 1 +.+.: (wq_completion)wg-kex-wg2#212 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) FD: 88 BD: 1 +.+.: (wq_completion)wg-crypt-wg0#109 ->(work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) ->(work_completion)(&peer->transmit_packet_work) FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#988 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#981 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#975 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#445 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#982 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#976 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#454 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#446 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#443 FD: 43 BD: 1 +.+.: (wq_completion)bond0#117 ->(work_completion)(&(&slave->notify_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#447 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#444 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#984 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#978 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#986 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#980 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#456 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#445 FD: 1 BD: 1 +...: rds_sock_lock FD: 1 BD: 1 +...: clock-AF_RDS FD: 1 BD: 1 ....: &rs->rs_recv_lock FD: 1 BD: 1 ....: rds_cong_monitor_lock FD: 1 BD: 1 ....: rds_cong_lock FD: 1 BD: 1 ....: &rs->rs_lock FD: 1 BD: 1 ....: &q->lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#988 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#982 ->&rq->__lock FD: 166 BD: 1 +.+.: (wq_completion)hci3#8 ->(work_completion)(&hdev->cmd_work) ->(work_completion)(&hdev->rx_work) ->(work_completion)(&hdev->tx_work) ->(work_completion)(&conn->pending_rx_work) ->(work_completion)(&(&hdev->cmd_timer)->work) ->(work_completion)(&(&conn->disc_work)->work) FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#175 FD: 136 BD: 50 +.+.: team->team_lock_key#118 ->fs_reclaim ->netpoll_srcu ->net_rwsem ->&tn->lock ->_xmit_ETHER ->&c->lock ->&dir->lock#2 ->input_pool.lock ->&ndev->lock ->&obj_hash[i].lock ->nl_table_lock ->nl_table_wait.lock ->&rq->__lock ->&in_dev->mc_tomb_lock ->&im->lock ->cbs_list_lock ->sysfs_symlink_target_lock ->lock ->&root->kernfs_rwsem ->&____s->seqcount#2 ->&____s->seqcount ->&n->list_lock ->lweventlist_lock ->(console_sem).lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#995 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#989 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#448 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#456 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#997 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#469 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#461 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#457 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#998 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#470 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#462 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#999 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#471 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#463 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#459 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#184 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#182 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#464 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#460 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#183 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#180 FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#29 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#27 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#63 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#61 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#60 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1025 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#476 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#190 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#188 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#468 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#185 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#477 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#469 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#470 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#466 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1029 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1019 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1013 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#467 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1033 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1017 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#480 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#468 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1018 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1025 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1019 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#481 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#473 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#469 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1020 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#474 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#470 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1037 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1028 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1021 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1039 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1022 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1040 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1030 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1031 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1024 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#483 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#475 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1042 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#484 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1032 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#476 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1025 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#472 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1026 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#485 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#189 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#186 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#473 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1044 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1034 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1027 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#66 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#64 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#63 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#30 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#486 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#478 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#474 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1045 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1035 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1028 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#191 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#67 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#188 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#65 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#64 FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#31 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#28 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#28 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1036 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1029 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1037 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1030 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#479 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#475 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1048 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1038 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1031 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#480 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#194 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#476 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#192 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1039 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1032 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#489 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#481 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#477 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1033 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1051 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1034 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#482 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#478 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#193 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#190 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1052 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1042 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1035 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#491 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#483 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#196 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#194 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#191 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1043 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1036 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#492 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#484 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#480 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1044 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1037 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#195 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#192 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1055 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_rx_wq#21 FD: 25 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#22 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#493 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#485 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#481 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#68 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#66 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#65 FD: 25 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#22 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1038 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc7_nci_rx_wq#22 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#22 FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#32 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#198 FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#29 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#29 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#196 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#193 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1056 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1046 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1047 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1040 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1048 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#199 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1041 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#197 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#194 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#482 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#69 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#67 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#66 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#200 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1059 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#198 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#195 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1042 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#495 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#487 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#483 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#70 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#68 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#67 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#33 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#30 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#30 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#201 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#199 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#196 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1050 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1043 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#202 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#497 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#200 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#489 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#485 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#197 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#490 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#486 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1051 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1044 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#68 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1062 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1052 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#203 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1045 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#201 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#198 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#72 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#70 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#69 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#491 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#487 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#492 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#488 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#204 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1053 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#202 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#199 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1046 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1054 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1047 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#493 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#489 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#205 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1065 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1055 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1048 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#502 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#494 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#490 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1066 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1056 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1067 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1057 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1050 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#503 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#495 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#491 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1058 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1051 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#504 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#492 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1069 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1059 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1052 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1070 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1060 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1053 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#505 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#493 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#206 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#204 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#201 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1061 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1054 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#506 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1072 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#498 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#494 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1062 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#207 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#205 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#202 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1056 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#507 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#499 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#495 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#508 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#500 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#496 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1064 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1057 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#497 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#510 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1065 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1058 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#511 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#503 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#498 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1076 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1066 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1059 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1077 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1060 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1068 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1061 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#499 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#513 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#505 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#500 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1069 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1062 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#208 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#206 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#203 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#514 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#506 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#501 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1071 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1081 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#207 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#204 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#502 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1082 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1072 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1064 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#210 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#208 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#73 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#71 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#205 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1083 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1073 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1065 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1074 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1066 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#503 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1085 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#211 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1075 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#209 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#206 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1067 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1086 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#517 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1077 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1068 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#509 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#504 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1078 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1069 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#518 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#505 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#212 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#210 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#207 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1079 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1070 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#519 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#511 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#506 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1090 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1080 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1071 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#520 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1072 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#512 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#507 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1082 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1073 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#513 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#508 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1093 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1083 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1074 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1075 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#74 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#72 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#71 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#522 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#514 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#509 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#213 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#211 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#208 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#510 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#214 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#212 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#209 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1095 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1085 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1076 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#524 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#516 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1096 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#511 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1086 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#215 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#213 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#210 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#75 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#73 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#72 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1087 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1078 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#525 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#517 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#512 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1079 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#526 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#518 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#513 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#216 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#214 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#211 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1089 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1080 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1100 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1090 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1081 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#527 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#514 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1091 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1082 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#528 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#520 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#515 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1083 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#521 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#516 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#217 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#215 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#212 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1084 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1094 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1104 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1105 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#530 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1095 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#522 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1086 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#517 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1106 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1096 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1087 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#518 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#218 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#213 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#76 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#74 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#73 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1107 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1097 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1088 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#532 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#524 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#219 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#217 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1108 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1098 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1089 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#214 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1099 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1090 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#525 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#520 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1111 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1101 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1092 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#220 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#215 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#534 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#526 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#521 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#221 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1112 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#219 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1102 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1093 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#535 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#527 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1113 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1103 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1094 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1114 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1104 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#528 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#523 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#220 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#217 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#537 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1115 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1105 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1096 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#529 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#524 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#77 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#75 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#74 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1116 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#34 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1106 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#31 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1097 FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#31 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#223 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#221 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#538 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#530 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1118 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#224 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#222 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#219 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1108 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1099 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1119 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1109 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1100 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#526 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#225 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#223 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#220 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1120 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#541 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1110 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#532 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1101 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#527 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#226 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#78 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#76 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#75 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#221 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#542 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1102 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#533 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#529 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#227 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1122 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#222 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1112 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1103 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#79 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#77 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#76 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#544 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#535 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#530 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#32 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#32 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1123 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1113 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1104 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1105 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1115 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1125 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#545 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#531 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1126 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1116 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1106 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1117 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1107 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1118 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#537 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1108 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#532 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1119 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1109 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#547 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#538 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#533 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#80 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#77 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#226 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#223 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1130 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1120 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1110 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#539 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#534 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1131 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1121 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1111 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#540 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#535 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1122 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1112 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1123 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1113 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#550 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#536 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1134 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#551 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1124 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#542 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#537 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1114 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1125 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1115 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1116 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1137 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1127 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1117 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#552 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#538 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_tx_wq#78 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_rx_wq#79 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#81 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#229 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#224 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1128 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1118 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#553 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#544 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#539 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#230 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#228 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#225 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1139 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1129 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1119 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1130 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1120 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1121 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#554 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#545 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#540 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1143 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1132 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1133 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1123 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1124 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1146 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1135 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1125 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#231 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#229 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#82 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#555 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#80 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#546 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#79 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#541 FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#36 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#33 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#33 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1126 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#556 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#547 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#542 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#232 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1148 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1137 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1127 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#548 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#543 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1138 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1128 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#558 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1150 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#549 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#544 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1139 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1129 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#545 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#233 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#230 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#227 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1141 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1131 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#546 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#83 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1153 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1142 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1132 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#234 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#81 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#231 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#228 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#80 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#552 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#547 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1154 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#84 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#82 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1143 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1133 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#81 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#562 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#553 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#548 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#232 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#229 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1134 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#563 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#554 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#549 FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#85 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#236 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#233 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#230 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1145 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1135 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#555 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#550 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1157 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#237 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#234 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1136 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#231 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1147 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1137 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#566 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#556 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#37 FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#34 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#551 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#34 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#235 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#232 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_rx_wq#1148 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1159 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#83 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#82 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#557 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#552 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#239 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#236 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#233 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1149 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1138 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#84 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#83 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#568 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1161 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1150 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#558 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1139 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#553 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#237 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#234 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1151 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1140 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1152 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1141 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#559 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#554 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1164 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1153 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#88 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#85 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1142 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#241 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#238 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#571 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1143 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#556 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#572 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1155 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1144 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#573 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#242 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#239 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#236 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#562 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#38 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#35 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#35 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#89 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#86 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#85 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1156 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1145 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#574 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#563 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#558 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1157 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1146 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_rx_wq#564 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#575 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1170 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#576 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#577 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1158 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#565 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#560 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1147 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1160 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1149 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1173 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1161 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1150 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#566 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#561 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1151 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#243 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#579 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#240 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#237 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#567 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#562 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1152 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#580 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#568 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#563 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#244 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#241 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#238 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1176 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1153 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#569 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1154 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#564 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1166 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#582 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#570 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#565 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#245 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#242 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#239 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#583 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#243 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#240 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#571 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1179 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1167 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1156 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#90 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#87 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#86 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1168 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1157 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#584 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1181 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1169 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1158 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1170 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#244 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#241 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1159 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#585 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#568 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1183 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#248 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#245 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#242 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1160 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#574 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#569 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1161 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1173 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#575 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#570 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1162 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#249 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#246 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#243 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1186 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1174 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1163 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#588 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#576 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#571 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1175 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1164 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1165 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#589 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#577 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#572 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1177 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1166 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1190 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1178 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1167 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#578 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#573 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1191 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1179 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1168 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#574 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1192 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#592 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1180 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#580 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#575 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1169 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1181 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1170 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1182 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1171 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1183 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1172 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1184 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1173 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#593 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#581 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#576 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1185 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#594 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#582 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#577 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#250 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#244 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1186 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1175 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#595 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#583 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#578 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1199 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1187 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1176 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#584 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#579 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#585 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#580 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#251 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#248 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#245 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#599 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#586 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1200 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#581 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1188 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1177 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1178 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#600 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#587 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#582 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1202 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1179 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#601 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#583 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1203 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1191 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1180 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#602 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#584 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1192 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1181 FD: 1 BD: 2 +.+.: virtio_fs_mutex FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#252 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#249 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#246 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1193 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1182 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1194 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1183 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#603 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#590 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#585 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#21 FD: 25 BD: 1 +.+.: (wq_completion)nfc9_nci_tx_wq#21 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_tx_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#17 FD: 1 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#18 FD: 25 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_tx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#19 FD: 25 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#19 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc33_nci_cmd_wq#16 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc33_nci_rx_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc33_nci_tx_wq#16 FD: 25 BD: 2 +.+.: (wq_completion)nfc32_nci_cmd_wq#14 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_rx_wq#14 FD: 1 BD: 1 +.+.: (wq_completion)nfc32_nci_tx_wq#14 FD: 25 BD: 2 +.+.: (wq_completion)nfc31_nci_cmd_wq#14 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc31_nci_rx_wq#14 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc35_nci_cmd_wq#16 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc31_nci_tx_wq#14 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc30_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc30_nci_rx_wq#15 FD: 25 BD: 1 +.+.: (wq_completion)nfc30_nci_tx_wq#15 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#20 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 2 +.+.: (wq_completion)nfc29_nci_cmd_wq#17 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc29_nci_rx_wq#17 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc29_nci_tx_wq#17 FD: 25 BD: 2 +.+.: (wq_completion)nfc28_nci_cmd_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_rx_wq#15 FD: 25 BD: 1 +.+.: (wq_completion)nfc28_nci_rx_wq#16 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc28_nci_tx_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc35_nci_tx_wq#15 FD: 25 BD: 2 +.+.: (wq_completion)nfc27_nci_cmd_wq#17 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc27_nci_rx_wq#17 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc27_nci_tx_wq#17 FD: 25 BD: 2 +.+.: (wq_completion)nfc34_nci_cmd_wq#15 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc26_nci_cmd_wq#21 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_rx_wq#21 FD: 1 BD: 1 +.+.: (wq_completion)nfc26_nci_tx_wq#21 FD: 25 BD: 1 +.+.: (wq_completion)nfc34_nci_rx_wq#15 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc25_nci_cmd_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc25_nci_tx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc34_nci_tx_wq#15 FD: 25 BD: 2 +.+.: (wq_completion)nfc24_nci_cmd_wq#15 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_rx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc24_nci_tx_wq#15 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#20 FD: 1 BD: 2 +.+.: (wq_completion)nfc23_nci_cmd_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_rx_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc23_nci_tx_wq#18 FD: 25 BD: 2 +.+.: (wq_completion)nfc21_nci_cmd_wq#17 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc21_nci_rx_wq#17 FD: 25 BD: 1 +.+.: (wq_completion)nfc21_nci_tx_wq#17 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc22_nci_cmd_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc22_nci_rx_wq#17 FD: 25 BD: 1 +.+.: (wq_completion)nfc22_nci_tx_wq#17 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc20_nci_rx_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc20_nci_tx_wq#18 FD: 25 BD: 2 +.+.: (wq_completion)nfc18_nci_cmd_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_rx_wq#18 FD: 1 BD: 1 +.+.: (wq_completion)nfc18_nci_tx_wq#18 FD: 1 BD: 2 +.+.: (wq_completion)nfc17_nci_cmd_wq#16 FD: 1 BD: 1 +.+.: (wq_completion)nfc17_nci_rx_wq#16 FD: 25 BD: 1 +.+.: (wq_completion)nfc17_nci_tx_wq#16 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc19_nci_cmd_wq#21 FD: 1 BD: 2 +.+.: (wq_completion)nfc16_nci_cmd_wq#19 FD: 25 BD: 1 +.+.: (wq_completion)nfc16_nci_rx_wq#19 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc16_nci_tx_wq#19 FD: 25 BD: 1 +.+.: (wq_completion)nfc19_nci_rx_wq#21 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq#17 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_rx_wq#17 FD: 1 BD: 1 +.+.: (wq_completion)nfc19_nci_tx_wq#21 FD: 25 BD: 2 +.+.: (wq_completion)nfc14_nci_cmd_wq#20 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc14_nci_rx_wq#20 FD: 25 BD: 2 +.+.: (wq_completion)nfc15_nci_cmd_wq#18 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc15_nci_rx_wq#18 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc15_nci_tx_wq#18 FD: 25 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#22 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc9_nci_rx_wq#22 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc13_nci_cmd_wq#16 FD: 25 BD: 1 +.+.: (wq_completion)nfc13_nci_rx_wq#16 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc13_nci_tx_wq#16 FD: 1 BD: 2 +.+.: (wq_completion)nfc12_nci_cmd_wq#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_rx_wq#19 FD: 1 BD: 1 +.+.: (wq_completion)nfc12_nci_tx_wq#19 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#24 FD: 25 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#22 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc9_nci_rx_wq#23 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 2 +.+.: (wq_completion)nfc9_nci_cmd_wq#23 FD: 25 BD: 2 +.+.: (wq_completion)nfc11_nci_cmd_wq#22 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc11_nci_rx_wq#22 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc11_nci_tx_wq#22 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc10_nci_cmd_wq#18 FD: 1 BD: 2 +.+.: (wq_completion)nfc8_nci_cmd_wq#25 FD: 1 BD: 1 +.+.: (wq_completion)nfc10_nci_rx_wq#18 FD: 25 BD: 1 +.+.: (wq_completion)nfc10_nci_tx_wq#18 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc7_nci_cmd_wq#23 FD: 25 BD: 1 +.+.: (wq_completion)nfc8_nci_rx_wq#24 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc7_nci_tx_wq#23 FD: 1 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#39 FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#36 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc8_nci_tx_wq#23 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#36 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#91 FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#88 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 25 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#87 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#253 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#250 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#247 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#604 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#591 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#586 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1207 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1195 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1184 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1196 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1185 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#92 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1209 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#254 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#89 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1197 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#88 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1186 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#248 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#605 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#592 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#587 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1198 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#606 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1187 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#593 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#255 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1211 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#249 FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1212 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1199 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1200 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1189 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1190 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#607 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#589 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1202 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#256 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#250 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#608 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1216 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#595 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1203 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1192 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#590 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#254 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#251 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1204 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1193 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#596 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#591 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#597 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#592 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1205 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc6_nci_cmd_wq#40 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1194 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_rx_wq#37 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#258 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc6_nci_tx_wq#37 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#255 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#252 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#611 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#593 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#259 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1219 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1206 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1195 FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#93 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#90 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#89 FD: 1 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#613 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#594 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#600 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#595 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1220 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1207 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1196 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#261 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#253 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1208 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1197 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#601 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#596 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#262 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#258 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#254 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1222 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1209 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1198 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1199 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#602 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#597 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#263 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#617 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#603 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#598 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#255 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1211 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#264 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#260 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#94 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#256 FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#91 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1225 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_tx_wq#90 FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#618 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1212 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1201 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#604 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#599 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1226 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1213 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1202 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1203 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1228 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#265 FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#266 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1204 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#619 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#605 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#600 FD: 25 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#267 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#261 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc3_nci_cmd_wq#620 ->&rq->__lock ->&cfs_rq->removed.lock ->&obj_hash[i].lock FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#257 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1229 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1216 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1205 FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#601 FD: 25 BD: 2 +.+.: (wq_completion)nfc2_nci_cmd_wq#1230 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1217 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1206 FD: 1 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1207 FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_rx_wq#1219 ->&rq->__lock FD: 25 BD: 1 +.+.: (wq_completion)nfc2_nci_tx_wq#1208 ->&rq->__lock FD: 1 BD: 2 +.+.: (wq_completion)nfc4_nci_cmd_wq#268 FD: 1 BD: 1 +.+.: (wq_completion)nfc4_nci_rx_wq#262 FD: 25 BD: 1 +.+.: (wq_completion)nfc4_nci_tx_wq#258 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc3_nci_rx_wq#607 FD: 25 BD: 1 +.+.: (wq_completion)nfc3_nci_tx_wq#602 ->&rq->__lock FD: 25 BD: 2 +.+.: (wq_completion)nfc5_nci_cmd_wq#95 ->&rq->__lock FD: 1 BD: 1 +.+.: (wq_completion)nfc5_nci_rx_wq#92 all lock chains: irq_context: 0 (console_sem).lock irq_context: 0 &obj_hash[i].lock irq_context: 0 &obj_hash[i].lock pool_lock irq_context: 0 cgroup_mutex irq_context: 0 fixmap_lock irq_context: 0 cpu_hotplug_lock irq_context: 0 cpu_hotplug_lock jump_label_mutex irq_context: 0 console_mutex irq_context: 0 input_pool.lock irq_context: 0 base_crng.lock irq_context: 0 cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 crng_init_wait.lock irq_context: 0 early_pfn_lock irq_context: 0 devtree_lock irq_context: 0 resource_lock irq_context: 0 restart_handler_list.lock irq_context: 0 system_transition_mutex irq_context: 0 pcpu_lock irq_context: 0 debug_hook_lock irq_context: 0 zonelist_update_seq irq_context: 0 zonelist_update_seq zonelist_update_seq.seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 &zone->lock irq_context: 0 &zone->lock &____s->seqcount irq_context: 0 &pcp->lock &zone->lock irq_context: 0 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &____s->seqcount irq_context: 0 pool_lock#2 irq_context: 0 pcpu_alloc_mutex irq_context: 0 pcpu_alloc_mutex pcpu_lock irq_context: 0 &n->list_lock irq_context: 0 &c->lock irq_context: 0 slab_mutex irq_context: 0 slab_mutex pool_lock#2 irq_context: 0 slab_mutex pcpu_alloc_mutex irq_context: 0 slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 trace_types_lock irq_context: 0 panic_notifier_list.lock irq_context: 0 die_chain.lock irq_context: 0 trace_event_sem irq_context: 0 batched_entropy_u32.lock irq_context: 0 batched_entropy_u32.lock crngs.lock irq_context: 0 batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 sysctl_lock irq_context: 0 &rq->__lock irq_context: 0 &rq->__lock rcu_read_lock &cfs_b->lock irq_context: 0 init_task.pi_lock irq_context: 0 init_task.pi_lock &rq->__lock irq_context: 0 init_task.vtime_seqcount irq_context: 0 slab_mutex &c->lock irq_context: 0 slab_mutex &pcp->lock &zone->lock irq_context: 0 slab_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 slab_mutex &____s->seqcount irq_context: 0 wq_pool_mutex irq_context: 0 wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 wq_pool_mutex &____s->seqcount irq_context: 0 wq_pool_mutex &c->lock irq_context: 0 wq_pool_mutex pool_lock#2 irq_context: 0 &wq->mutex irq_context: 0 &wq->mutex &pool->lock irq_context: 0 wq_pool_mutex &wq->mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 shrinker_mutex irq_context: 0 rcu_node_0 irq_context: 0 rcu_state.barrier_lock irq_context: 0 rcu_state.barrier_lock rcu_node_0 irq_context: 0 &rnp->exp_poll_lock irq_context: 0 &rnp->exp_poll_lock rcu_read_lock &pool->lock irq_context: 0 &rnp->exp_poll_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 trace_event_sem trace_event_ida.xa_lock irq_context: 0 trace_event_sem trace_event_ida.xa_lock &pcp->lock &zone->lock irq_context: 0 trace_event_sem trace_event_ida.xa_lock &____s->seqcount irq_context: 0 trace_event_sem trace_event_ida.xa_lock &c->lock irq_context: 0 trace_event_sem trace_event_ida.xa_lock pool_lock#2 irq_context: 0 trigger_cmd_mutex irq_context: 0 free_vmap_area_lock irq_context: 0 vmap_area_lock irq_context: 0 acpi_probe_mutex irq_context: 0 acpi_probe_mutex pool_lock#2 irq_context: 0 acpi_probe_mutex free_vmap_area_lock irq_context: 0 acpi_probe_mutex vmap_area_lock irq_context: 0 acpi_probe_mutex &pcp->lock &zone->lock irq_context: 0 acpi_probe_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 acpi_probe_mutex &____s->seqcount irq_context: 0 acpi_probe_mutex init_mm.page_table_lock irq_context: 0 acpi_probe_mutex resource_lock irq_context: 0 acpi_probe_mutex &c->lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 acpi_probe_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 acpi_probe_mutex (console_sem).lock irq_context: 0 acpi_probe_mutex irq_domain_mutex irq_context: 0 acpi_probe_mutex pcpu_alloc_mutex irq_context: 0 acpi_probe_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 acpi_probe_mutex irq_domain_mutex pool_lock#2 irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 acpi_probe_mutex &domain->mutex irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &pcp->lock &zone->lock irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &____s->seqcount irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &c->lock irq_context: 0 acpi_probe_mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock pool_lock irq_context: 0 acpi_probe_mutex &domain->mutex pool_lock#2 irq_context: 0 acpi_probe_mutex &domain->mutex &irq_desc_lock_class irq_context: 0 acpi_probe_mutex &desc->request_mutex irq_context: 0 acpi_probe_mutex &desc->request_mutex &irq_desc_lock_class irq_context: 0 acpi_probe_mutex &irq_desc_lock_class irq_context: 0 acpi_probe_mutex cpu_pm_notifier.lock irq_context: 0 acpi_probe_mutex &obj_hash[i].lock irq_context: 0 acpi_probe_mutex purge_vmap_area_lock irq_context: 0 acpi_probe_mutex iort_msi_chip_lock irq_context: 0 acpi_probe_mutex &zone->lock irq_context: 0 acpi_probe_mutex &zone->lock &____s->seqcount irq_context: 0 acpi_probe_mutex its_lock irq_context: 0 acpi_probe_mutex resource_lock irq_context: 0 acpi_probe_mutex efi_mem_reserve_persistent_lock irq_context: 0 acpi_probe_mutex lpi_range_lock irq_context: 0 acpi_probe_mutex syscore_ops_lock irq_context: 0 acpi_probe_mutex its_lock &its->lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-down irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-up irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex resource_lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex pool_lock#2 irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex resource_lock irq_context: 0 timekeeper_lock irq_context: 0 timekeeper_lock tk_core.seq.seqcount irq_context: 0 timekeeper_lock tk_core.seq.seqcount &obj_hash[i].lock irq_context: 0 acpi_probe_mutex &desc->request_mutex &irq_desc_lock_class irq_controller_lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex (console_sem).lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex clockevents_lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex clockevents_lock tk_core.seq.seqcount irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex clockevents_lock tick_broadcast_lock irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex clockevents_lock jiffies_seq.seqcount irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex &irq_desc_lock_class irq_context: 0 acpi_probe_mutex cpu_hotplug_lock cpuhp_state_mutex &irq_desc_lock_class irq_controller_lock irq_context: 0 acpi_probe_mutex clocksource_mutex irq_context: 0 clockevents_lock irq_context: 0 tk_core.seq.seqcount irq_context: 0 &base->lock irq_context: 0 &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &pool->lock irq_context: 0 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 batched_entropy_u64.lock irq_context: 0 batched_entropy_u64.lock crngs.lock irq_context: 0 pmus_lock irq_context: 0 pmus_lock pcpu_alloc_mutex irq_context: 0 pmus_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 pmus_lock pool_lock#2 irq_context: 0 pmus_lock &obj_hash[i].lock irq_context: 0 &swhash->hlist_mutex irq_context: 0 pmus_lock &cpuctx_mutex irq_context: 0 pmus_lock &obj_hash[i].lock pool_lock irq_context: 0 tty_ldiscs_lock irq_context: 0 console_lock irq_context: 0 console_lock pool_lock#2 irq_context: 0 console_lock &obj_hash[i].lock irq_context: 0 console_lock &pcp->lock &zone->lock irq_context: 0 console_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 console_lock &____s->seqcount irq_context: 0 console_lock &c->lock irq_context: 0 console_lock kbd_event_lock irq_context: 0 console_lock kbd_event_lock led_lock irq_context: 0 console_lock (console_sem).lock irq_context: 0 console_lock console_owner_lock irq_context: 0 init_task.alloc_lock irq_context: 0 acpi_ioremap_lock irq_context: 0 acpi_ioremap_lock pool_lock#2 irq_context: 0 semaphore->lock irq_context: 0 *(&acpi_gbl_reference_count_lock) irq_context: 0 hrtimer_bases.lock irq_context: 0 hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 percpu_counters_lock irq_context: 0 tomoyo_policy_lock irq_context: 0 tomoyo_policy_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex irq_context: 0 pernet_ops_rwsem crngs.lock irq_context: 0 pernet_ops_rwsem net_rwsem irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock irq_context: 0 rtnl_mutex irq_context: 0 rtnl_mutex &c->lock irq_context: 0 rtnl_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &____s->seqcount irq_context: 0 rtnl_mutex pool_lock#2 irq_context: 0 lock irq_context: 0 lock kernfs_idr_lock irq_context: 0 lock kernfs_idr_lock pool_lock#2 irq_context: 0 &root->kernfs_rwsem irq_context: 0 file_systems_lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 dq_list_lock irq_context: 0 sb_lock irq_context: 0 &type->s_umount_key/1 irq_context: 0 &type->s_umount_key/1 pool_lock#2 irq_context: 0 &type->s_umount_key/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key/1 shrinker_mutex irq_context: 0 &type->s_umount_key/1 shrinker_mutex pool_lock#2 irq_context: 0 &type->s_umount_key/1 list_lrus_mutex irq_context: 0 &type->s_umount_key/1 sb_lock irq_context: 0 &type->s_umount_key/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key/1 &c->lock irq_context: 0 &type->s_umount_key/1 &____s->seqcount irq_context: 0 &type->s_umount_key/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key/1 percpu_counters_lock irq_context: 0 &type->s_umount_key/1 crngs.lock irq_context: 0 &type->s_umount_key/1 &sbinfo->stat_lock irq_context: 0 &type->s_umount_key/1 &sb->s_type->i_lock_key irq_context: 0 &type->s_umount_key/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key/1 batched_entropy_u32.lock irq_context: 0 &type->s_umount_key/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key/1 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &type->s_umount_key/1 &dentry->d_lock irq_context: 0 mnt_id_ida.xa_lock irq_context: 0 &dentry->d_lock irq_context: 0 mount_lock irq_context: 0 mount_lock mount_lock.seqcount irq_context: 0 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#2/1 irq_context: 0 &type->s_umount_key#2/1 pool_lock#2 irq_context: 0 &type->s_umount_key#2/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#2/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#2/1 shrinker_mutex irq_context: 0 &type->s_umount_key#2/1 &c->lock irq_context: 0 &type->s_umount_key#2/1 &____s->seqcount irq_context: 0 &type->s_umount_key#2/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#2/1 sb_lock irq_context: 0 &type->s_umount_key#2/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#2/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#2/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#2/1 &sb->s_type->i_lock_key#2 irq_context: 0 &type->s_umount_key#2/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#2/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#2/1 &sb->s_type->i_lock_key#2 &dentry->d_lock irq_context: 0 &type->s_umount_key#2/1 &dentry->d_lock irq_context: 0 ucounts_lock irq_context: 0 proc_inum_ida.xa_lock irq_context: 0 init_fs.lock irq_context: 0 init_fs.lock init_fs.seq.seqcount irq_context: hardirq jiffies_lock irq_context: hardirq jiffies_lock jiffies_seq.seqcount irq_context: hardirq log_wait.lock irq_context: 0 &type->s_umount_key#3/1 irq_context: 0 &type->s_umount_key#3/1 pool_lock#2 irq_context: 0 &type->s_umount_key#3/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#3/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#3/1 shrinker_mutex irq_context: 0 &type->s_umount_key#3/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#3/1 sb_lock irq_context: 0 &type->s_umount_key#3/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#3/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#3/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#3/1 &____s->seqcount irq_context: 0 &type->s_umount_key#3/1 &c->lock irq_context: 0 &type->s_umount_key#3/1 &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#3/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#3/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#3/1 &sb->s_type->i_lock_key#3 &dentry->d_lock irq_context: 0 &type->s_umount_key#3/1 &dentry->d_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-down irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-up irq_context: 0 proc_subdir_lock irq_context: 0 proc_subdir_lock irq_context: 0 pernet_ops_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 &type->s_umount_key#4/1 irq_context: 0 &type->s_umount_key#4/1 pool_lock#2 irq_context: 0 &type->s_umount_key#4/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#4/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#4/1 shrinker_mutex irq_context: 0 &type->s_umount_key#4/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#4/1 sb_lock irq_context: 0 &type->s_umount_key#4/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#4/1 &sb->s_type->i_lock_key#4 irq_context: 0 &type->s_umount_key#4/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#4/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#4/1 &sb->s_type->i_lock_key#4 &dentry->d_lock irq_context: 0 &type->s_umount_key#4/1 &dentry->d_lock irq_context: 0 cgroup_mutex pcpu_alloc_mutex irq_context: 0 cgroup_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 cgroup_mutex &c->lock irq_context: 0 cgroup_mutex &____s->seqcount irq_context: 0 cgroup_mutex pool_lock#2 irq_context: 0 cgroup_mutex lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 cgroup_mutex &root->kernfs_rwsem irq_context: 0 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cgroup_mutex &obj_hash[i].lock irq_context: 0 cgroup_mutex cgroup_file_kn_lock irq_context: 0 cgroup_mutex &obj_hash[i].lock pool_lock irq_context: 0 cgroup_mutex css_set_lock irq_context: 0 lock cgroup_idr_lock irq_context: 0 lock cgroup_idr_lock pool_lock#2 irq_context: 0 cpuset_mutex irq_context: 0 cpuset_mutex callback_lock irq_context: 0 cgroup_mutex blkcg_pol_mutex irq_context: 0 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex irq_context: 0 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 cgroup_mutex lock cgroup_idr_lock irq_context: 0 cgroup_mutex lock cgroup_idr_lock pool_lock#2 irq_context: 0 cgroup_mutex percpu_counters_lock irq_context: 0 cgroup_mutex shrinker_mutex irq_context: 0 cgroup_mutex shrinker_mutex pool_lock#2 irq_context: 0 cgroup_mutex &base->lock irq_context: 0 cgroup_mutex &base->lock &obj_hash[i].lock irq_context: 0 cgroup_mutex devcgroup_mutex irq_context: 0 cgroup_mutex cpu_hotplug_lock irq_context: 0 cgroup_mutex cpu_hotplug_lock freezer_mutex irq_context: 0 rcu_state.exp_mutex rcu_node_0 irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[0] irq_context: 0 rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[1] irq_context: 0 init_sighand.siglock irq_context: 0 init_mm.page_table_lock irq_context: 0 init_files.file_lock irq_context: 0 rcu_read_lock init_sighand.siglock irq_context: 0 lock pidmap_lock irq_context: 0 lock pidmap_lock pool_lock#2 irq_context: 0 pidmap_lock irq_context: 0 cgroup_threadgroup_rwsem irq_context: 0 cgroup_threadgroup_rwsem css_set_lock irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem tk_core.seq.seqcount irq_context: 0 cgroup_threadgroup_rwsem tasklist_lock irq_context: 0 cgroup_threadgroup_rwsem tasklist_lock init_sighand.siglock irq_context: 0 cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock irq_context: 0 cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 &p->pi_lock irq_context: 0 &p->pi_lock &rq->__lock irq_context: 0 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &p->pi_lock &rq->__lock &base->lock irq_context: 0 &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &p->pi_lock irq_context: 0 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (kthreadd_done).wait.lock irq_context: 0 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock irq_context: 0 &p->alloc_lock irq_context: 0 &p->alloc_lock &____s->seqcount#2 irq_context: 0 fs_reclaim irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kthread_create_lock irq_context: 0 &x->wait irq_context: 0 rcu_read_lock &sighand->siglock irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem tasklist_lock &sighand->siglock irq_context: 0 &x->wait &p->pi_lock irq_context: 0 &x->wait &p->pi_lock &rq->__lock irq_context: 0 &x->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (worker)->lock irq_context: 0 wq_pool_mutex fs_reclaim irq_context: 0 wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 wq_pool_mutex kthread_create_lock irq_context: 0 wq_pool_mutex &p->pi_lock irq_context: 0 wq_pool_mutex &p->pi_lock &rq->__lock irq_context: 0 wq_pool_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wq_pool_mutex &x->wait irq_context: 0 wq_pool_mutex &rq->__lock irq_context: 0 wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wq_pool_mutex &obj_hash[i].lock irq_context: 0 wq_pool_attach_mutex irq_context: 0 wq_mayday_lock irq_context: 0 &xa->xa_lock irq_context: 0 &pool->lock irq_context: 0 &pool->lock &p->pi_lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (&pool->mayday_timer) irq_context: 0 &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rnp->exp_poll_wq) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rnp->exp_poll_wq) &rnp->exp_poll_lock irq_context: 0 (null) irq_context: 0 (null) tk_core.seq.seqcount irq_context: 0 (&wq_watchdog_timer) irq_context: 0 (wq_completion)events_unbound irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) allocation_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq allocation_wait.lock irq_context: hardirq allocation_wait.lock &p->pi_lock irq_context: hardirq allocation_wait.lock &p->pi_lock &rq->__lock irq_context: hardirq allocation_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 batched_entropy_u8.lock irq_context: 0 batched_entropy_u8.lock crngs.lock irq_context: 0 kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &base->lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex irq_context: 0 rcu_tasks.cbs_gbl_lock irq_context: hardirq rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &x->wait#2 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex irq_context: 0 rcu_tasks_trace.cbs_gbl_lock irq_context: 0 rcu_tasks__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[2] irq_context: 0 rcu_tasks.tasks_gp_mutex &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex &obj_hash[i].lock pool_lock irq_context: 0 rcu_tasks.tasks_gp_mutex &base->lock irq_context: 0 rcu_tasks.tasks_gp_mutex &base->lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex &ACCESS_PRIVATE(sdp, lock) irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#3 irq_context: 0 rcu_tasks.tasks_gp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex tasks_rcu_exit_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex kernel/rcu/tasks.h:152 irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[3] irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_tasks__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#2 irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#2 &p->pi_lock irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#2 &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex &x->wait#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_tasks__percpu.cbs_pcpu_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_tasks__percpu.cbs_pcpu_lock &base->lock irq_context: 0 rcu_read_lock rcu_tasks__percpu.cbs_pcpu_lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock &ACCESS_PRIVATE(rtpcp, lock) irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[0] irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &x->wait#2 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &x->wait#2 &p->pi_lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &x->wait#2 &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &x->wait#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &rq->__lock irq_context: 0 rcu_read_lock rcu_tasks_trace__percpu.cbs_pcpu_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_tasks_trace__percpu.cbs_pcpu_lock &base->lock irq_context: 0 rcu_read_lock rcu_tasks_trace__percpu.cbs_pcpu_lock &base->lock &obj_hash[i].lock irq_context: 0 (memory_chain).rwsem irq_context: 0 cpu_hotplug_lock smpboot_threads_lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock fs_reclaim irq_context: 0 cpu_hotplug_lock smpboot_threads_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock smpboot_threads_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock smpboot_threads_lock kthread_create_lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &p->pi_lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &x->wait irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &rq->__lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock smpboot_threads_lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock hrtimer_bases.lock irq_context: 0 cpu_hotplug_lock smpboot_threads_lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 cpu_hotplug_lock smpboot_threads_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &rcu_state.gp_wq irq_context: 0 &stop_pi_lock irq_context: 0 &stop_pi_lock &rq->__lock irq_context: 0 &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &stopper->lock irq_context: 0 (module_notify_list).rwsem irq_context: 0 ddebug_lock irq_context: 0 iort_msi_chip_lock irq_context: 0 irq_domain_mutex irq_context: 0 irq_domain_mutex fs_reclaim irq_context: 0 irq_domain_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 irq_domain_mutex pool_lock#2 irq_context: 0 cci_probing irq_context: 0 cci_probing devtree_lock irq_context: 0 resource_lock irq_context: 0 fixmap_lock fs_reclaim irq_context: 0 fixmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 fixmap_lock &____s->seqcount irq_context: 0 fixmap_lock &c->lock irq_context: 0 fixmap_lock pool_lock#2 irq_context: 0 rcu_read_lock ptlock_ptr(ptdesc) irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex clockevents_lock irq_context: 0 watchdog_mutex irq_context: 0 watchdog_mutex cpu_hotplug_lock irq_context: 0 watchdog_mutex cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 watchdog_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 watchdog_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)events irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &x->wait#4 irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) hrtimer_bases.lock irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&sscs.work) &x->wait#5 irq_context: 0 watchdog_mutex cpu_hotplug_lock &x->wait#5 irq_context: 0 &newf->file_lock irq_context: 0 init_fs.lock &dentry->d_lock irq_context: 0 &p->vtime.seqcount irq_context: 0 cpu_hotplug_lock mem_hotplug_lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock mem_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[1] irq_context: 0 cpu_hotplug_lock mem_hotplug_lock.waiters.lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_hotplug_lock mem_hotplug_lock.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock cpu_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[2] irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.waiters.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.rss.gp_wait.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpuset_hotplug_work irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock fs_reclaim irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock pool_lock#2 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock kthread_create_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &x->wait irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock smpboot_threads_lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &swhash->hlist_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pmus_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pmus_lock &cpuctx_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pcp_batch_high_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &xa->xa_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock fs_reclaim irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pool_lock#2 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &c->lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &____s->seqcount irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock kthread_create_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &x->wait irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock wq_pool_attach_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pcpu_alloc_mutex irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock rcu_node_0 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock (cpu_running).wait.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &base->lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &base->lock &obj_hash[i].lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&timer.timer) irq_context: softirq (&timer.timer) &p->pi_lock irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &rcu_state.gp_wq &p->pi_lock irq_context: softirq &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: softirq &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback &obj_hash[i].lock irq_context: softirq rcu_callback pool_lock#2 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock cpu_hotplug_lock.waiters.lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback mem_hotplug_lock.rss.gp_wait.lock irq_context: softirq rcu_callback mem_hotplug_lock.rss.gp_wait.lock &obj_hash[i].lock irq_context: softirq rcu_callback cpu_hotplug_lock.rss.gp_wait.lock irq_context: 0 its_lock irq_context: 0 its_lock &its->lock irq_context: 0 clockevents_lock tick_broadcast_lock irq_context: 0 clockevents_lock jiffies_seq.seqcount irq_context: 0 clockevents_lock tk_core.seq.seqcount irq_context: 0 &irq_desc_lock_class irq_context: 0 &irq_desc_lock_class irq_controller_lock irq_context: 0 (cpu_running).wait.lock irq_context: 0 (cpu_running).wait.lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock (&timer.timer) irq_context: 0 &x->wait#6 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &x->wait#6 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock sparse_irq_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &x->wait#6 irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state-up smpboot_threads_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up sparse_irq_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up sparse_irq_lock &irq_desc_lock_class irq_context: 0 cpu_hotplug_lock cpuhp_state-up &swhash->hlist_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up pmus_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up pmus_lock &cpuctx_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up &x->wait#4 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &obj_hash[i].lock pool_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up hrtimer_bases.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &x->wait#7 irq_context: 0 cpu_hotplug_lock cpuhp_state-up wq_pool_mutex wq_pool_attach_mutex &pool->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_node_0 irq_context: 0 cpu_hotplug_lock cpuhp_state-up resource_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &pcp->lock &zone->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up &c->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up resource_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock irq_context: hardirq &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock &rt_b->rt_runtime_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock &rt_b->rt_runtime_lock &rt_rq->rt_runtime_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &rq->__lock rcu_read_lock &cfs_b->lock irq_context: 0 &x->wait#6 &p->pi_lock irq_context: 0 &x->wait#6 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#6 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.waiters.lock &p->pi_lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.waiters.lock &p->pi_lock &rq->__lock irq_context: 0 cpu_add_remove_lock cpu_hotplug_lock.waiters.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock stop_cpus_mutex irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stopper->lock irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &rq->__lock irq_context: 0 &x->wait#8 irq_context: 0 rcu_tasks_trace.tasks_gp_mutex rcu_state.exp_mutex rcu_state.exp_wake_mutex &rnp->exp_wq[3] irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &obj_hash[i].lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &base->lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &base->lock &obj_hash[i].lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock stop_cpus_mutex &x->wait#8 irq_context: 0 &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events rdist_memreserve_cpuhp_cleanup_work irq_context: 0 (wq_completion)events rdist_memreserve_cpuhp_cleanup_work cpu_hotplug_lock irq_context: 0 (wq_completion)events rdist_memreserve_cpuhp_cleanup_work cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 sched_domains_mutex irq_context: 0 sched_domains_mutex fs_reclaim irq_context: 0 sched_domains_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sched_domains_mutex pool_lock#2 irq_context: 0 sched_domains_mutex &obj_hash[i].lock irq_context: 0 sched_domains_mutex pcpu_alloc_mutex irq_context: 0 sched_domains_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sched_domains_mutex &c->lock irq_context: 0 sched_domains_mutex &____s->seqcount irq_context: 0 sched_domains_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 sched_domains_mutex rcu_read_lock pool_lock#2 irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &cp->lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &rt_b->rt_runtime_lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock &rt_b->rt_runtime_lock &rt_rq->rt_runtime_lock irq_context: 0 sched_domains_mutex rcu_read_lock &rq->__lock rcu_read_lock &cfs_b->lock irq_context: 0 sched_domains_mutex pcpu_lock irq_context: 0 slab_mutex fs_reclaim irq_context: 0 slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#5/1 irq_context: 0 &type->s_umount_key#5/1 fs_reclaim irq_context: 0 &type->s_umount_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#5/1 pool_lock#2 irq_context: 0 &type->s_umount_key#5/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#5/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#5/1 shrinker_mutex irq_context: 0 &type->s_umount_key#5/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#5/1 sb_lock irq_context: 0 &type->s_umount_key#5/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#5/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#5/1 percpu_counters_lock irq_context: 0 &type->s_umount_key#5/1 crngs.lock irq_context: 0 &type->s_umount_key#5/1 &sbinfo->stat_lock irq_context: 0 &type->s_umount_key#5/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#5/1 &sb->s_type->i_lock_key#5 irq_context: 0 &type->s_umount_key#5/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#5/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#5/1 batched_entropy_u32.lock irq_context: 0 &type->s_umount_key#5/1 &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 &type->s_umount_key#5/1 &dentry->d_lock irq_context: 0 (setup_done).wait.lock irq_context: 0 namespace_sem irq_context: 0 namespace_sem fs_reclaim irq_context: 0 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 namespace_sem &pcp->lock &zone->lock irq_context: 0 namespace_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 namespace_sem &____s->seqcount irq_context: 0 namespace_sem &c->lock irq_context: 0 namespace_sem pool_lock#2 irq_context: 0 namespace_sem mnt_id_ida.xa_lock irq_context: 0 namespace_sem pcpu_alloc_mutex irq_context: 0 namespace_sem pcpu_alloc_mutex pcpu_lock irq_context: 0 namespace_sem &dentry->d_lock irq_context: 0 namespace_sem mount_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &p->alloc_lock init_fs.lock irq_context: 0 rcu_read_lock &____s->seqcount#3 irq_context: 0 file_systems_lock irq_context: 0 &type->s_umount_key#6 irq_context: 0 &type->s_umount_key#6 fs_reclaim irq_context: 0 &type->s_umount_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#6 pool_lock#2 irq_context: 0 &type->s_umount_key#6 &dentry->d_lock irq_context: 0 &type->s_umount_key#6 &____s->seqcount irq_context: 0 &type->s_umount_key#6 &c->lock irq_context: 0 &type->s_umount_key#6 &lru->node[i].lock irq_context: 0 &type->s_umount_key#6 &sbinfo->stat_lock irq_context: 0 &type->s_umount_key#6 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key irq_context: 0 &sb->s_type->i_mutex_key namespace_sem irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key namespace_sem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem mount_lock mount_lock.seqcount pool_lock#2 irq_context: 0 rcu_read_lock &sb->s_type->i_lock_key#2 irq_context: 0 rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_lock_key#5 irq_context: 0 &fs->lock irq_context: 0 &fs->lock &____s->seqcount#3 irq_context: 0 (setup_done).wait.lock &p->pi_lock irq_context: 0 req_lock irq_context: 0 of_mutex irq_context: 0 of_mutex fs_reclaim irq_context: 0 of_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 of_mutex pool_lock#2 irq_context: 0 of_mutex lock irq_context: 0 of_mutex lock kernfs_idr_lock irq_context: 0 of_mutex &root->kernfs_rwsem irq_context: 0 of_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &x->wait#9 irq_context: 0 &k->list_lock irq_context: 0 bus_type_sem irq_context: 0 &root->kernfs_rwsem irq_context: 0 &dev->power.lock irq_context: 0 dpm_list_mtx irq_context: 0 uevent_sock_mutex irq_context: 0 running_helpers_waitq.lock irq_context: 0 sysfs_symlink_target_lock irq_context: 0 &k->k_lock irq_context: 0 &dev->mutex &k->list_lock irq_context: 0 &dev->mutex &k->k_lock irq_context: 0 &dev->mutex &dev->power.lock irq_context: 0 subsys mutex irq_context: 0 memory_blocks.xa_lock irq_context: 0 memory_blocks.xa_lock pool_lock#2 irq_context: 0 lock kernfs_idr_lock &c->lock irq_context: 0 lock kernfs_idr_lock &____s->seqcount irq_context: softirq (&timer.timer) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_tasks_trace.tasks_gp_mutex (&timer.timer) irq_context: 0 rcu_tasks_trace.tasks_gp_mutex (console_sem).lock irq_context: softirq &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 subsys mutex#2 irq_context: 0 subsys mutex#3 irq_context: 0 dev_pm_qos_mtx irq_context: 0 dev_pm_qos_mtx fs_reclaim irq_context: 0 dev_pm_qos_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 dev_pm_qos_mtx pool_lock#2 irq_context: 0 dev_pm_qos_mtx &dev->power.lock irq_context: 0 dev_pm_qos_mtx pm_qos_lock irq_context: 0 dev_pm_qos_sysfs_mtx irq_context: 0 dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 dev_pm_qos_sysfs_mtx fs_reclaim irq_context: 0 dev_pm_qos_sysfs_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 dev_pm_qos_sysfs_mtx pool_lock#2 irq_context: 0 dev_pm_qos_sysfs_mtx lock irq_context: 0 dev_pm_qos_sysfs_mtx lock kernfs_idr_lock irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 dev_pm_qos_mtx &c->lock irq_context: 0 dev_pm_qos_mtx &pcp->lock &zone->lock irq_context: 0 dev_pm_qos_mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 dev_pm_qos_mtx &____s->seqcount irq_context: 0 register_lock irq_context: 0 register_lock proc_subdir_lock irq_context: 0 register_lock fs_reclaim irq_context: 0 register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_lock pool_lock#2 irq_context: 0 register_lock proc_inum_ida.xa_lock irq_context: 0 register_lock proc_subdir_lock irq_context: 0 register_lock &c->lock irq_context: 0 register_lock &____s->seqcount irq_context: 0 register_lock proc_inum_ida.xa_lock pool_lock#2 irq_context: 0 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 batched_entropy_u64.lock crngs.lock base_crng.lock irq_context: 0 &x->wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_pm_notifier.lock irq_context: 0 (cpufreq_policy_notifier_list).rwsem irq_context: 0 (pm_chain_head).rwsem irq_context: 0 cpufreq_governor_mutex irq_context: 0 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 rcu_state.exp_mutex &rq->__lock irq_context: 0 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] irq_context: 0 clocksource_mutex irq_context: 0 syscore_ops_lock irq_context: softirq &(&kfence_timer)->timer irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rcu_state.expedited_wq irq_context: hardirq &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rq->__lock irq_context: 0 binfmt_lock irq_context: 0 pin_fs_lock irq_context: 0 &type->s_umount_key#7/1 irq_context: 0 &type->s_umount_key#7/1 fs_reclaim irq_context: 0 &type->s_umount_key#7/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#7/1 pool_lock#2 irq_context: 0 &type->s_umount_key#7/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#7/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#7/1 shrinker_mutex irq_context: 0 &type->s_umount_key#7/1 &____s->seqcount irq_context: 0 &type->s_umount_key#7/1 &c->lock irq_context: 0 &type->s_umount_key#7/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#7/1 sb_lock irq_context: 0 &type->s_umount_key#7/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#7/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#7/1 &sb->s_type->i_lock_key#6 irq_context: 0 &type->s_umount_key#7/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#7/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#7/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#7/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#7/1 &sb->s_type->i_lock_key#6 &dentry->d_lock irq_context: 0 &type->s_umount_key#7/1 &dentry->d_lock irq_context: 0 rcu_read_lock mount_lock irq_context: 0 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 irq_context: 0 &sb->s_type->i_mutex_key#2 &sb->s_type->i_lock_key#6 irq_context: 0 &sb->s_type->i_mutex_key#2 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#2 rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#2 &sb->s_type->i_lock_key#6 &dentry->d_lock irq_context: 0 &type->s_umount_key#8/1 irq_context: 0 &type->s_umount_key#8/1 fs_reclaim irq_context: 0 &type->s_umount_key#8/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#8/1 pool_lock#2 irq_context: 0 &type->s_umount_key#8/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#8/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#8/1 shrinker_mutex irq_context: 0 &type->s_umount_key#8/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#8/1 sb_lock irq_context: 0 &type->s_umount_key#8/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#8/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#8/1 &sb->s_type->i_lock_key#7 irq_context: 0 &type->s_umount_key#8/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#8/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#8/1 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &type->s_umount_key#8/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#3 irq_context: 0 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 chrdevs_lock irq_context: 0 cb_lock irq_context: 0 cb_lock genl_mutex irq_context: 0 cb_lock genl_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex pool_lock#2 irq_context: 0 subsys mutex#4 irq_context: 0 async_lock irq_context: 0 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 regulator_list_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex device_links_srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex fwnode_link_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex device_links_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->devres_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pinctrl_list_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pinctrl_maps_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &x->wait#9 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex regulator_nesting_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex regulator_ww_class_mutex regulator_nesting_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex devtree_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex bus_type_sem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &____s->seqcount irq_context: 0 (reboot_notifier_list).rwsem irq_context: 0 purge_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dpm_list_mtx irq_context: 0 purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex uevent_sock_mutex irq_context: 0 purge_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex running_helpers_waitq.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#5 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#5 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pin_fs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &x->wait#6 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex rcu_read_lock &rq->__lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex regulator_list_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex deferred_probe_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex probe_waitqueue.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_done.lock irq_context: 0 &type->s_umount_key#9/1 irq_context: 0 &type->s_umount_key#9/1 fs_reclaim irq_context: 0 &type->s_umount_key#9/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#9/1 pool_lock#2 irq_context: 0 &type->s_umount_key#9/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#9/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#9/1 shrinker_mutex irq_context: 0 &type->s_umount_key#9/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#9/1 sb_lock irq_context: 0 &type->s_umount_key#9/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#9/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#9/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#9/1 &____s->seqcount irq_context: 0 &type->s_umount_key#9/1 &c->lock irq_context: 0 &type->s_umount_key#9/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#9/1 &sb->s_type->i_lock_key#8 irq_context: 0 &type->s_umount_key#9/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#9/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#9/1 &sb->s_type->i_lock_key#8 &dentry->d_lock irq_context: 0 &type->s_umount_key#9/1 &dentry->d_lock irq_context: 0 pernet_ops_rwsem fs_reclaim irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem &c->lock irq_context: 0 pernet_ops_rwsem sysctl_lock irq_context: 0 vmap_purge_lock irq_context: 0 vmap_purge_lock purge_vmap_area_lock irq_context: 0 vmap_purge_lock free_vmap_area_lock irq_context: 0 vmap_purge_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 vmap_purge_lock free_vmap_area_lock pool_lock#2 irq_context: 0 vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock irq_context: 0 vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &obj_hash[i].lock irq_context: 0 &fp->aux->used_maps_mutex irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex pcpu_lock irq_context: 0 proto_list_mutex irq_context: 0 targets_mutex irq_context: 0 nl_table_lock irq_context: 0 nl_table_wait.lock irq_context: 0 net_family_lock irq_context: 0 pernet_ops_rwsem net_generic_ids.xa_lock irq_context: 0 pernet_ops_rwsem mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem &dir->lock irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex irq_context: 0 rtnl_mutex fs_reclaim irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sparse_irq_lock irq_context: 0 sparse_irq_lock fs_reclaim irq_context: 0 sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sparse_irq_lock pool_lock#2 irq_context: 0 sparse_irq_lock &c->lock irq_context: 0 sparse_irq_lock &____s->seqcount irq_context: 0 sparse_irq_lock lock irq_context: 0 sparse_irq_lock lock kernfs_idr_lock irq_context: 0 sparse_irq_lock &root->kernfs_rwsem irq_context: 0 sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sparse_irq_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &cma->lock irq_context: 0 cma_mutex irq_context: 0 cma_mutex &zone->lock irq_context: 0 cma_mutex &zone->lock &____s->seqcount irq_context: 0 cma_mutex pcpu_drain_mutex &pcp->lock irq_context: 0 cma_mutex pcpu_drain_mutex &pcp->lock &zone->lock irq_context: 0 cma_mutex pcpu_drain_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 cma_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cma_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 cma_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 cma_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cma_mutex &obj_hash[i].lock irq_context: 0 cma_mutex lock#2 irq_context: 0 &pool->lock#2 irq_context: 0 cma_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 cma_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 freezer_lock irq_context: 0 &list->lock irq_context: 0 kauditd_wait.lock irq_context: 0 audit_backlog_wait.lock irq_context: 0 lock#3 irq_context: 0 lock#3 &zone->lock irq_context: 0 pcp_batch_high_lock irq_context: 0 khugepaged_mutex irq_context: 0 &(&priv->bus_notifier)->rwsem irq_context: 0 gdp_mutex irq_context: 0 gdp_mutex &k->list_lock irq_context: 0 gdp_mutex fs_reclaim irq_context: 0 gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 gdp_mutex pool_lock#2 irq_context: 0 gdp_mutex lock irq_context: 0 gdp_mutex lock kernfs_idr_lock irq_context: 0 gdp_mutex &root->kernfs_rwsem irq_context: 0 gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 subsys mutex#6 irq_context: 0 subsys mutex#6 &k->k_lock irq_context: 0 subsys mutex#7 irq_context: 0 subsys mutex#7 &k->list_lock irq_context: 0 subsys mutex#7 &k->k_lock irq_context: 0 regmap_debugfs_early_lock irq_context: 0 (acpi_reconfig_chain).rwsem irq_context: 0 __i2c_board_lock irq_context: 0 quarantine_lock irq_context: 0 core_lock irq_context: 0 core_lock &k->list_lock irq_context: 0 core_lock &k->k_lock irq_context: 0 cb_lock genl_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex nl_table_wait.lock irq_context: 0 remove_cache_srcu irq_context: 0 remove_cache_srcu quarantine_lock irq_context: 0 nl_table_lock irq_context: 0 thermal_governor_lock irq_context: 0 thermal_governor_lock thermal_list_lock irq_context: 0 cpuidle_lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 cpuidle_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpuidle_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 cpuidle_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 cpuidle_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpuidle_lock &obj_hash[i].lock irq_context: 0 cpuidle_lock (console_sem).lock irq_context: 0 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_lock_key#8 irq_context: 0 &dir->lock irq_context: 0 k-sk_lock-AF_QIPCRTR irq_context: 0 k-sk_lock-AF_QIPCRTR k-slock-AF_QIPCRTR irq_context: 0 k-slock-AF_QIPCRTR irq_context: 0 k-sk_lock-AF_QIPCRTR fs_reclaim irq_context: 0 k-sk_lock-AF_QIPCRTR fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 k-sk_lock-AF_QIPCRTR qrtr_ports.xa_lock irq_context: 0 k-sk_lock-AF_QIPCRTR pool_lock#2 irq_context: 0 k-sk_lock-AF_QIPCRTR qrtr_node_lock irq_context: 0 k-sk_lock-AF_QIPCRTR &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex irq_context: 0 resource_lock pool_lock#2 irq_context: 0 resource_lock &obj_hash[i].lock irq_context: 0 resource_lock &c->lock irq_context: 0 resource_lock &____s->seqcount irq_context: 0 crngs.lock irq_context: 0 (crypto_chain).rwsem irq_context: 0 tty_mutex irq_context: 0 iova_cache_mutex irq_context: 0 iova_cache_mutex cpu_hotplug_lock irq_context: 0 iova_cache_mutex cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 iova_cache_mutex slab_mutex irq_context: 0 iova_cache_mutex slab_mutex fs_reclaim irq_context: 0 iova_cache_mutex slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 iova_cache_mutex slab_mutex pool_lock#2 irq_context: 0 iova_cache_mutex slab_mutex pcpu_alloc_mutex irq_context: 0 iova_cache_mutex slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 gdp_mutex &c->lock irq_context: 0 gdp_mutex &____s->seqcount irq_context: 0 subsys mutex#8 irq_context: 0 subsys mutex#8 &k->k_lock irq_context: 0 device_links_lock irq_context: 0 uidhash_lock irq_context: 0 &type->s_umount_key#25/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#25/1 &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#25/1 &sb->s_type->i_lock_key#3 &xa->xa_lock#8 irq_context: 0 rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 oom_reaper_wait.lock irq_context: 0 subsys mutex#9 irq_context: 0 &pgdat->kcompactd_wait irq_context: 0 hugetlb_lock irq_context: 0 &type->s_umount_key#25/1 lock#4 &lruvec->lru_lock irq_context: 0 memory_tier_lock irq_context: 0 memory_tier_lock fs_reclaim irq_context: 0 memory_tier_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 memory_tier_lock pool_lock#2 irq_context: 0 memory_tier_lock &x->wait#9 irq_context: 0 memory_tier_lock &obj_hash[i].lock irq_context: 0 memory_tier_lock &k->list_lock irq_context: 0 memory_tier_lock &c->lock irq_context: 0 memory_tier_lock &____s->seqcount irq_context: 0 memory_tier_lock lock irq_context: 0 memory_tier_lock lock kernfs_idr_lock irq_context: 0 memory_tier_lock &root->kernfs_rwsem irq_context: 0 memory_tier_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 memory_tier_lock bus_type_sem irq_context: 0 memory_tier_lock sysfs_symlink_target_lock irq_context: 0 memory_tier_lock &k->k_lock irq_context: 0 memory_tier_lock &root->kernfs_rwsem irq_context: 0 memory_tier_lock &dev->power.lock irq_context: 0 memory_tier_lock dpm_list_mtx irq_context: 0 memory_tier_lock uevent_sock_mutex irq_context: 0 memory_tier_lock running_helpers_waitq.lock irq_context: 0 memory_tier_lock &dev->mutex &k->list_lock irq_context: 0 memory_tier_lock &dev->mutex &k->k_lock irq_context: 0 memory_tier_lock &dev->mutex &dev->power.lock irq_context: 0 memory_tier_lock subsys mutex#10 irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 memory_tier_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 memory_tier_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 memory_tier_lock rcu_state.exp_mutex &rq->__lock irq_context: hardirq &rcu_state.expedited_wq &p->pi_lock irq_context: 0 ksm_thread_mutex irq_context: 0 ksm_thread_wait.lock irq_context: 0 khugepaged_mutex fs_reclaim irq_context: 0 khugepaged_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 khugepaged_mutex pool_lock#2 irq_context: 0 khugepaged_mutex kthread_create_lock irq_context: 0 khugepaged_mutex &p->pi_lock irq_context: 0 khugepaged_mutex &p->pi_lock &rq->__lock irq_context: 0 khugepaged_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 khugepaged_mutex &x->wait irq_context: 0 khugepaged_mutex &rq->__lock irq_context: 0 khugepaged_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 khugepaged_mutex &obj_hash[i].lock irq_context: 0 khugepaged_mutex lock#3 irq_context: 0 khugepaged_mutex lock#3 &zone->lock irq_context: 0 khugepaged_mutex pcp_batch_high_lock irq_context: 0 cgroup_mutex fs_reclaim irq_context: 0 cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 damon_ops_lock irq_context: 0 crypto_alg_sem irq_context: 0 crypto_alg_sem (crypto_chain).rwsem irq_context: 0 cpu_hotplug_lock fs_reclaim irq_context: 0 cpu_hotplug_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock pcpu_alloc_mutex irq_context: 0 cpu_hotplug_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 cpu_hotplug_lock &wq->mutex irq_context: 0 cpu_hotplug_lock &wq->mutex &pool->lock irq_context: 0 cpu_hotplug_lock kthread_create_lock irq_context: 0 cpu_hotplug_lock &p->pi_lock irq_context: 0 cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock &rq->__lock irq_context: 0 lock#2 irq_context: 0 khugepaged_mm_lock irq_context: 0 khugepaged_wait.lock irq_context: 0 cpu_hotplug_lock &x->wait irq_context: 0 cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock (worker)->lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock (worker)->lock &p->pi_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock (worker)->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock (worker)->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock &____s->seqcount irq_context: 0 cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock hrtimer_bases.lock irq_context: 0 cpu_hotplug_lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 cpu_hotplug_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 bio_slab_lock irq_context: 0 bio_slab_lock fs_reclaim irq_context: 0 bio_slab_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 bio_slab_lock pool_lock#2 irq_context: 0 bio_slab_lock slab_mutex irq_context: 0 bio_slab_lock slab_mutex fs_reclaim irq_context: 0 bio_slab_lock slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 bio_slab_lock slab_mutex pool_lock#2 irq_context: 0 bio_slab_lock slab_mutex &c->lock irq_context: 0 bio_slab_lock slab_mutex &____s->seqcount irq_context: 0 bio_slab_lock slab_mutex pcpu_alloc_mutex irq_context: 0 bio_slab_lock slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 bio_slab_lock bio_slabs.xa_lock irq_context: 0 bio_slab_lock bio_slabs.xa_lock pool_lock#2 irq_context: 0 bio_slab_lock bio_slabs.xa_lock &c->lock irq_context: 0 bio_slab_lock bio_slabs.xa_lock &____s->seqcount irq_context: 0 major_names_lock irq_context: 0 major_names_lock fs_reclaim irq_context: 0 major_names_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 major_names_lock pool_lock#2 irq_context: 0 major_names_lock major_names_spinlock irq_context: 0 rcu_tasks.tasks_gp_mutex (&timer.timer) irq_context: 0 rcu_tasks.tasks_gp_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &type->s_umount_key#25/1 lock#5 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&rtpcp->lazy_timer) irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks_trace__percpu.cbs_pcpu_lock irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks__percpu.cbs_pcpu_lock irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks__percpu.cbs_pcpu_lock &obj_hash[i].lock irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks__percpu.cbs_pcpu_lock &base->lock irq_context: softirq (&rtpcp->lazy_timer) rcu_tasks__percpu.cbs_pcpu_lock &base->lock &obj_hash[i].lock irq_context: 0 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) (&timer.timer) irq_context: 0 rcu_tasks.tasks_gp_mutex (console_sem).lock irq_context: 0 &pgdat->kswapd_lock irq_context: softirq drivers/char/random.c:251 irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (next_reseed).work irq_context: 0 (wq_completion)events_unbound (next_reseed).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work &base->lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work input_pool.lock irq_context: 0 (wq_completion)events_unbound (next_reseed).work base_crng.lock irq_context: 0 &pool->lock &base->lock irq_context: 0 &pool->lock &base->lock &obj_hash[i].lock irq_context: softirq mm/vmstat.c:2022 irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (shepherd).work irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (shepherd).work cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (shepherd).work &obj_hash[i].lock irq_context: 0 (wq_completion)events (shepherd).work &base->lock irq_context: 0 (wq_completion)events (shepherd).work &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mm_percpu_wq irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &obj_hash[i].lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &base->lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &rq->__lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 slab_mutex batched_entropy_u8.lock irq_context: 0 slab_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 slab_mutex batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 slab_mutex kfence_freelist_lock irq_context: 0 console_lock fs_reclaim irq_context: 0 console_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 console_lock &x->wait#9 irq_context: 0 console_lock &k->list_lock irq_context: 0 console_lock gdp_mutex irq_context: 0 console_lock gdp_mutex &k->list_lock irq_context: 0 console_lock gdp_mutex fs_reclaim irq_context: 0 console_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 console_lock gdp_mutex pool_lock#2 irq_context: 0 console_lock gdp_mutex lock irq_context: 0 console_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 console_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 console_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 console_lock lock irq_context: 0 console_lock lock kernfs_idr_lock irq_context: 0 console_lock &root->kernfs_rwsem irq_context: 0 console_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 console_lock bus_type_sem irq_context: 0 console_lock sysfs_symlink_target_lock irq_context: 0 console_lock &root->kernfs_rwsem irq_context: 0 console_lock &dev->power.lock irq_context: 0 console_lock dpm_list_mtx irq_context: 0 console_lock uevent_sock_mutex irq_context: 0 console_lock running_helpers_waitq.lock irq_context: 0 console_lock subsys mutex#11 irq_context: 0 console_lock subsys mutex#11 &k->k_lock irq_context: 0 shrink_qlist.lock irq_context: 0 remove_cache_srcu_srcu_usage.lock irq_context: 0 remove_cache_srcu_srcu_usage.lock &obj_hash[i].lock irq_context: 0 &ACCESS_PRIVATE(sdp, lock) irq_context: 0 remove_cache_srcu irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu_srcu_usage.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex remove_cache_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) remove_cache_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &x->wait#3 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex remove_cache_srcu_srcu_usage.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock flush_lock irq_context: 0 cpu_hotplug_lock flush_lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock flush_lock (work_completion)(&sfw->work) irq_context: 0 cpu_hotplug_lock flush_lock rcu_read_lock (wq_completion)slub_flushwq irq_context: 0 cpu_hotplug_lock flush_lock &x->wait#10 irq_context: 0 cpu_hotplug_lock flush_lock &rq->__lock irq_context: 0 cpu_hotplug_lock flush_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)slub_flushwq irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) &c->lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) &n->list_lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&sfw->work) &obj_hash[i].lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)slub_flushwq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &base->lock &obj_hash[i].lock irq_context: softirq &(&ssp->srcu_sup->work)->timer irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &n->list_lock &c->lock irq_context: 0 &meta->lock irq_context: 0 clk_debug_lock irq_context: 0 clocks_mutex irq_context: 0 acpi_scan_lock irq_context: 0 acpi_scan_lock semaphore->lock irq_context: 0 acpi_scan_lock fs_reclaim irq_context: 0 acpi_scan_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock pool_lock#2 irq_context: 0 acpi_scan_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &x->wait#9 irq_context: 0 acpi_scan_lock acpi_device_lock irq_context: 0 acpi_scan_lock acpi_device_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_device_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_device_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_device_lock &xa->xa_lock#2 irq_context: 0 acpi_scan_lock acpi_device_lock semaphore->lock irq_context: 0 acpi_scan_lock acpi_device_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &k->list_lock irq_context: 0 acpi_scan_lock lock irq_context: 0 acpi_scan_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock bus_type_sem irq_context: 0 acpi_scan_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &k->k_lock irq_context: 0 acpi_scan_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &c->lock irq_context: 0 acpi_scan_lock &____s->seqcount irq_context: 0 acpi_scan_lock &dev->power.lock irq_context: 0 acpi_scan_lock dpm_list_mtx irq_context: 0 acpi_scan_lock &dev->mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex &dev->power.lock irq_context: 0 acpi_scan_lock subsys mutex#12 irq_context: 0 acpi_scan_lock uevent_sock_mutex irq_context: 0 acpi_scan_lock &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 acpi_scan_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock acpi_device_lock &c->lock irq_context: 0 acpi_scan_lock acpi_device_lock &____s->seqcount irq_context: 0 acpi_scan_lock &n->list_lock irq_context: 0 acpi_scan_lock &n->list_lock &c->lock irq_context: 0 acpi_scan_lock quarantine_lock irq_context: 0 acpi_scan_lock acpi_ioremap_lock irq_context: 0 acpi_scan_lock acpi_ioremap_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_ioremap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_ioremap_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_ioremap_lock &____s->seqcount irq_context: 0 acpi_scan_lock acpi_ioremap_lock &c->lock irq_context: 0 acpi_scan_lock acpi_ioremap_lock free_vmap_area_lock irq_context: 0 acpi_scan_lock acpi_ioremap_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock acpi_ioremap_lock free_vmap_area_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_ioremap_lock vmap_area_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock fs_reclaim irq_context: 0 acpi_scan_lock &device->physical_node_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &device->physical_node_lock pool_lock#2 irq_context: 0 acpi_scan_lock &device->physical_node_lock lock irq_context: 0 acpi_scan_lock &device->physical_node_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &device->physical_node_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &device->physical_node_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &device->physical_node_lock &c->lock irq_context: 0 acpi_scan_lock &device->physical_node_lock &____s->seqcount irq_context: 0 acpi_scan_lock irq_domain_mutex irq_context: 0 acpi_scan_lock &domain->mutex irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock fs_reclaim irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock &____s->seqcount irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock &c->lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &domain->mutex &irq_desc_lock_class irq_context: 0 acpi_scan_lock &domain->mutex fs_reclaim irq_context: 0 acpi_scan_lock &domain->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &domain->mutex pool_lock#2 irq_context: 0 acpi_scan_lock resource_lock irq_context: 0 acpi_scan_lock &(&priv->bus_notifier)->rwsem irq_context: 0 acpi_scan_lock &(&priv->bus_notifier)->rwsem &device->physical_node_lock irq_context: 0 acpi_scan_lock fwnode_link_lock irq_context: 0 acpi_scan_lock fwnode_link_lock &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex &device->physical_node_lock irq_context: 0 acpi_scan_lock &dev->mutex device_links_srcu irq_context: 0 acpi_scan_lock &dev->mutex fwnode_link_lock irq_context: 0 acpi_scan_lock &dev->mutex device_links_lock irq_context: 0 acpi_scan_lock &dev->mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex &c->lock irq_context: 0 acpi_scan_lock &dev->mutex &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex &dev->devres_lock irq_context: 0 acpi_scan_lock &dev->mutex pinctrl_list_mutex irq_context: 0 acpi_scan_lock &dev->mutex pinctrl_maps_mutex irq_context: 0 acpi_scan_lock &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 acpi_scan_lock &dev->mutex sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex lock irq_context: 0 acpi_scan_lock &dev->mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock semaphore->lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock wakeup_ida.xa_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &x->wait#9 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock bus_type_sem irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &c->lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock uevent_sock_mutex irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock events_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_wakeup_lock irq_context: 0 acpi_scan_lock &dev->mutex resource_lock irq_context: 0 acpi_scan_lock &dev->mutex free_vmap_area_lock irq_context: 0 acpi_scan_lock &dev->mutex vmap_area_lock irq_context: 0 acpi_scan_lock &dev->mutex init_mm.page_table_lock irq_context: 0 acpi_scan_lock &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock &dev->mutex chrdevs_lock irq_context: 0 acpi_scan_lock &dev->mutex tty_mutex irq_context: 0 acpi_scan_lock &dev->mutex proc_subdir_lock irq_context: 0 acpi_scan_lock &dev->mutex proc_inum_ida.xa_lock irq_context: 0 acpi_scan_lock &dev->mutex proc_subdir_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex port_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &x->wait#9 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex bus_type_sem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &c->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->power.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex dpm_list_mtx irq_context: 0 acpi_scan_lock &dev->mutex port_mutex uevent_sock_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex device_links_srcu irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex fwnode_link_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex device_links_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->devres_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex pinctrl_list_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex pinctrl_maps_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex deferred_probe_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex uevent_sock_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex probe_waitqueue.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex subsys mutex#14 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &xa->xa_lock#3 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &rq->__lock irq_context: 0 (wq_completion)pm irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock &dev->power.wait_queue irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex (console_sem).lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &port_lock_key irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex &port_lock_key irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex syslog_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex (console_sem).lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex console_lock console_srcu console_owner_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex console_lock console_srcu console_owner irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &type->s_umount_key#25/1 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#25/1 crypto_alg_sem irq_context: 0 &type->s_umount_key#25/1 lock#2 irq_context: 0 &type->s_umount_key#25/1 lock#2 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&group->avgs_work)->timer irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq mm/memcontrol.c:679 irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work cgroup_rstat_lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work cgroup_rstat_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &base->lock irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &base->lock &obj_hash[i].lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_mutex &rq->__lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex batched_entropy_u8.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex batched_entropy_u8.lock crngs.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex kfence_freelist_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex ctrl_ida.xa_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &x->wait#9 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->power.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &c->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex semaphore->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex dpm_list_mtx irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex uevent_sock_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &dev->power.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &dev->power.lock hrtimer_bases.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &dev->mutex &dev->power.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex subsys mutex#15 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &n->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &n->list_lock &c->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &k->k_lock klist_remove_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex deferred_probe_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex device_links_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex &k->list_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex fs_reclaim irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex bus_type_sem irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex req_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &p->pi_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &x->wait#11 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &rq->__lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers irq_context: 0 sb_writers mount_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 fs_reclaim irq_context: 0 sb_writers &type->i_mutex_dir_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sbinfo->stat_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &s->s_inode_list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tk_core.seq.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 batched_entropy_u32.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &xattrs->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &xattrs->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 smack_known_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 smack_known_lock &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 smack_known_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 tk_core.seq.seqcount irq_context: 0 &x->wait#11 irq_context: 0 &x->wait#11 &p->pi_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex subsys mutex#16 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex subsys mutex#16 &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex chrdevs_lock irq_context: 0 acpi_scan_lock &dev->mutex fwnode_link_lock &k->k_lock irq_context: 0 acpi_scan_lock &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 acpi_scan_lock &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 acpi_scan_lock &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex deferred_probe_mutex irq_context: 0 acpi_scan_lock &dev->mutex uevent_sock_mutex irq_context: 0 acpi_scan_lock &dev->mutex running_helpers_waitq.lock irq_context: 0 acpi_scan_lock &dev->mutex probe_waitqueue.lock irq_context: 0 acpi_scan_lock subsys mutex#4 irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &c->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &dev->mutex &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_lock console_srcu console_owner_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_lock console_srcu console_owner irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock &c->lock irq_context: 0 acpi_scan_lock &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &c->lock irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &port->mutex &device->physical_node_lock &____s->seqcount irq_context: 0 acpi_scan_lock &dev->mutex port_mutex &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock (console_sem).lock irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner_lock irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 acpi_scan_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 acpi_scan_lock &domain->mutex sparse_irq_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 acpi_scan_lock free_vmap_area_lock irq_context: 0 acpi_scan_lock vmap_area_lock irq_context: 0 acpi_scan_lock init_mm.page_table_lock irq_context: 0 acpi_scan_lock io_range_mutex irq_context: 0 acpi_scan_lock pci_bus_sem irq_context: 0 acpi_scan_lock gdp_mutex irq_context: 0 acpi_scan_lock gdp_mutex &k->list_lock irq_context: 0 acpi_scan_lock gdp_mutex fs_reclaim irq_context: 0 acpi_scan_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock gdp_mutex pool_lock#2 irq_context: 0 acpi_scan_lock gdp_mutex lock irq_context: 0 acpi_scan_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock subsys mutex#17 irq_context: 0 acpi_scan_lock subsys mutex#17 &k->k_lock irq_context: 0 acpi_scan_lock acpi_hp_context_lock irq_context: 0 acpi_scan_lock acpi_hp_context_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_hp_context_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_hp_context_lock pool_lock#2 irq_context: 0 acpi_scan_lock bridge_mutex irq_context: 0 acpi_scan_lock pci_bus_sem irq_context: 0 acpi_scan_lock pci_lock irq_context: 0 acpi_scan_lock pci_acpi_companion_lookup_sem irq_context: 0 acpi_scan_lock pci_slot_mutex irq_context: 0 acpi_scan_lock resource_alignment_lock irq_context: 0 acpi_scan_lock device_links_srcu irq_context: 0 acpi_scan_lock &dev->power.lock &dev->power.lock/1 irq_context: 0 acpi_scan_lock iort_msi_chip_lock irq_context: 0 acpi_scan_lock subsys mutex#18 irq_context: 0 acpi_scan_lock devtree_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock semaphore->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock batched_entropy_u8.lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock kfence_freelist_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock wakeup_ida.xa_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &x->wait#9 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->list_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &k->list_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &c->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &____s->seqcount irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex pool_lock#2 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock kernfs_idr_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock bus_type_sem irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock sysfs_symlink_target_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &c->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &____s->seqcount irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock uevent_sock_mutex irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &pcp->lock &zone->lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &obj_hash[i].lock pool_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock running_helpers_waitq.lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->k_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 &k->k_lock irq_context: 0 acpi_scan_lock acpi_pm_notifier_install_lock acpi_pm_notifier_lock events_lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock &dev->mutex &dev->power.lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock &dev->mutex &k->list_lock irq_context: 0 acpi_scan_lock pci_rescan_remove_lock &dev->mutex &k->k_lock irq_context: 0 acpi_scan_lock acpi_link_lock irq_context: 0 acpi_scan_lock acpi_link_lock fs_reclaim irq_context: 0 acpi_scan_lock acpi_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 acpi_scan_lock acpi_link_lock pool_lock#2 irq_context: 0 acpi_scan_lock acpi_link_lock semaphore->lock irq_context: 0 acpi_scan_lock acpi_link_lock &obj_hash[i].lock irq_context: 0 acpi_scan_lock acpi_link_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 acpi_scan_lock acpi_link_lock (console_sem).lock irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner_lock irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 acpi_scan_lock acpi_link_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 acpi_scan_lock acpi_link_lock &c->lock irq_context: 0 acpi_scan_lock acpi_link_lock &____s->seqcount irq_context: 0 acpi_scan_lock acpi_dep_list_lock irq_context: 0 acpi_scan_lock power_resource_list_lock irq_context: 0 acpi_device_lock irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem pool_lock#2 irq_context: 0 klist_remove_lock irq_context: 0 &k->k_lock klist_remove_lock irq_context: 0 kernfs_idr_lock irq_context: 0 console_lock console_srcu console_owner_lock irq_context: 0 console_lock console_srcu console_owner irq_context: 0 console_lock console_srcu console_owner &port_lock_key irq_context: 0 console_lock console_srcu console_owner console_owner_lock irq_context: 0 k-sk_lock-AF_NETLINK irq_context: 0 k-sk_lock-AF_NETLINK k-slock-AF_NETLINK irq_context: 0 k-sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 k-slock-AF_NETLINK irq_context: 0 cpu_hotplug_lock cpuhp_state-up fs_reclaim irq_context: 0 cpu_hotplug_lock cpuhp_state-up fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock cpuhp_state-up lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up lock kernfs_idr_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &root->kernfs_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state-up &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &type->s_umount_key#10/1 irq_context: 0 &type->s_umount_key#10/1 fs_reclaim irq_context: 0 &type->s_umount_key#10/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#10/1 pool_lock#2 irq_context: 0 &type->s_umount_key#10/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#10/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#10/1 shrinker_mutex irq_context: 0 &type->s_umount_key#10/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#10/1 sb_lock irq_context: 0 &type->s_umount_key#10/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#10/1 &____s->seqcount irq_context: 0 &type->s_umount_key#10/1 &c->lock irq_context: 0 &type->s_umount_key#10/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#10/1 &sb->s_type->i_lock_key#9 irq_context: 0 &type->s_umount_key#10/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#10/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#10/1 &sb->s_type->i_lock_key#9 &dentry->d_lock irq_context: 0 &type->s_umount_key#10/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#11/1 irq_context: 0 &type->s_umount_key#11/1 fs_reclaim irq_context: 0 &type->s_umount_key#11/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#11/1 pool_lock#2 irq_context: 0 &type->s_umount_key#11/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#11/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#11/1 shrinker_mutex irq_context: 0 &type->s_umount_key#11/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#11/1 sb_lock irq_context: 0 &type->s_umount_key#11/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#11/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#11/1 &sb->s_type->i_lock_key#10 irq_context: 0 &type->s_umount_key#11/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#11/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#11/1 &c->lock irq_context: 0 &type->s_umount_key#11/1 &____s->seqcount irq_context: 0 &type->s_umount_key#11/1 &sb->s_type->i_lock_key#10 &dentry->d_lock irq_context: 0 &type->s_umount_key#11/1 &dentry->d_lock irq_context: 0 &mm->mmap_lock irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim &mapping->i_mmap_rwsem irq_context: 0 &mm->mmap_lock reservation_ww_class_acquire reservation_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start dma_fence_map irq_context: 0 &mm->mmap_lock irq_context: 0 key irq_context: 0 attribute_container_mutex irq_context: 0 triggers_list_lock irq_context: 0 leds_list_lock irq_context: 0 bus_type_sem irq_context: 0 (usb_notifier_list).rwsem irq_context: 0 &device->physical_node_lock irq_context: 0 rc_map_lock irq_context: 0 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &root->kernfs_rwsem &____s->seqcount irq_context: 0 pci_lock irq_context: 0 subsys mutex#19 irq_context: 0 &(&priv->bus_notifier)->rwsem irq_context: 0 &(&priv->bus_notifier)->rwsem iommu_probe_device_lock irq_context: 0 &(&priv->bus_notifier)->rwsem iommu_probe_device_lock iommu_device_lock irq_context: 0 (efi_runtime_lock).lock irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &x->wait#12 irq_context: 0 (wq_completion)efi_rts_wq irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) cpu_asid_lock irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) efi_rt_lock irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) &x->wait#12 irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) &x->wait#12 &p->pi_lock irq_context: 0 (efivars_lock).lock irq_context: 0 devfreq_list_lock irq_context: 0 &entry->access irq_context: 0 info_mutex irq_context: 0 info_mutex proc_subdir_lock irq_context: 0 info_mutex fs_reclaim irq_context: 0 info_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 info_mutex &c->lock irq_context: 0 info_mutex &____s->seqcount irq_context: 0 info_mutex pool_lock#2 irq_context: 0 info_mutex proc_inum_ida.xa_lock irq_context: 0 info_mutex proc_subdir_lock irq_context: 0 kobj_ns_type_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &x->wait#9 irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex bus_type_sem irq_context: 0 pernet_ops_rwsem rtnl_mutex sysfs_symlink_target_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex running_helpers_waitq.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &dir->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex input_pool.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock irq_context: 0 rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock nl_table_lock irq_context: 0 rcu_read_lock nl_table_wait.lock irq_context: 0 qdisc_mod_lock irq_context: 0 bt_proto_lock irq_context: 0 hci_cb_list_lock irq_context: 0 mgmt_chan_list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 rate_ctrl_mutex irq_context: 0 rate_ctrl_mutex fs_reclaim irq_context: 0 rate_ctrl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rate_ctrl_mutex pool_lock#2 irq_context: 0 netlbl_domhsh_lock irq_context: 0 netlbl_unlhsh_lock irq_context: 0 rcu_read_lock netlbl_domhsh_lock irq_context: 0 rcu_read_lock netlbl_domhsh_lock pool_lock#2 irq_context: 0 misc_mtx irq_context: 0 misc_mtx fs_reclaim irq_context: 0 misc_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx pool_lock#2 irq_context: 0 misc_mtx &x->wait#9 irq_context: 0 misc_mtx &obj_hash[i].lock irq_context: 0 misc_mtx &____s->seqcount irq_context: 0 misc_mtx &k->list_lock irq_context: 0 misc_mtx gdp_mutex irq_context: 0 misc_mtx gdp_mutex &k->list_lock irq_context: 0 misc_mtx gdp_mutex fs_reclaim irq_context: 0 misc_mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx gdp_mutex pool_lock#2 irq_context: 0 misc_mtx gdp_mutex lock irq_context: 0 misc_mtx gdp_mutex lock kernfs_idr_lock irq_context: 0 misc_mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 misc_mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 misc_mtx lock irq_context: 0 misc_mtx lock kernfs_idr_lock irq_context: 0 misc_mtx &root->kernfs_rwsem irq_context: 0 misc_mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 misc_mtx bus_type_sem irq_context: 0 misc_mtx sysfs_symlink_target_lock irq_context: 0 misc_mtx &c->lock irq_context: 0 misc_mtx &root->kernfs_rwsem irq_context: 0 misc_mtx &dev->power.lock irq_context: 0 misc_mtx dpm_list_mtx irq_context: 0 misc_mtx req_lock irq_context: 0 misc_mtx &p->pi_lock irq_context: 0 misc_mtx &x->wait#11 irq_context: 0 misc_mtx &rq->__lock irq_context: 0 misc_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx uevent_sock_mutex irq_context: 0 misc_mtx running_helpers_waitq.lock irq_context: 0 misc_mtx subsys mutex#21 irq_context: 0 misc_mtx subsys mutex#21 &k->k_lock irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex irq_context: 0 input_mutex irq_context: 0 input_mutex input_devices_poll_wait.lock irq_context: 0 (netlink_chain).rwsem irq_context: 0 proto_tab_lock irq_context: 0 random_ready_notifier.lock irq_context: 0 random_ready_notifier.lock crngs.lock irq_context: 0 misc_mtx misc_minors_ida.xa_lock irq_context: 0 misc_mtx &obj_hash[i].lock pool_lock irq_context: hardirq &rq->__lock &rt_b->rt_runtime_lock irq_context: hardirq &rq->__lock &rt_b->rt_runtime_lock tk_core.seq.seqcount irq_context: hardirq &rq->__lock &rt_b->rt_runtime_lock hrtimer_bases.lock irq_context: hardirq &rq->__lock &rt_b->rt_runtime_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &rq->__lock &rt_rq->rt_runtime_lock irq_context: 0 wtd_deferred_reg_mutex irq_context: 0 &type->s_umount_key#12/1 irq_context: 0 &type->s_umount_key#12/1 fs_reclaim irq_context: 0 &type->s_umount_key#12/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#12/1 pool_lock#2 irq_context: 0 &type->s_umount_key#12/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#12/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#12/1 shrinker_mutex irq_context: 0 &type->s_umount_key#12/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#12/1 sb_lock irq_context: 0 &type->s_umount_key#12/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#12/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#12/1 &sb->s_type->i_lock_key#11 irq_context: 0 &type->s_umount_key#12/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#12/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#12/1 &sb->s_type->i_lock_key#11 &dentry->d_lock irq_context: 0 &type->s_umount_key#12/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#11 irq_context: 0 clocksource_mutex cpu_hotplug_lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stopper->lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &rq->__lock irq_context: 0 clocksource_mutex cpu_hotplug_lock stop_cpus_mutex &x->wait#8 irq_context: 0 clocksource_mutex (console_sem).lock irq_context: 0 clocksource_mutex console_lock console_srcu console_owner_lock irq_context: 0 clocksource_mutex console_lock console_srcu console_owner irq_context: 0 clocksource_mutex console_lock console_srcu console_owner &port_lock_key irq_context: hardirq tick_broadcast_lock irq_context: hardirq tick_broadcast_lock jiffies_lock irq_context: 0 (wq_completion)events timer_update_work irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events timer_update_work timer_keys_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 jiffies_seq.seqcount irq_context: 0 clocksource_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &type->s_umount_key#13/1 irq_context: 0 &type->s_umount_key#13/1 fs_reclaim irq_context: 0 &type->s_umount_key#13/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#13/1 pool_lock#2 irq_context: 0 &type->s_umount_key#13/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#13/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#13/1 shrinker_mutex irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#13/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#13/1 sb_lock irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#13/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#25/1 lock#2 (work_completion)(work) irq_context: 0 &type->s_umount_key#13/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#25/1 lock#2 rcu_read_lock (wq_completion)mm_percpu_wq irq_context: 0 &type->s_umount_key#13/1 &sb->s_type->i_lock_key#12 irq_context: 0 &type->s_umount_key#25/1 lock#2 &x->wait#10 irq_context: 0 &type->s_umount_key#13/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#13/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#25/1 lock#2 &rq->__lock irq_context: 0 &type->s_umount_key#13/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#25/1 lock#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#13/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) irq_context: 0 &type->s_umount_key#13/1 &____s->seqcount irq_context: 0 &type->s_umount_key#13/1 &sb->s_type->i_lock_key#12 &dentry->d_lock irq_context: 0 &type->s_umount_key#13/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#13/1 bit_wait_table + i irq_context: 0 &type->s_umount_key#13/1 bit_wait_table + i &p->pi_lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#4 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#5 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) irq_context: 0 &sb->s_type->i_mutex_key#5 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 irq_context: 0 &sb->s_type->i_mutex_key#5 &sb->s_type->i_lock_key#12 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#5 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#5 fs_reclaim irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#5 pool_lock#2 irq_context: 0 &type->s_umount_key#25/1 &x->wait#23 irq_context: 0 &sb->s_type->i_mutex_key#5 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#5 &dentry->d_lock &wq irq_context: 0 &type->s_umount_key#26/1 irq_context: 0 &sb->s_type->i_mutex_key#5 &____s->seqcount irq_context: 0 &type->s_umount_key#26/1 fs_reclaim irq_context: 0 &type->s_umount_key#26/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#5 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#5 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#26/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#5 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#26/1 &c->lock irq_context: 0 &type->s_umount_key#26/1 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#5 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#26/1 shrinker_mutex irq_context: 0 &sb->s_type->i_mutex_key#5 &sb->s_type->i_lock_key#12 &dentry->d_lock irq_context: 0 &type->s_umount_key#26/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#26/1 sb_lock irq_context: 0 &type->s_umount_key#26/1 inode_hash_lock irq_context: 0 &type->s_umount_key#26/1 inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#15/1 irq_context: 0 &type->s_umount_key#26/1 bdev_lock irq_context: 0 &type->s_umount_key#15/1 fs_reclaim irq_context: 0 &type->s_umount_key#26/1 &disk->open_mutex irq_context: 0 &type->s_umount_key#15/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26/1 &disk->open_mutex bdev_lock irq_context: 0 &type->s_umount_key#15/1 pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 &disk->open_mutex bdev_lock &bdev->bd_holder_lock irq_context: 0 &type->s_umount_key#15/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#26/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#15/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#26/1 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#26/1 lock#4 irq_context: 0 &type->s_umount_key#15/1 shrinker_mutex irq_context: 0 &type->s_umount_key#26/1 &mapping->i_private_lock irq_context: 0 &type->s_umount_key#15/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#15/1 sb_lock irq_context: 0 &type->s_umount_key#26/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#15/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#26/1 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#15/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26/1 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#26/1 bit_wait_table + i irq_context: 0 &type->s_umount_key#26/1 &rq->__lock irq_context: 0 &type->s_umount_key#15/1 &sb->s_type->i_lock_key#13 irq_context: 0 &type->s_umount_key#26/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#15/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#15/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#26/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#15/1 &sb->s_type->i_lock_key#13 &dentry->d_lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_lock_key#3 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#15/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#26/1 lock#4 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#26/1 lock#5 irq_context: 0 &type->s_umount_key#26/1 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#26/1 crypto_alg_sem irq_context: 0 &type->s_umount_key#26/1 pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 &xa->xa_lock#8 pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 percpu_counters_lock irq_context: 0 &type->s_umount_key#26/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#26/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#26/1 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 &type->s_umount_key#26/1 inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#26/1 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 rcu_read_lock &retval->lock irq_context: 0 &type->s_umount_key#16/1 irq_context: 0 &type->s_umount_key#26/1 rcu_read_lock &____s->seqcount irq_context: 0 &type->s_umount_key#16/1 fs_reclaim irq_context: 0 &type->s_umount_key#16/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#16/1 &c->lock irq_context: 0 &type->s_umount_key#16/1 &____s->seqcount irq_context: hardirq &retval->lock irq_context: 0 &type->s_umount_key#16/1 pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_lock_key#22 irq_context: 0 &type->s_umount_key#16/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 irq_context: 0 &type->s_umount_key#16/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 &type->s_umount_key#16/1 shrinker_mutex irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem irq_context: 0 &type->s_umount_key#16/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#16/1 sb_lock irq_context: 0 &type->s_umount_key#16/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#16/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem pool_lock#2 irq_context: 0 &type->s_umount_key#16/1 &sb->s_type->i_lock_key#14 irq_context: 0 &type->s_umount_key#16/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#16/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#16/1 &sb->s_type->i_lock_key#14 &dentry->d_lock irq_context: 0 &type->s_umount_key#16/1 &dentry->d_lock irq_context: 0 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#17/1 irq_context: 0 &type->s_umount_key#17/1 fs_reclaim irq_context: 0 &type->s_umount_key#17/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#17/1 pool_lock#2 irq_context: 0 &type->s_umount_key#17/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#17/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#17/1 &c->lock irq_context: 0 &type->s_umount_key#17/1 &____s->seqcount irq_context: 0 &type->s_umount_key#17/1 shrinker_mutex irq_context: 0 &type->s_umount_key#17/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#17/1 sb_lock irq_context: 0 &type->s_umount_key#17/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#17/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#17/1 &sb->s_type->i_lock_key#15 irq_context: 0 &type->s_umount_key#17/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#17/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#17/1 &sb->s_type->i_lock_key#15 &dentry->d_lock irq_context: 0 &type->s_umount_key#17/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#15 irq_context: 0 kclist_lock irq_context: 0 kclist_lock resource_lock irq_context: 0 kclist_lock fs_reclaim irq_context: 0 kclist_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kclist_lock pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &type->s_umount_key#18/1 irq_context: 0 &type->s_umount_key#18/1 fs_reclaim irq_context: 0 &type->s_umount_key#18/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#18/1 pool_lock#2 irq_context: 0 &type->s_umount_key#18/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#18/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#18/1 shrinker_mutex irq_context: 0 &type->s_umount_key#18/1 &c->lock irq_context: 0 &type->s_umount_key#18/1 &____s->seqcount irq_context: 0 &type->s_umount_key#18/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#18/1 sb_lock irq_context: 0 &type->s_umount_key#18/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#18/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#18/1 &sb->s_type->i_lock_key#16 irq_context: 0 &type->s_umount_key#18/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#18/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#18/1 &sb->s_type->i_lock_key#16 &dentry->d_lock irq_context: 0 &type->s_umount_key#18/1 &dentry->d_lock irq_context: 0 misc_mtx &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &x->wait#11 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#11 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#2 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#2 &____s->seqcount irq_context: 0 tomoyo_ss irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss pool_lock#2 irq_context: 0 tomoyo_ss tomoyo_policy_lock irq_context: 0 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 tomoyo_ss (console_sem).lock irq_context: 0 tomoyo_ss console_lock console_srcu console_owner_lock irq_context: 0 tomoyo_ss console_lock console_srcu console_owner irq_context: 0 tomoyo_ss console_lock console_srcu console_owner &port_lock_key irq_context: 0 tomoyo_ss console_lock console_srcu console_owner console_owner_lock irq_context: hardirq &rt_b->rt_runtime_lock irq_context: hardirq &rt_b->rt_runtime_lock tk_core.seq.seqcount irq_context: hardirq &rt_rq->rt_runtime_lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 pnp_lock irq_context: 0 pnp_lock fs_reclaim irq_context: 0 pnp_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pnp_lock pool_lock#2 irq_context: 0 &device->physical_node_lock sysfs_symlink_target_lock irq_context: 0 &device->physical_node_lock fs_reclaim irq_context: 0 &device->physical_node_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &device->physical_node_lock pool_lock#2 irq_context: 0 &device->physical_node_lock lock irq_context: 0 &device->physical_node_lock lock kernfs_idr_lock irq_context: 0 &device->physical_node_lock &root->kernfs_rwsem irq_context: 0 &device->physical_node_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 fwnode_link_lock irq_context: 0 fwnode_link_lock &k->k_lock irq_context: 0 &dev->mutex device_links_srcu irq_context: 0 &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex device_links_lock irq_context: 0 &dev->mutex fs_reclaim irq_context: 0 &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex &dev->devres_lock irq_context: 0 &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex pinctrl_list_mutex irq_context: 0 &dev->mutex pinctrl_maps_mutex irq_context: 0 &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex lock irq_context: 0 &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex pnp_lock irq_context: 0 &dev->mutex resource_lock irq_context: 0 &dev->mutex (console_sem).lock irq_context: 0 &dev->mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex console_lock console_srcu console_owner irq_context: 0 &dev->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &dev->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex fwnode_link_lock &k->k_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex probe_waitqueue.lock irq_context: 0 subsys mutex#22 irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &type->s_umount_key#26/1 proc_subdir_lock irq_context: 0 &type->s_umount_key#26/1 proc_inum_ida.xa_lock irq_context: 0 &type->s_umount_key#26/1 proc_subdir_lock irq_context: 0 &type->s_umount_key#26/1 &journal->j_state_lock irq_context: softirq (&net->can.stattimer) irq_context: softirq (&net->can.stattimer) &obj_hash[i].lock irq_context: softirq (&net->can.stattimer) &base->lock irq_context: softirq (&net->can.stattimer) &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#26/1 kthread_create_lock irq_context: 0 &type->s_umount_key#26/1 &p->pi_lock irq_context: 0 &type->s_umount_key#26/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#26/1 &x->wait irq_context: 0 &type->s_umount_key#26/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &journal->j_wait_done_commit irq_context: 0 &type->s_umount_key#26/1 &journal->j_wait_done_commit irq_context: 0 &journal->j_state_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_done_commit irq_context: 0 subsys mutex#23 irq_context: 0 subsys mutex#23 &k->k_lock irq_context: 0 subsys mutex#16 irq_context: 0 subsys mutex#16 &k->k_lock irq_context: 0 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 fill_pool_map-wait-type-override pool_lock irq_context: 0 subsys mutex#24 irq_context: 0 subsys mutex#24 &k->k_lock irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 fill_pool_map-wait-type-override &c->lock irq_context: 0 fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq led_lock irq_context: 0 misc_mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 subsys mutex#25 irq_context: 0 subsys mutex#25 &k->list_lock irq_context: 0 subsys mutex#25 &k->k_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &xa->xa_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex kthread_create_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &p->pi_lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &x->wait irq_context: 0 cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock wq_pool_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &p->pi_lock &rq->__lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex wq_pool_attach_mutex irq_context: 0 cpu_hotplug_lock wq_pool_mutex &pool->lock &p->pi_lock irq_context: 0 netevent_notif_chain.lock irq_context: 0 clients_rwsem irq_context: 0 clients_rwsem fs_reclaim irq_context: 0 clients_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 clients_rwsem clients.xa_lock irq_context: 0 devices_rwsem irq_context: softirq rcu_callback pcpu_lock irq_context: 0 clients_rwsem clients.xa_lock pool_lock#2 irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (blocking_lsm_notifier_chain).rwsem irq_context: 0 (inetaddr_chain).rwsem irq_context: 0 inet6addr_chain.lock irq_context: 0 buses_mutex irq_context: 0 offload_lock irq_context: 0 inetsw_lock irq_context: 0 (wq_completion)events pcpu_balance_work irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex fs_reclaim irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex pool_lock#2 irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex free_vmap_area_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex vmap_area_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &____s->seqcount irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex init_mm.page_table_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &c->lock irq_context: 0 ptype_lock irq_context: 0 (wq_completion)events_power_efficient irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &tbl->lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &tbl->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work irq_context: 0 pernet_ops_rwsem &net->rules_mod_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &base->lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem slab_mutex irq_context: 0 pernet_ops_rwsem slab_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem slab_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem slab_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock pool_lock irq_context: 0 tcp_ulp_list_lock irq_context: 0 xfrm_state_afinfo_lock irq_context: 0 xfrm_policy_afinfo_lock irq_context: 0 xfrm_input_afinfo_lock irq_context: 0 pernet_ops_rwsem percpu_counters_lock irq_context: 0 rtnl_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex krc.lock irq_context: 0 rtnl_mutex krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex krc.lock hrtimer_bases.lock irq_context: 0 rtnl_mutex krc.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex krc.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex krc.lock &base->lock irq_context: 0 rtnl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem k-slock-AF_INET/1 irq_context: 0 (wq_completion)events_highpri irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) fs_reclaim irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) &____s->seqcount irq_context: 0 (wq_completion)events_highpri (work_completion)(&(&krcp->page_cache_work)->work) krc.lock irq_context: 0 &hashinfo->lock irq_context: 0 k-slock-AF_INET/1 irq_context: 0 rcu_read_lock &c->lock irq_context: 0 rcu_read_lock &____s->seqcount irq_context: 0 tcp_cong_list_lock irq_context: 0 mptcp_sched_list_lock irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem cache_list_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) cache_list_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (rpc_pipefs_notifier_list).rwsem irq_context: 0 svc_xprt_class_lock irq_context: 0 xprt_list_lock irq_context: 0 xprt_list_lock (console_sem).lock irq_context: 0 xprt_list_lock console_lock console_srcu console_owner_lock irq_context: 0 xprt_list_lock console_lock console_srcu console_owner irq_context: 0 xprt_list_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 xprt_list_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 umhelper_sem irq_context: 0 umhelper_sem usermodehelper_disabled_waitq.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock init_fs.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_mutex_key &dentry->d_lock &wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 mount_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &s->s_inode_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_log_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_log_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_policy_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key &sb->s_type->i_lock_key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key &wb->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key &wb->list_lock &sb->s_type->i_lock_key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &sb->s_type->i_lock_key#2 irq_context: 0 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &sighand->siglock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) free_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_mm.page_table_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u64.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_files.file_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_fs.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) init_fs.lock &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->alloc_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &sighand->siglock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) pidmap_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem css_set_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem tasklist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem tasklist_lock &sighand->siglock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) input_pool.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key tomoyo_ss &rq->__lock irq_context: 0 umh_sysctl_lock irq_context: 0 async_done.lock irq_context: 0 &drv->dynids.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_done.lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_done.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) async_done.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key irq_context: 0 &sb->s_type->i_mutex_key fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &x->wait#6 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock &wq irq_context: 0 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &dentry->d_lock pool_lock#2 irq_context: 0 &tsk->futex_exit_mutex irq_context: 0 &tsk->futex_exit_mutex &p->pi_lock irq_context: 0 &p->alloc_lock &fs->lock irq_context: 0 &child->perf_event_mutex irq_context: 0 css_set_lock irq_context: 0 tasklist_lock irq_context: 0 tasklist_lock &pid->wait_pidfd irq_context: 0 tasklist_lock &sighand->siglock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit irq_context: 0 tasklist_lock &sighand->siglock input_pool.lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override pool_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 &obj_hash[i].lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pool_lock#2 irq_context: 0 tasklist_lock &obj_hash[i].lock irq_context: 0 &cfs_rq->removed.lock irq_context: 0 misc_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 misc_mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 subsys mutex#26 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u64.lock crngs.lock irq_context: 0 subsys mutex#27 irq_context: 0 subsys mutex#27 &k->list_lock irq_context: 0 subsys mutex#27 &k->k_lock irq_context: 0 &sb->s_type->i_mutex_key &c->lock irq_context: 0 &sb->s_type->i_mutex_key &____s->seqcount irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock pool_lock#2 irq_context: 0 subsys mutex#28 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock &obj_hash[i].lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pmus_lock fs_reclaim irq_context: 0 pmus_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pmus_lock &k->list_lock irq_context: 0 pmus_lock lock irq_context: 0 pmus_lock lock kernfs_idr_lock irq_context: 0 pmus_lock &root->kernfs_rwsem irq_context: 0 pmus_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pmus_lock uevent_sock_mutex irq_context: 0 pmus_lock rcu_read_lock &pool->lock irq_context: 0 pmus_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pmus_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pmus_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pmus_lock running_helpers_waitq.lock irq_context: 0 pmus_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pmus_lock &x->wait#9 irq_context: 0 pmus_lock bus_type_sem irq_context: 0 pmus_lock sysfs_symlink_target_lock irq_context: 0 pmus_lock &k->k_lock irq_context: 0 pmus_lock &c->lock irq_context: 0 pmus_lock &____s->seqcount irq_context: 0 pmus_lock &root->kernfs_rwsem irq_context: 0 pmus_lock &dev->power.lock irq_context: 0 pmus_lock dpm_list_mtx irq_context: 0 pmus_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pmus_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pmus_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pmus_lock &dev->mutex &k->list_lock irq_context: 0 pmus_lock &dev->mutex &k->k_lock irq_context: 0 pmus_lock &dev->mutex &dev->power.lock irq_context: 0 pmus_lock subsys mutex#29 irq_context: 0 pmus_lock &pcp->lock &zone->lock irq_context: 0 pmus_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pmus_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pmus_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pmus_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pmus_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock &obj_hash[i].lock pool_lock irq_context: 0 key_user_lock irq_context: 0 key_serial_lock irq_context: 0 key_construction_mutex irq_context: 0 &type->lock_class irq_context: 0 &type->lock_class keyring_serialise_link_lock irq_context: 0 &type->lock_class keyring_serialise_link_lock fs_reclaim irq_context: 0 &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->lock_class keyring_serialise_link_lock pool_lock#2 irq_context: 0 &type->lock_class keyring_serialise_link_lock &obj_hash[i].lock irq_context: 0 keyring_serialise_link_lock irq_context: 0 &pgdat->kswapd_lock fs_reclaim irq_context: 0 &pgdat->kswapd_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pgdat->kswapd_lock pool_lock#2 irq_context: 0 &pgdat->kswapd_lock kthread_create_lock irq_context: 0 &pgdat->kswapd_lock &p->pi_lock irq_context: 0 &pgdat->kswapd_lock &x->wait irq_context: 0 &pgdat->kswapd_lock &rq->__lock irq_context: 0 &pgdat->kswapd_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pgdat->kswapd_lock &obj_hash[i].lock irq_context: 0 &pgdat->kswapd_wait irq_context: 0 list_lrus_mutex irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex fs_reclaim irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex pool_lock#2 irq_context: 0 drivers_lock irq_context: 0 damon_dbgfs_lock irq_context: 0 damon_dbgfs_lock fs_reclaim irq_context: 0 damon_dbgfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 damon_dbgfs_lock pool_lock#2 irq_context: 0 damon_dbgfs_lock damon_ops_lock irq_context: 0 damon_dbgfs_lock pin_fs_lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 damon_dbgfs_lock &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 misc_mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#19/1 irq_context: 0 &type->s_umount_key#19/1 fs_reclaim irq_context: 0 &type->s_umount_key#19/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#19/1 pool_lock#2 irq_context: 0 &type->s_umount_key#19/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#19/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#19/1 shrinker_mutex irq_context: 0 &type->s_umount_key#19/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#19/1 sb_lock irq_context: 0 &type->s_umount_key#19/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#19/1 &c->lock irq_context: 0 &type->s_umount_key#19/1 &____s->seqcount irq_context: 0 &type->s_umount_key#19/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#19/1 &sb->s_type->i_lock_key#17 irq_context: 0 &type->s_umount_key#19/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#19/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#19/1 &sb->s_type->i_lock_key#17 &dentry->d_lock irq_context: 0 &type->s_umount_key#19/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#20/1 irq_context: 0 &type->s_umount_key#20/1 fs_reclaim irq_context: 0 &type->s_umount_key#20/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#20/1 pool_lock#2 irq_context: 0 &type->s_umount_key#20/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#20/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#20/1 &c->lock irq_context: 0 &type->s_umount_key#20/1 &____s->seqcount irq_context: 0 &type->s_umount_key#20/1 shrinker_mutex irq_context: 0 &type->s_umount_key#20/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#20/1 sb_lock irq_context: 0 &type->s_umount_key#20/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#20/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#20/1 &sb->s_type->i_lock_key#18 irq_context: 0 &type->s_umount_key#20/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#20/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#20/1 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &type->s_umount_key#20/1 &dentry->d_lock irq_context: 0 configfs_subsystem_mutex irq_context: 0 &sb->s_type->i_mutex_key#6/1 irq_context: 0 &sb->s_type->i_mutex_key#6/1 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &sb->s_type->i_lock_key#18 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 irq_context: 0 &x->wait#11 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 slab_mutex rcu_read_lock &pool->lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 slab_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ecryptfs_kthread_ctl.wait irq_context: 0 ecryptfs_daemon_hash_mux irq_context: 0 ecryptfs_daemon_hash_mux fs_reclaim irq_context: 0 ecryptfs_daemon_hash_mux fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ecryptfs_daemon_hash_mux &c->lock irq_context: 0 ecryptfs_daemon_hash_mux &____s->seqcount irq_context: 0 ecryptfs_daemon_hash_mux pool_lock#2 irq_context: 0 ecryptfs_msg_ctx_lists_mux irq_context: 0 ecryptfs_msg_ctx_lists_mux &ecryptfs_msg_ctx_arr[i].mux irq_context: 0 pernet_ops_rwsem tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem &k->list_lock irq_context: 0 pernet_ops_rwsem lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem running_helpers_waitq.lock irq_context: 0 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfs_version_lock irq_context: 0 key_types_sem irq_context: 0 key_types_sem (console_sem).lock irq_context: 0 key_types_sem console_lock console_srcu console_owner_lock irq_context: 0 key_types_sem console_lock console_srcu console_owner irq_context: 0 key_types_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 key_types_sem console_lock console_srcu console_owner console_owner_lock irq_context: softirq rcu_callback put_task_map-wait-type-override &obj_hash[i].lock irq_context: softirq rcu_callback put_task_map-wait-type-override pool_lock#2 irq_context: 0 pnfs_spinlock irq_context: 0 pernet_ops_rwsem &sn->pipefs_sb_lock irq_context: 0 pernet_ops_rwsem krc.lock irq_context: 0 pernet_ops_rwsem krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem krc.lock hrtimer_bases.lock irq_context: 0 pernet_ops_rwsem krc.lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem krc.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem krc.lock &base->lock irq_context: 0 pernet_ops_rwsem krc.lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem &s->s_inode_list_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 nls_lock irq_context: softirq &(&cache_cleaner)->timer irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_callback put_task_map-wait-type-override &obj_hash[i].lock pool_lock irq_context: 0 jffs2_compressor_list_lock irq_context: 0 next_tag_value_lock irq_context: softirq (&tcp_orphan_timer) irq_context: softirq (&tcp_orphan_timer) &obj_hash[i].lock irq_context: softirq (&tcp_orphan_timer) &base->lock irq_context: softirq (&tcp_orphan_timer) &base->lock &obj_hash[i].lock irq_context: 0 log_redrive_lock irq_context: 0 &TxAnchor.LazyLock irq_context: 0 &TxAnchor.LazyLock jfs_commit_thread_wait.lock irq_context: 0 jfsTxnLock irq_context: 0 ocfs2_stack_lock irq_context: 0 ocfs2_stack_lock (console_sem).lock irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner_lock irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 ocfs2_stack_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 o2hb_callback_sem irq_context: 0 o2net_handler_lock irq_context: 0 slab_mutex &rq->__lock irq_context: 0 subsys mutex#30 irq_context: 0 subsys mutex#30 &k->k_lock irq_context: 0 &type->s_umount_key#21/1 irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#21/1 fs_reclaim irq_context: 0 &type->s_umount_key#21/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#21/1 pool_lock#2 irq_context: 0 &type->s_umount_key#21/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#21/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#21/1 shrinker_mutex irq_context: 0 &type->s_umount_key#21/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#21/1 sb_lock irq_context: 0 &type->s_umount_key#21/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#21/1 &____s->seqcount irq_context: 0 &type->s_umount_key#21/1 &c->lock irq_context: 0 &type->s_umount_key#21/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#21/1 &sb->s_type->i_lock_key#19 irq_context: 0 &type->s_umount_key#21/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#21/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#21/1 &sb->s_type->i_lock_key#19 &dentry->d_lock irq_context: 0 &type->s_umount_key#21/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#22/1 irq_context: 0 &type->s_umount_key#22/1 fs_reclaim irq_context: 0 &type->s_umount_key#22/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#22/1 pool_lock#2 irq_context: 0 &type->s_umount_key#22/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#22/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#22/1 shrinker_mutex irq_context: 0 &type->s_umount_key#22/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#22/1 sb_lock irq_context: 0 &type->s_umount_key#22/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#22/1 &c->lock irq_context: 0 &type->s_umount_key#22/1 &____s->seqcount irq_context: 0 &type->s_umount_key#22/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#22/1 &sb->s_type->i_lock_key#20 irq_context: 0 &type->s_umount_key#22/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#22/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#22/1 &sb->s_type->i_lock_key#20 &dentry->d_lock irq_context: 0 &type->s_umount_key#22/1 &dentry->d_lock irq_context: 0 cipso_v4_doi_list_lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 crypto_alg_sem irq_context: 0 alg_types_sem irq_context: 0 alg_types_sem fs_reclaim irq_context: 0 alg_types_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 alg_types_sem pool_lock#2 irq_context: 0 dma_list_mutex irq_context: 0 asymmetric_key_parsers_sem irq_context: 0 asymmetric_key_parsers_sem (console_sem).lock irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner_lock irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 asymmetric_key_parsers_sem console_lock console_srcu console_owner console_owner_lock irq_context: 0 blkcg_pol_register_mutex irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex irq_context: 0 blkcg_pol_register_mutex cgroup_mutex irq_context: 0 blkcg_pol_register_mutex cgroup_mutex &root->kernfs_rwsem irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex fs_reclaim irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 blkcg_pol_register_mutex blkcg_pol_mutex pool_lock#2 irq_context: 0 blkcg_pol_register_mutex cgroup_mutex fs_reclaim irq_context: 0 blkcg_pol_register_mutex cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 blkcg_pol_register_mutex cgroup_mutex pool_lock#2 irq_context: 0 blkcg_pol_register_mutex cgroup_mutex &c->lock irq_context: 0 blkcg_pol_register_mutex cgroup_mutex &____s->seqcount irq_context: 0 blkcg_pol_register_mutex cgroup_mutex lock irq_context: 0 blkcg_pol_register_mutex cgroup_mutex lock kernfs_idr_lock irq_context: 0 blkcg_pol_register_mutex cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 elv_list_lock irq_context: 0 crc_t10dif_mutex irq_context: 0 crc_t10dif_mutex crypto_alg_sem irq_context: 0 crc_t10dif_mutex fs_reclaim irq_context: 0 crc_t10dif_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 crc_t10dif_mutex pool_lock#2 irq_context: 0 crc64_rocksoft_mutex irq_context: 0 crc64_rocksoft_mutex crypto_alg_sem irq_context: 0 crc64_rocksoft_mutex fs_reclaim irq_context: 0 crc64_rocksoft_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 crc64_rocksoft_mutex pool_lock#2 irq_context: 0 ts_mod_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) kfence_freelist_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 &obj_hash[i].lock pool_lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_callback put_task_map-wait-type-override fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_callback put_task_map-wait-type-override fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rq->__lock irq_context: 0 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &c->lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &x->wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 tasklist_lock quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &cfs_rq->removed.lock irq_context: 0 fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) irq_context: 0 (wq_completion)events (work_completion)(&p->wq) vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) pool_lock#2 irq_context: softirq rcu_callback put_task_map-wait-type-override quarantine_lock irq_context: softirq rcu_callback &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &obj_hash[i].lock pool_lock irq_context: softirq rcu_callback &meta->lock irq_context: softirq rcu_callback kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pci_ep_cfs_subsys.su_mutex irq_context: 0 &default_group_class[depth - 1]#2/1 irq_context: 0 &default_group_class[depth - 1]#2/1 fs_reclaim irq_context: 0 &default_group_class[depth - 1]#2/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &default_group_class[depth - 1]#2/1 pool_lock#2 irq_context: 0 &default_group_class[depth - 1]#2/1 &dentry->d_lock irq_context: 0 &default_group_class[depth - 1]#2/1 configfs_dirent_lock irq_context: 0 &default_group_class[depth - 1]#2/1 &____s->seqcount irq_context: 0 &default_group_class[depth - 1]#2/1 mmu_notifier_invalidate_range_start irq_context: 0 &default_group_class[depth - 1]#2/1 &sb->s_type->i_lock_key#18 irq_context: 0 &default_group_class[depth - 1]#2/1 &s->s_inode_list_lock irq_context: 0 &default_group_class[depth - 1]#2/1 tk_core.seq.seqcount irq_context: 0 &default_group_class[depth - 1]#2/1 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &default_group_class[depth - 1]#2/1 &sb->s_type->i_mutex_key#7/2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &sb->s_type->i_mutex_key#7/2 irq_context: 0 pci_epf_mutex irq_context: 0 ipmi_interfaces_mutex irq_context: 0 ipmi_interfaces_mutex &k->list_lock irq_context: 0 ipmi_interfaces_mutex fs_reclaim irq_context: 0 ipmi_interfaces_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ipmi_interfaces_mutex pool_lock#2 irq_context: 0 ipmi_interfaces_mutex lock irq_context: 0 ipmi_interfaces_mutex lock kernfs_idr_lock irq_context: 0 ipmi_interfaces_mutex &root->kernfs_rwsem irq_context: 0 ipmi_interfaces_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 ipmi_interfaces_mutex &k->k_lock irq_context: 0 ipmi_interfaces_mutex &c->lock irq_context: 0 ipmi_interfaces_mutex &____s->seqcount irq_context: 0 ipmi_interfaces_mutex uevent_sock_mutex irq_context: 0 ipmi_interfaces_mutex &obj_hash[i].lock irq_context: 0 ipmi_interfaces_mutex rcu_read_lock &pool->lock irq_context: 0 ipmi_interfaces_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 ipmi_interfaces_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 ipmi_interfaces_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 ipmi_interfaces_mutex running_helpers_waitq.lock irq_context: 0 ipmi_interfaces_mutex pcpu_alloc_mutex irq_context: 0 ipmi_interfaces_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 ipmi_interfaces_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 ipmi_interfaces_mutex kthread_create_lock irq_context: 0 ipmi_interfaces_mutex &p->pi_lock irq_context: 0 ipmi_interfaces_mutex &p->pi_lock &rq->__lock irq_context: 0 ipmi_interfaces_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ipmi_interfaces_mutex &rq->__lock irq_context: 0 ipmi_interfaces_mutex &x->wait irq_context: 0 ipmi_interfaces_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ipmi_interfaces_mutex wq_pool_mutex irq_context: 0 ipmi_interfaces_mutex wq_pool_mutex &wq->mutex irq_context: 0 ipmi_interfaces_mutex &base->lock irq_context: 0 ipmi_interfaces_mutex &base->lock &obj_hash[i].lock irq_context: 0 ipmi_interfaces_mutex panic_notifier_list.lock irq_context: 0 smi_watchers_mutex irq_context: 0 smi_watchers_mutex &ipmi_interfaces_srcu irq_context: 0 smi_infos_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex &device->physical_node_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock semaphore->lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock fs_reclaim irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock pool_lock#2 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock pool_lock#2 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock wakeup_ida.xa_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &x->wait#9 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &obj_hash[i].lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->list_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &k->list_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock bus_type_sem irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock sysfs_symlink_target_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &c->lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &____s->seqcount irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock uevent_sock_mutex irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock running_helpers_waitq.lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock &k->k_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock subsys mutex#13 &k->k_lock irq_context: 0 &dev->mutex acpi_pm_notifier_install_lock acpi_pm_notifier_lock events_lock irq_context: 0 &dev->mutex acpi_wakeup_lock irq_context: 0 &dev->mutex semaphore->lock irq_context: 0 &dev->mutex *(&acpi_gbl_reference_count_lock) irq_context: 0 &dev->mutex irq_domain_mutex irq_context: 0 &dev->mutex &domain->mutex irq_context: 0 &dev->mutex kthread_create_lock irq_context: 0 &dev->mutex &p->pi_lock irq_context: 0 &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &rq->__lock irq_context: 0 &dev->mutex &x->wait irq_context: 0 &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &desc->request_mutex irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class irq_controller_lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 &dev->mutex &desc->wait_for_threads irq_context: 0 &desc->wait_for_threads irq_context: 0 &desc->wait_for_threads &p->pi_lock irq_context: 0 &p->pi_lock &rq->__lock &rt_b->rt_runtime_lock irq_context: 0 &p->pi_lock &rq->__lock &rt_b->rt_runtime_lock tk_core.seq.seqcount irq_context: 0 &dev->mutex register_lock irq_context: 0 &p->pi_lock &rq->__lock &rt_b->rt_runtime_lock hrtimer_bases.lock irq_context: 0 &dev->mutex register_lock proc_subdir_lock irq_context: 0 &p->pi_lock &rq->__lock &rt_b->rt_runtime_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &dev->mutex register_lock fs_reclaim irq_context: 0 &dev->mutex register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->pi_lock &rq->__lock &rt_rq->rt_runtime_lock irq_context: 0 &dev->mutex register_lock pool_lock#2 irq_context: 0 &dev->mutex register_lock proc_inum_ida.xa_lock irq_context: 0 &dev->mutex register_lock proc_subdir_lock irq_context: 0 &x->wait#7 irq_context: 0 &dev->mutex register_lock &rq->__lock irq_context: 0 &dev->mutex register_lock &c->lock irq_context: 0 &dev->mutex register_lock &____s->seqcount irq_context: 0 &dev->mutex &irq_desc_lock_class irq_context: 0 &dev->mutex proc_subdir_lock irq_context: 0 &dev->mutex proc_inum_ida.xa_lock irq_context: 0 &dev->mutex proc_subdir_lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &n->list_lock irq_context: 0 &dev->mutex &c->lock irq_context: 0 &dev->mutex &x->wait#9 irq_context: 0 &dev->mutex gdp_mutex irq_context: 0 &dev->mutex gdp_mutex &k->list_lock irq_context: 0 &dev->mutex gdp_mutex fs_reclaim irq_context: 0 &dev->mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex gdp_mutex lock irq_context: 0 &dev->mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex bus_type_sem irq_context: 0 &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex dpm_list_mtx irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex subsys mutex#31 irq_context: 0 &dev->mutex subsys mutex#31 &k->k_lock irq_context: 0 &dev->mutex input_mutex irq_context: 0 &dev->mutex input_mutex fs_reclaim irq_context: 0 &dev->mutex input_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex input_mutex pool_lock#2 irq_context: 0 &dev->mutex input_mutex &dev->mutex#2 irq_context: 0 &dev->mutex input_mutex input_devices_poll_wait.lock irq_context: 0 &dev->mutex wakeup_ida.xa_lock irq_context: 0 &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &dev->mutex subsys mutex#13 irq_context: 0 &dev->mutex subsys mutex#13 &k->k_lock irq_context: 0 &dev->mutex events_lock irq_context: softirq rcu_callback quarantine_lock irq_context: 0 register_count_mutex irq_context: 0 register_count_mutex &k->list_lock irq_context: 0 register_count_mutex fs_reclaim irq_context: 0 register_count_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_count_mutex pool_lock#2 irq_context: 0 register_count_mutex lock irq_context: 0 register_count_mutex lock kernfs_idr_lock irq_context: 0 register_count_mutex &root->kernfs_rwsem irq_context: 0 register_count_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 register_count_mutex &k->k_lock irq_context: 0 register_count_mutex &c->lock irq_context: 0 register_count_mutex &____s->seqcount irq_context: 0 register_count_mutex uevent_sock_mutex irq_context: 0 register_count_mutex &obj_hash[i].lock irq_context: 0 register_count_mutex rcu_read_lock &pool->lock irq_context: 0 register_count_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 register_count_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 register_count_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 register_count_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex cpu_add_remove_lock irq_context: 0 &dev->mutex thermal_cdev_ida.xa_lock irq_context: 0 &dev->mutex cpufreq_driver_lock irq_context: 0 &dev->mutex subsys mutex#32 irq_context: 0 &dev->mutex subsys mutex#32 &k->k_lock irq_context: 0 &dev->mutex thermal_list_lock irq_context: 0 &dev->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) uevent_sock_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) running_helpers_waitq.lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_callback &base->lock irq_context: softirq rcu_callback &base->lock &obj_hash[i].lock irq_context: 0 scmi_requested_devices_mtx irq_context: 0 scmi_requested_devices_mtx fs_reclaim irq_context: 0 scmi_requested_devices_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 scmi_requested_devices_mtx pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key rcu_read_lock &dentry->d_lock irq_context: 0 &dev->mutex iommu_probe_device_lock irq_context: 0 &dev->mutex quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock pool_lock#2 irq_context: 0 &dev->mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex acpi_link_lock irq_context: 0 &dev->mutex acpi_link_lock fs_reclaim irq_context: 0 &dev->mutex acpi_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex acpi_link_lock pool_lock#2 irq_context: 0 &dev->mutex acpi_link_lock &____s->seqcount irq_context: 0 &dev->mutex acpi_link_lock semaphore->lock irq_context: 0 &dev->mutex acpi_link_lock &obj_hash[i].lock irq_context: 0 &dev->mutex acpi_link_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 &dev->mutex acpi_link_lock (console_sem).lock irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex acpi_link_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex &drv->dynids.lock irq_context: 0 &dev->mutex pci_lock irq_context: softirq rcu_callback put_task_map-wait-type-override &meta->lock irq_context: softirq rcu_callback put_task_map-wait-type-override kfence_freelist_lock irq_context: 0 &dev->mutex virtio_index_ida.xa_lock irq_context: 0 &dev->mutex &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex &dev->mutex &k->list_lock irq_context: 0 &dev->mutex &dev->mutex &k->k_lock irq_context: 0 &dev->mutex subsys mutex#33 irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rq->__lock irq_context: 0 vdpa_dev_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#2 irq_context: 0 &type->i_mutex_dir_key#2 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#2 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#2 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 &dentry->d_lock &wq irq_context: 0 &x->wait#11 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 subsys mutex#34 irq_context: 0 subsys mutex#34 &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) quarantine_lock irq_context: 0 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq lib/debugobjects.c:101 irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (debug_obj_work).work irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_node_0 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 quarantine_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rcu_read_lock rcu_node_0 irq_context: 0 port_mutex irq_context: 0 port_mutex fs_reclaim irq_context: 0 port_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 port_mutex &c->lock irq_context: 0 port_mutex &____s->seqcount irq_context: 0 port_mutex pool_lock#2 irq_context: 0 port_mutex &x->wait#9 irq_context: 0 port_mutex &obj_hash[i].lock irq_context: 0 port_mutex &k->list_lock irq_context: 0 port_mutex lock irq_context: 0 port_mutex lock kernfs_idr_lock irq_context: 0 port_mutex &root->kernfs_rwsem irq_context: 0 port_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 port_mutex bus_type_sem irq_context: 0 port_mutex sysfs_symlink_target_lock irq_context: 0 port_mutex &k->k_lock irq_context: 0 port_mutex &root->kernfs_rwsem irq_context: 0 port_mutex &dev->power.lock irq_context: 0 port_mutex dpm_list_mtx irq_context: 0 port_mutex uevent_sock_mutex irq_context: 0 port_mutex rcu_read_lock &pool->lock irq_context: 0 port_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 port_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 port_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 port_mutex running_helpers_waitq.lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock irq_context: 0 port_mutex &dev->mutex &k->list_lock irq_context: 0 port_mutex &dev->mutex &k->k_lock irq_context: 0 port_mutex &dev->mutex device_links_srcu irq_context: 0 port_mutex &dev->mutex fwnode_link_lock irq_context: 0 port_mutex &dev->mutex device_links_lock irq_context: 0 port_mutex &dev->mutex fs_reclaim irq_context: 0 port_mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 port_mutex &dev->mutex pool_lock#2 irq_context: 0 port_mutex &dev->mutex &dev->devres_lock irq_context: 0 port_mutex &dev->mutex pinctrl_list_mutex irq_context: 0 port_mutex &dev->mutex pinctrl_maps_mutex irq_context: 0 port_mutex &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 port_mutex &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 port_mutex &dev->mutex &obj_hash[i].lock irq_context: 0 port_mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 port_mutex &dev->mutex lock irq_context: 0 port_mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 port_mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 port_mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 port_mutex &dev->mutex deferred_probe_mutex irq_context: 0 port_mutex &dev->mutex &c->lock irq_context: 0 port_mutex &dev->mutex &____s->seqcount irq_context: 0 port_mutex &dev->mutex uevent_sock_mutex irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 port_mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 port_mutex &dev->mutex probe_waitqueue.lock irq_context: 0 port_mutex subsys mutex#14 irq_context: 0 port_mutex &xa->xa_lock#3 irq_context: 0 port_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 port_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 port_mutex &dev->mutex &pcp->lock &zone->lock irq_context: 0 port_mutex &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 port_mutex &dev->mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex &port->mutex irq_context: 0 port_mutex &port->mutex fs_reclaim irq_context: 0 port_mutex &port->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 port_mutex &port->mutex pool_lock#2 irq_context: 0 port_mutex &port->mutex console_mutex irq_context: 0 port_mutex &port->mutex ctrl_ida.xa_lock irq_context: 0 port_mutex &port->mutex &x->wait#9 irq_context: 0 port_mutex &port->mutex &obj_hash[i].lock irq_context: 0 port_mutex &port->mutex &dev->power.lock irq_context: 0 port_mutex &port->mutex &k->list_lock irq_context: 0 port_mutex &port->mutex lock irq_context: 0 port_mutex &port->mutex lock kernfs_idr_lock irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 port_mutex &port->mutex bus_type_sem irq_context: 0 port_mutex &port->mutex sysfs_symlink_target_lock irq_context: 0 port_mutex &port->mutex &k->k_lock irq_context: 0 port_mutex &port->mutex dpm_list_mtx irq_context: 0 port_mutex &port->mutex uevent_sock_mutex irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex &port->mutex running_helpers_waitq.lock irq_context: 0 port_mutex &port->mutex &dev->mutex &dev->power.lock irq_context: 0 port_mutex &port->mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 port_mutex &port->mutex &dev->mutex &k->list_lock irq_context: 0 port_mutex &port->mutex &dev->mutex &k->k_lock irq_context: 0 port_mutex &port->mutex &dev->mutex &dev->power.lock hrtimer_bases.lock irq_context: 0 port_mutex &port->mutex &dev->mutex &dev->power.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 port_mutex &port->mutex subsys mutex#15 irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem irq_context: 0 port_mutex &port->mutex dev_pm_qos_sysfs_mtx irq_context: 0 port_mutex &port->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 port_mutex &port->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 port_mutex &port->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 port_mutex &port->mutex kernfs_idr_lock irq_context: 0 port_mutex &port->mutex &k->k_lock klist_remove_lock irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 port_mutex &port->mutex deferred_probe_mutex irq_context: 0 port_mutex &port->mutex device_links_lock irq_context: 0 port_mutex &port->mutex mmu_notifier_invalidate_range_start irq_context: 0 port_mutex &port->mutex &obj_hash[i].lock pool_lock irq_context: 0 port_mutex &port->mutex gdp_mutex irq_context: 0 port_mutex &port->mutex gdp_mutex &k->list_lock irq_context: 0 port_mutex &port->mutex gdp_mutex fs_reclaim irq_context: 0 port_mutex &port->mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 port_mutex &port->mutex gdp_mutex pool_lock#2 irq_context: 0 port_mutex &port->mutex gdp_mutex lock irq_context: 0 port_mutex &port->mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 port_mutex &port->mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 port_mutex &port->mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 port_mutex &port->mutex &c->lock irq_context: 0 port_mutex &port->mutex &____s->seqcount irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 port_mutex &port->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex &port->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 port_mutex &port->mutex req_lock irq_context: 0 port_mutex &port->mutex &p->pi_lock irq_context: 0 port_mutex &port->mutex &p->pi_lock &rq->__lock irq_context: 0 port_mutex &port->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex &port->mutex &rq->__lock irq_context: 0 port_mutex &port->mutex &x->wait#11 irq_context: 0 port_mutex &port->mutex subsys mutex#16 irq_context: 0 port_mutex &port->mutex subsys mutex#16 &k->k_lock irq_context: 0 port_mutex &port->mutex chrdevs_lock irq_context: 0 port_mutex &port->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 port_mutex &port->mutex &cfs_rq->removed.lock irq_context: 0 port_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 port_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 port_mutex &port->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 port_mutex &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 fill_pool_map-wait-type-override &obj_hash[i].lock pool_lock irq_context: 0 (work_completion)(&buf->work) irq_context: 0 &dev->mutex rng_index_ida.xa_lock irq_context: 0 &dev->mutex &md->mutex irq_context: 0 &dev->mutex free_vmap_area_lock irq_context: 0 &dev->mutex free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &dev->mutex free_vmap_area_lock pool_lock#2 irq_context: 0 &dev->mutex vmap_area_lock irq_context: 0 &dev->mutex &md->mutex pci_lock irq_context: 0 &dev->mutex &md->mutex &rq->__lock irq_context: 0 &dev->mutex &md->mutex fs_reclaim irq_context: 0 &dev->mutex &md->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &xa->xa_lock#5 irq_context: 0 &dev->mutex &md->mutex &xa->xa_lock#5 pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock &its->lock irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock fs_reclaim irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock &zone->lock irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock &____s->seqcount irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &its->dev_alloc_lock lpi_range_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fs_reclaim irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &md->mutex &domain->mutex fs_reclaim irq_context: 0 &dev->mutex &md->mutex &domain->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &md->mutex &domain->mutex pool_lock#2 irq_context: 0 &dev->mutex &md->mutex &domain->mutex &irq_desc_lock_class irq_context: 0 &dev->mutex &md->mutex &irq_desc_lock_class irq_context: 0 &dev->mutex &md->mutex tmpmask_lock irq_context: 0 &dev->mutex &md->mutex &its->lock irq_context: 0 &dev->mutex &md->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &md->mutex lock irq_context: 0 &dev->mutex &md->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex &md->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &md->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &c->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &____s->seqcount irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class &its->lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock tmpmask_lock irq_context: 0 &dev->mutex &zone->lock irq_context: 0 &dev->mutex &zone->lock &____s->seqcount irq_context: 0 &dev->mutex &dev->vqs_list_lock irq_context: 0 &dev->mutex &vp_dev->lock irq_context: 0 &dev->mutex rng_mutex irq_context: 0 &dev->mutex rng_mutex &x->wait#13 irq_context: 0 &dev->mutex rng_mutex fs_reclaim irq_context: 0 &dev->mutex rng_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rng_mutex pool_lock#2 irq_context: 0 &dev->mutex rng_mutex kthread_create_lock irq_context: 0 &dev->mutex rng_mutex &p->pi_lock irq_context: 0 &dev->mutex rng_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rng_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rng_mutex &rq->__lock irq_context: hardirq &x->wait#14 irq_context: 0 &dev->mutex rng_mutex &x->wait irq_context: 0 &dev->mutex rng_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rng_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex reading_mutex irq_context: 0 &dev->mutex input_pool.lock irq_context: 0 &dev->mutex &dev->config_lock irq_context: 0 rng_mutex irq_context: 0 reading_mutex irq_context: 0 reading_mutex &x->wait#14 irq_context: softirq drivers/char/random.c:1010 irq_context: softirq drivers/char/random.c:1010 input_pool.lock irq_context: 0 misc_mtx &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx rcu_read_lock &rq->__lock irq_context: 0 misc_mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &base->lock irq_context: 0 tasklist_lock &base->lock &obj_hash[i].lock irq_context: 0 &dev->devres_lock irq_context: 0 &dev->managed.lock irq_context: 0 &type->s_umount_key#23/1 irq_context: 0 &type->s_umount_key#23/1 fs_reclaim irq_context: 0 &type->s_umount_key#23/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#23/1 pool_lock#2 irq_context: 0 &type->s_umount_key#23/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#23/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#23/1 shrinker_mutex irq_context: 0 &type->s_umount_key#23/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#23/1 sb_lock irq_context: 0 &type->s_umount_key#23/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#23/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#23/1 &sb->s_type->i_lock_key#21 irq_context: 0 &type->s_umount_key#23/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#23/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#23/1 &sb->s_type->i_lock_key#21 &dentry->d_lock irq_context: 0 &type->s_umount_key#23/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#21 irq_context: 0 lock drm_minor_lock irq_context: 0 lock drm_minor_lock pool_lock#2 irq_context: 0 lock drm_minor_lock &c->lock irq_context: 0 lock drm_minor_lock &____s->seqcount irq_context: 0 stack_depot_init_mutex irq_context: 0 subsys mutex#35 irq_context: 0 subsys mutex#35 &k->k_lock irq_context: 0 drm_minor_lock irq_context: 0 &dev->mode_config.idr_mutex irq_context: 0 &dev->mode_config.idr_mutex fs_reclaim irq_context: 0 &dev->mode_config.idr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mode_config.idr_mutex pool_lock#2 irq_context: 0 crtc_ww_class_acquire irq_context: 0 crtc_ww_class_acquire crtc_ww_class_mutex irq_context: 0 crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_acquire irq_context: 0 crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_acquire reservation_ww_class_mutex irq_context: 0 &dev->mode_config.blob_lock irq_context: 0 &xa->xa_lock#6 irq_context: 0 &xa->xa_lock#7 irq_context: 0 &dev->mode_config.connector_list_lock irq_context: 0 &dev->vbl_lock irq_context: 0 drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 drm_connector_list_iter fs_reclaim irq_context: 0 drm_connector_list_iter fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex irq_context: 0 drm_connector_list_iter &connector->mutex fs_reclaim irq_context: 0 drm_connector_list_iter &connector->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter &connector->mutex pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex &x->wait#9 irq_context: 0 drm_connector_list_iter &connector->mutex &obj_hash[i].lock irq_context: 0 drm_connector_list_iter &connector->mutex &k->list_lock irq_context: 0 drm_connector_list_iter &connector->mutex lock irq_context: 0 drm_connector_list_iter &connector->mutex lock kernfs_idr_lock irq_context: 0 drm_connector_list_iter &connector->mutex &root->kernfs_rwsem irq_context: 0 drm_connector_list_iter &connector->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 drm_connector_list_iter &connector->mutex bus_type_sem irq_context: 0 drm_connector_list_iter &connector->mutex sysfs_symlink_target_lock irq_context: 0 drm_connector_list_iter &connector->mutex &c->lock irq_context: 0 drm_connector_list_iter &connector->mutex &____s->seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &root->kernfs_rwsem irq_context: 0 drm_connector_list_iter &connector->mutex &dev->power.lock irq_context: 0 drm_connector_list_iter &connector->mutex dpm_list_mtx irq_context: 0 drm_connector_list_iter &connector->mutex uevent_sock_mutex irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 drm_connector_list_iter &connector->mutex running_helpers_waitq.lock irq_context: 0 drm_connector_list_iter &connector->mutex &k->k_lock irq_context: 0 drm_connector_list_iter &connector->mutex subsys mutex#35 irq_context: 0 drm_connector_list_iter &connector->mutex subsys mutex#35 &k->k_lock irq_context: 0 drm_connector_list_iter &connector->mutex pin_fs_lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 drm_connector_list_iter &connector->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 drm_connector_list_iter &connector->mutex &dev->mode_config.idr_mutex irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 drm_connector_list_iter &connector->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 drm_connector_list_iter &connector->mutex connector_list_lock irq_context: 0 &dev->filelist_mutex irq_context: 0 &dev->clientlist_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock drm_connector_list_iter pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex &c->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex crtc_ww_class_acquire crtc_ww_class_mutex &____s->seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex &dev->mode_config.mutex &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock &client->modeset_mutex drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &sbinfo->stat_lock irq_context: 0 &dev->clientlist_mutex &helper->lock mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex &helper->lock &sb->s_type->i_lock_key irq_context: 0 &dev->clientlist_mutex &helper->lock &s->s_inode_list_lock irq_context: 0 &dev->clientlist_mutex &helper->lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex &helper->lock batched_entropy_u32.lock irq_context: 0 &dev->clientlist_mutex &helper->lock batched_entropy_u32.lock crngs.lock irq_context: 0 &dev->clientlist_mutex &helper->lock &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &mgr->vm_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &mgr->vm_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock lock &file_private->table_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->object_name_lock lock &file_private->table_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &node->vm_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &file_private->table_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->mode_config.idr_mutex irq_context: 0 &dev->clientlist_mutex &helper->lock &dev->mode_config.fb_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &file->fbs_lock irq_context: 0 &dev->clientlist_mutex &helper->lock &prime_fpriv->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &node->vm_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &node->vm_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &file_private->table_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock &file_private->table_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock free_vmap_area_lock irq_context: 0 &dev->clientlist_mutex &helper->lock vmap_area_lock irq_context: 0 &dev->clientlist_mutex &helper->lock init_mm.page_table_lock irq_context: 0 &dev->clientlist_mutex &helper->lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex &helper->lock free_vmap_area_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex &helper->lock &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex &helper->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock irq_context: 0 &dev->clientlist_mutex registration_lock fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock &x->wait#9 irq_context: 0 &dev->clientlist_mutex registration_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock lock irq_context: 0 &dev->clientlist_mutex registration_lock lock kernfs_idr_lock irq_context: 0 &dev->clientlist_mutex registration_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock bus_type_sem irq_context: 0 &dev->clientlist_mutex registration_lock sysfs_symlink_target_lock irq_context: 0 &dev->clientlist_mutex registration_lock &c->lock irq_context: 0 &dev->clientlist_mutex registration_lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock &dev->power.lock irq_context: 0 &dev->clientlist_mutex registration_lock dpm_list_mtx irq_context: 0 &dev->clientlist_mutex registration_lock req_lock irq_context: 0 &dev->clientlist_mutex registration_lock &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock &x->wait#11 irq_context: 0 &dev->clientlist_mutex registration_lock uevent_sock_mutex irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock running_helpers_waitq.lock irq_context: 0 &dev->clientlist_mutex registration_lock &k->k_lock irq_context: 0 &dev->clientlist_mutex registration_lock subsys mutex#11 irq_context: 0 &dev->clientlist_mutex registration_lock subsys mutex#11 &k->k_lock irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock vt_switch_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock (console_sem).lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &fb_info->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock vt_event_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &base->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &base->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &c->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &x->wait#9 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock gdp_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock gdp_mutex &k->list_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock lock kernfs_idr_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock bus_type_sem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock sysfs_symlink_target_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &root->kernfs_rwsem irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &dev->power.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock dpm_list_mtx irq_context: 0 &dev->clientlist_mutex registration_lock console_lock uevent_sock_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock running_helpers_waitq.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock subsys mutex#6 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock subsys mutex#6 &k->k_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->mode_config.idr_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->mode_config.blob_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter &dev->mode_config.connector_list_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &c->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &crtc->commit_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &xa->xa_lock#8 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &sb->s_type->i_lock_key irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &info->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex lock#4 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &xa->xa_lock#8 &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &xa->xa_lock#8 pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex lock#4 &lruvec->lru_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex &c->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex free_vmap_area_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex vmap_area_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex reservation_ww_class_mutex init_mm.page_table_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &____s->seqcount#5 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &x->wait#15 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &dev->vbl_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock &____s->seqcount#5 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &x->wait#15 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex (work_completion)(&vkms_state->composer_work) irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->damage_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->damage_lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock reservation_ww_class_mutex irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter fs_reclaim irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex drm_connector_list_iter pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vbl_lock &dev->vblank_time_lock tk_core.seq.seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &vkms_out->lock &dev->event_lock &dev->vblank_time_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &base->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pool_lock irq_context: hardirq &vkms_out->lock irq_context: hardirq &vkms_out->lock &dev->event_lock irq_context: hardirq &vkms_out->lock &dev->event_lock &dev->vblank_time_lock irq_context: hardirq &vkms_out->lock &dev->event_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock irq_context: hardirq &vkms_out->lock &dev->event_lock &dev->vblank_time_lock &(&vblank->seqlock)->lock &____s->seqcount#5 irq_context: hardirq &vkms_out->lock &dev->event_lock &vblank->queue irq_context: hardirq &vkms_out->lock &dev->event_lock &____s->seqcount#5 irq_context: hardirq &vkms_out->lock &dev->event_lock &obj_hash[i].lock irq_context: hardirq &vkms_out->lock &dev->event_lock &base->lock irq_context: hardirq &vkms_out->lock &dev->event_lock &base->lock &obj_hash[i].lock irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#15 irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#15 &p->pi_lock irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#15 &p->pi_lock &rq->__lock irq_context: hardirq &vkms_out->lock &dev->event_lock &x->wait#15 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &vkms_out->lock &dev->event_lock pool_lock#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex (&timer.timer) irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex (work_completion)(&vkms_state->composer_work)#2 irq_context: 0 &dev->clientlist_mutex registration_lock console_lock (console_sem).lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&helper->damage_work) &helper->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &pcp->lock &zone->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &helper->lock &dev->master_mutex &client->modeset_mutex crtc_ww_class_acquire crtc_ww_class_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &lock->wait_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &p->pi_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->clientlist_mutex registration_lock console_lock batched_entropy_u8.lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock kfence_freelist_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock &meta->lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->clientlist_mutex registration_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->clientlist_mutex (console_sem).lock irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->clientlist_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->clientlist_mutex kernel_fb_helper_lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem pool_lock#2 irq_context: 0 drivers_lock#2 irq_context: 0 devices_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex lock kernfs_idr_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &root->kernfs_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &c->lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up &x->wait#9 irq_context: 0 cpu_hotplug_lock cpuhp_state-up fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state-up fill_pool_map-wait-type-override pool_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up &k->list_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up bus_type_sem irq_context: 0 cpu_hotplug_lock cpuhp_state-up &k->k_lock irq_context: 0 cpu_hotplug_lock cpuhp_state-up lock kernfs_idr_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state-up &pcp->lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &cfs_rq->removed.lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &obj_hash[i].lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 blk_queue_ida.xa_lock irq_context: 0 &sb->s_type->i_lock_key#3 irq_context: 0 &xa->xa_lock#9 irq_context: 0 lock &q->queue_lock irq_context: 0 lock &q->queue_lock &blkcg->lock irq_context: 0 &q->queue_lock irq_context: 0 &q->queue_lock pool_lock#2 irq_context: 0 &q->queue_lock pcpu_lock irq_context: 0 &q->queue_lock &obj_hash[i].lock irq_context: 0 &q->queue_lock percpu_counters_lock irq_context: 0 &q->queue_lock &obj_hash[i].lock pool_lock irq_context: 0 &q->queue_lock &blkcg->lock irq_context: 0 &bdev->bd_size_lock irq_context: 0 subsys mutex#36 irq_context: 0 subsys mutex#36 &k->k_lock irq_context: 0 dev_hotplug_mutex irq_context: 0 dev_hotplug_mutex &dev->power.lock irq_context: 0 &q->sysfs_dir_lock irq_context: 0 &q->sysfs_dir_lock fs_reclaim irq_context: 0 &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &q->sysfs_dir_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 percpu_ref_switch_lock irq_context: 0 gdp_mutex &pcp->lock &zone->lock irq_context: 0 gdp_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 subsys mutex#37 irq_context: 0 subsys mutex#37 &k->k_lock irq_context: 0 cgwb_lock irq_context: 0 bdi_lock irq_context: 0 inode_hash_lock irq_context: 0 inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 bdev_lock irq_context: 0 &disk->open_mutex irq_context: 0 &disk->open_mutex fs_reclaim irq_context: 0 &disk->open_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &disk->open_mutex pool_lock#2 irq_context: 0 &disk->open_mutex free_vmap_area_lock irq_context: 0 &disk->open_mutex vmap_area_lock irq_context: 0 &disk->open_mutex &____s->seqcount irq_context: 0 &disk->open_mutex init_mm.page_table_lock irq_context: 0 &disk->open_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &disk->open_mutex &xa->xa_lock#8 irq_context: 0 &disk->open_mutex lock#4 irq_context: 0 &disk->open_mutex mmu_notifier_invalidate_range_start irq_context: 0 &disk->open_mutex &pcp->lock &zone->lock irq_context: 0 &disk->open_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &disk->open_mutex &c->lock irq_context: 0 &disk->open_mutex &mapping->i_private_lock irq_context: 0 &disk->open_mutex tk_core.seq.seqcount irq_context: 0 &disk->open_mutex &ret->b_uptodate_lock irq_context: 0 &disk->open_mutex &obj_hash[i].lock irq_context: 0 &disk->open_mutex &xa->xa_lock#8 pool_lock#2 irq_context: 0 &disk->open_mutex purge_vmap_area_lock irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#8 irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#8 &obj_hash[i].lock irq_context: 0 &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#8 pool_lock#2 irq_context: 0 &disk->open_mutex lock#4 &lruvec->lru_lock irq_context: 0 &disk->open_mutex lock#5 irq_context: 0 &disk->open_mutex &lruvec->lru_lock irq_context: 0 pcpu_alloc_mutex &rq->__lock irq_context: 0 lock &q->queue_lock &blkcg->lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &disk->open_mutex purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex purge_vmap_area_lock pool_lock#2 irq_context: 0 &q->queue_lock &c->lock irq_context: 0 &q->queue_lock &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &q->queue_lock &pcp->lock &zone->lock irq_context: 0 &q->queue_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &disk->open_mutex &rq->__lock irq_context: 0 misc_mtx &cfs_rq->removed.lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 loop_ctl_mutex irq_context: 0 loop_ctl_mutex fs_reclaim irq_context: 0 loop_ctl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 loop_ctl_mutex pool_lock#2 irq_context: 0 &q->sysfs_lock irq_context: 0 &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_lock pool_lock#2 irq_context: 0 &q->sysfs_lock &obj_hash[i].lock irq_context: 0 &q->sysfs_lock &c->lock irq_context: 0 &q->sysfs_lock &____s->seqcount irq_context: 0 &q->sysfs_lock cpu_hotplug_lock irq_context: 0 &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 &q->sysfs_lock fs_reclaim irq_context: 0 &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_lock &xa->xa_lock#10 irq_context: 0 &set->tag_list_lock irq_context: 0 &q->mq_freeze_lock irq_context: 0 &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex percpu_ref_switch_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex pin_fs_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &stats->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u64.lock crngs.lock base_crng.lock irq_context: 0 cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 nbd_index_mutex irq_context: 0 nbd_index_mutex fs_reclaim irq_context: 0 nbd_index_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nbd_index_mutex pool_lock#2 irq_context: 0 set->srcu irq_context: 0 (work_completion)(&(&q->requeue_work)->work) irq_context: 0 (work_completion)(&(&hctx->run_work)->work) irq_context: 0 &q->debugfs_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex set->srcu irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &q->sysfs_lock &rq->__lock irq_context: 0 &q->sysfs_lock &cfs_rq->removed.lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: softirq &(&ops->cursor_work)->timer irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq &(&ops->cursor_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) (console_sem).lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock &helper->damage_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) console_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &base->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&ops->cursor_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock &c->lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock &____s->seqcount irq_context: 0 zram_index_mutex irq_context: 0 zram_index_mutex fs_reclaim irq_context: 0 zram_index_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex pool_lock#2 irq_context: 0 zram_index_mutex blk_queue_ida.xa_lock irq_context: 0 zram_index_mutex &obj_hash[i].lock irq_context: 0 zram_index_mutex pcpu_alloc_mutex irq_context: 0 zram_index_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 zram_index_mutex &obj_hash[i].lock pool_lock irq_context: 0 zram_index_mutex &rq->__lock irq_context: 0 zram_index_mutex bio_slab_lock irq_context: 0 zram_index_mutex percpu_counters_lock irq_context: 0 zram_index_mutex mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &sb->s_type->i_lock_key#3 irq_context: 0 zram_index_mutex &s->s_inode_list_lock irq_context: 0 zram_index_mutex &xa->xa_lock#9 irq_context: 0 zram_index_mutex lock irq_context: 0 zram_index_mutex lock &q->queue_lock irq_context: 0 zram_index_mutex lock &q->queue_lock &blkcg->lock irq_context: 0 zram_index_mutex &q->queue_lock irq_context: 0 zram_index_mutex &q->queue_lock pool_lock#2 irq_context: 0 zram_index_mutex &q->queue_lock pcpu_lock irq_context: 0 zram_index_mutex &q->queue_lock &obj_hash[i].lock irq_context: 0 zram_index_mutex &q->queue_lock percpu_counters_lock irq_context: 0 zram_index_mutex &q->queue_lock &blkcg->lock irq_context: 0 zram_index_mutex &x->wait#9 irq_context: 0 zram_index_mutex &bdev->bd_size_lock irq_context: 0 zram_index_mutex &k->list_lock irq_context: 0 zram_index_mutex gdp_mutex irq_context: 0 zram_index_mutex gdp_mutex &k->list_lock irq_context: 0 zram_index_mutex lock kernfs_idr_lock irq_context: 0 zram_index_mutex &root->kernfs_rwsem irq_context: 0 zram_index_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 zram_index_mutex bus_type_sem irq_context: 0 zram_index_mutex sysfs_symlink_target_lock irq_context: 0 zram_index_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 zram_index_mutex &c->lock irq_context: 0 zram_index_mutex &____s->seqcount irq_context: 0 zram_index_mutex &root->kernfs_rwsem irq_context: 0 zram_index_mutex &dev->power.lock irq_context: 0 zram_index_mutex dpm_list_mtx irq_context: 0 zram_index_mutex req_lock irq_context: 0 zram_index_mutex &p->pi_lock irq_context: 0 zram_index_mutex &x->wait#11 irq_context: 0 zram_index_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 zram_index_mutex subsys mutex#36 irq_context: 0 zram_index_mutex subsys mutex#36 &k->k_lock irq_context: 0 zram_index_mutex dev_hotplug_mutex irq_context: 0 zram_index_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock fs_reclaim irq_context: 0 zram_index_mutex &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &q->sysfs_dir_lock pool_lock#2 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &c->lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &____s->seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 zram_index_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 zram_index_mutex percpu_ref_switch_lock irq_context: 0 zram_index_mutex uevent_sock_mutex irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 zram_index_mutex running_helpers_waitq.lock irq_context: 0 zram_index_mutex subsys mutex#37 irq_context: 0 zram_index_mutex subsys mutex#37 &k->k_lock irq_context: 0 zram_index_mutex cgwb_lock irq_context: 0 zram_index_mutex pin_fs_lock irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 zram_index_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 zram_index_mutex bdi_lock irq_context: 0 zram_index_mutex inode_hash_lock irq_context: 0 zram_index_mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 zram_index_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 zram_index_mutex (console_sem).lock irq_context: 0 zram_index_mutex console_lock console_srcu console_owner_lock irq_context: 0 zram_index_mutex console_lock console_srcu console_owner irq_context: 0 zram_index_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 zram_index_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 subsys mutex#38 irq_context: 0 subsys mutex#38 &k->k_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]#3 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]#3 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]#3 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]#3 configfs_dirent_lock irq_context: 0 &q->sysfs_lock &pcp->lock &zone->lock irq_context: 0 &q->sysfs_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &q->sysfs_lock &xa->xa_lock#10 pool_lock#2 irq_context: 0 &lock irq_context: 0 &lock nullb_indexes.xa_lock irq_context: 0 &q->sysfs_dir_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 &disk->open_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex rcu_read_lock &base->lock irq_context: 0 &disk->open_mutex rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex rcu_read_lock &ret->b_uptodate_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key &rq->__lock irq_context: 0 nfc_index_ida.xa_lock irq_context: 0 nfc_devlist_mutex irq_context: 0 nfc_devlist_mutex fs_reclaim irq_context: 0 nfc_devlist_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nfc_devlist_mutex pool_lock#2 irq_context: 0 nfc_devlist_mutex &k->list_lock irq_context: 0 nfc_devlist_mutex gdp_mutex irq_context: 0 nfc_devlist_mutex gdp_mutex &k->list_lock irq_context: 0 nfc_devlist_mutex gdp_mutex fs_reclaim irq_context: 0 nfc_devlist_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nfc_devlist_mutex gdp_mutex pool_lock#2 irq_context: 0 nfc_devlist_mutex gdp_mutex lock irq_context: 0 nfc_devlist_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 nfc_devlist_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 nfc_devlist_mutex lock irq_context: 0 nfc_devlist_mutex lock kernfs_idr_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 nfc_devlist_mutex bus_type_sem irq_context: 0 nfc_devlist_mutex sysfs_symlink_target_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex &c->lock irq_context: 0 nfc_devlist_mutex &____s->seqcount irq_context: 0 nfc_devlist_mutex &dev->power.lock irq_context: 0 nfc_devlist_mutex dpm_list_mtx irq_context: 0 nfc_devlist_mutex &pcp->lock &zone->lock irq_context: 0 nfc_devlist_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 nfc_devlist_mutex uevent_sock_mutex irq_context: 0 nfc_devlist_mutex &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 nfc_devlist_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 nfc_devlist_mutex running_helpers_waitq.lock irq_context: 0 nfc_devlist_mutex subsys mutex#39 irq_context: 0 nfc_devlist_mutex subsys mutex#39 &k->k_lock irq_context: 0 llcp_devices_lock irq_context: 0 &dev->mutex rfkill_global_mutex irq_context: 0 &dev->mutex rfkill_global_mutex fs_reclaim irq_context: 0 &dev->mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rfkill_global_mutex pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex &k->list_lock irq_context: 0 &dev->mutex rfkill_global_mutex lock irq_context: 0 &dev->mutex rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex rfkill_global_mutex bus_type_sem irq_context: 0 &dev->mutex rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex rfkill_global_mutex &c->lock irq_context: 0 &dev->mutex rfkill_global_mutex &____s->seqcount irq_context: 0 &dev->mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rfkill_global_mutex &dev->power.lock irq_context: 0 &dev->mutex rfkill_global_mutex dpm_list_mtx irq_context: 0 &dev->mutex rfkill_global_mutex &rfkill->lock irq_context: 0 &dev->mutex rfkill_global_mutex uevent_sock_mutex irq_context: 0 &dev->mutex rfkill_global_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex rfkill_global_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex rfkill_global_mutex &k->k_lock irq_context: 0 &dev->mutex rfkill_global_mutex subsys mutex#40 irq_context: 0 &dev->mutex rfkill_global_mutex subsys mutex#40 &k->k_lock irq_context: 0 &dev->mutex rfkill_global_mutex triggers_list_lock irq_context: 0 &dev->mutex rfkill_global_mutex leds_list_lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rfkill_global_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rfkill->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 dma_heap_minors.xa_lock irq_context: 0 subsys mutex#41 irq_context: 0 subsys mutex#41 &k->k_lock irq_context: 0 heap_list_lock irq_context: 0 dma_heap_minors.xa_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &rq->__lock irq_context: 0 major_names_lock &c->lock irq_context: 0 major_names_lock &____s->seqcount irq_context: 0 subsys mutex#42 irq_context: 0 subsys mutex#42 &k->list_lock irq_context: 0 subsys mutex#42 &k->k_lock irq_context: softirq &(&kfence_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 nvmf_hosts_mutex irq_context: 0 subsys mutex#43 irq_context: 0 subsys mutex#43 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex semaphore->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex iommu_probe_device_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex *(&acpi_gbl_reference_count_lock) irq_context: 0 nvmf_transports_rwsem irq_context: 0 subsys mutex#44 irq_context: 0 subsys mutex#44 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock semaphore->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock *(&acpi_gbl_reference_count_lock) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock (console_sem).lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex acpi_link_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex irq_domain_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &domain->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &drv->dynids.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cpu_add_remove_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_instance_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex chrdevs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex req_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &x->wait#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#45 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex subsys mutex#45 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex dev_pm_qos_mtx pm_qos_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex resource_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex free_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock pools_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pools_reg_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex (console_sem).lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pci_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex pci_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &xa->xa_lock#5 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->dev_alloc_lock lpi_range_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &domain->mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex tmpmask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex hrtimer_bases.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &sb->s_type->i_lock_key#18 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 irq_context: 0 nvmet_config_sem irq_context: 0 subsys mutex#46 irq_context: 0 subsys mutex#46 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex &irq_desc_lock_class &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock tmpmask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex register_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pcpu_alloc_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex batched_entropy_u32.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex blk_queue_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock cpu_hotplug_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &q->sysfs_lock &xa->xa_lock#10 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &set->tag_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &ctrl->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &ctrl->lock &ctrl->state_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &hctx->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &hctx->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &nvmeq->sq_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &x->wait#16 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &base->lock &obj_hash[i].lock irq_context: softirq &x->wait#16 irq_context: softirq &x->wait#16 &p->pi_lock irq_context: softirq &x->wait#16 &p->pi_lock &rq->__lock irq_context: softirq &x->wait#16 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex (&timer.timer) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock bus_type_sem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &dev->power.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock dpm_list_mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock uevent_sock_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock running_helpers_waitq.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock subsys mutex#47 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex nvme_subsystems_lock subsys mutex#47 &k->k_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &xa->xa_lock#11 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class irq_resend_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &ent->pde_unload_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex pci_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &irq_desc_lock_class irq_context: 0 lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &obj_hash[i].lock irq_context: 0 lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock lpi_range_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock lpi_range_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock lpi_range_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex &its->dev_alloc_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &ent->pde_unload_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex sparse_irq_lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &xa->xa_lock#5 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex purge_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock pci_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock free_vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &xa->xa_lock#5 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->dev_alloc_lock lpi_range_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex tmpmask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &its->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &md->mutex &domain->mutex sparse_irq_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock tmpmask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &irq_desc_lock_class irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock proc_subdir_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &cma->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex cma_mutex &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class tmp_mask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &irq_desc_lock_class tmp_mask_lock tmpmask_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock register_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &dev->shutdown_lock &desc->request_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &md->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &sb->s_type->i_lock_key#18 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &sb->s_type->i_lock_key#18 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &sb->s_type->i_lock_key#18 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &sb->s_type->i_lock_key#18 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7 configfs_dirent_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &default_group_class[depth - 1]/2 &default_group_class[depth - 1]#4/2 &default_group_class[depth - 1]#5/2 &default_group_class[depth - 1]#6/2 &default_group_class[depth - 1]#7/2 irq_context: 0 backend_mutex irq_context: 0 scsi_mib_index_lock irq_context: 0 hba_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &ctrl->namespaces_rwsem irq_context: 0 device_mutex irq_context: 0 device_mutex fs_reclaim irq_context: 0 device_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 device_mutex pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->async_event_work) irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->async_event_work) &nvmeq->sq_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &hctx->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &hctx->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &nvmeq->sq_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &x->wait#16 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &base->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock (&timer.timer) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex (work_completion)(&ctrl->scan_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex rcu_read_lock (wq_completion)nvme-wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&entry->work) &dev->mutex &x->wait#10 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &ctrl->namespaces_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock blk_queue_ida.xa_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock pcpu_alloc_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock cpu_hotplug_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock &xa->xa_lock#10 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_lock &xa->xa_lock#10 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &set->tag_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock bio_slab_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock percpu_counters_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &s->s_inode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &xa->xa_lock#9 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock lock &q->queue_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock lock &q->queue_lock &blkcg->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->mq_freeze_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock pcpu_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock percpu_counters_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->queue_lock &blkcg->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &x->wait#9 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock nvme_subsystems_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock irq_context: 0 &hba->device_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock &xa->xa_lock#11 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock pcpu_alloc_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &subsys->lock pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &bdev->bd_size_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &ctrl->namespaces_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock ext_devt_ida.xa_lock irq_context: 0 rcu_read_lock init_fs.seq.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &k->list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock bus_type_sem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &dev->power.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock dpm_list_mtx irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock req_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &x->wait#11 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#36 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#36 &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock dev_hotplug_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock dev_hotplug_mutex &dev->power.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &____s->seqcount irq_context: 0 part_parser_lock irq_context: 0 mtd_table_mutex irq_context: 0 chip_drvs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 (kmod_concurrent_max).lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex percpu_ref_switch_lock irq_context: 0 &x->wait#17 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex pin_fs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &sig->wait_chldexit irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &q->sysfs_dir_lock &q->sysfs_lock &stats->lock irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock gdp_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock gdp_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &____s->seqcount#4 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &prev->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &(&sig->stats_lock)->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &(&sig->stats_lock)->lock &____s->seqcount#4 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) css_set_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock input_pool.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock gdp_mutex &k->list_lock irq_context: 0 mtd_table_mutex fs_reclaim irq_context: 0 mtd_table_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex pool_lock#2 irq_context: 0 mtd_table_mutex &x->wait#9 irq_context: 0 mtd_table_mutex &obj_hash[i].lock irq_context: 0 mtd_table_mutex &k->list_lock irq_context: 0 mtd_table_mutex gdp_mutex irq_context: 0 mtd_table_mutex gdp_mutex &k->list_lock irq_context: 0 mtd_table_mutex gdp_mutex fs_reclaim irq_context: 0 mtd_table_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex gdp_mutex &____s->seqcount irq_context: 0 mtd_table_mutex gdp_mutex pool_lock#2 irq_context: 0 mtd_table_mutex gdp_mutex &obj_hash[i].lock irq_context: 0 mtd_table_mutex gdp_mutex lock irq_context: 0 mtd_table_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 mtd_table_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 mtd_table_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock uevent_sock_mutex irq_context: 0 mtd_table_mutex lock irq_context: 0 mtd_table_mutex lock kernfs_idr_lock irq_context: 0 mtd_table_mutex &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 mtd_table_mutex bus_type_sem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mtd_table_mutex &c->lock irq_context: 0 mtd_table_mutex &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock running_helpers_waitq.lock irq_context: 0 mtd_table_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 mtd_table_mutex &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &dev->power.lock irq_context: 0 mtd_table_mutex dpm_list_mtx irq_context: 0 mtd_table_mutex req_lock irq_context: 0 mtd_table_mutex &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#37 irq_context: 0 mtd_table_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#37 &k->k_lock irq_context: 0 mtd_table_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock cgwb_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock pin_fs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 irq_context: 0 mtd_table_mutex &x->wait#11 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 mtd_table_mutex &rq->__lock irq_context: 0 mtd_table_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 mtd_table_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 mtd_table_mutex uevent_sock_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock bdi_lock irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 mtd_table_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock inode_hash_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 mtd_table_mutex running_helpers_waitq.lock irq_context: 0 mtd_table_mutex subsys mutex#48 irq_context: 0 mtd_table_mutex subsys mutex#48 &k->k_lock irq_context: 0 mtd_table_mutex devtree_lock irq_context: 0 mtd_table_mutex nvmem_ida.xa_lock irq_context: 0 mtd_table_mutex nvmem_cell_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock bdev_lock irq_context: 0 mtd_table_mutex &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex fs_reclaim irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex free_vmap_area_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex vmap_area_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex init_mm.page_table_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &rq->__lock irq_context: 0 mtd_table_mutex &dev->mutex &dev->power.lock irq_context: 0 mtd_table_mutex &dev->mutex &k->list_lock irq_context: 0 mtd_table_mutex &dev->mutex &k->k_lock irq_context: 0 mtd_table_mutex subsys mutex#49 irq_context: 0 mtd_table_mutex nvmem_mutex irq_context: 0 mtd_table_mutex pin_fs_lock irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 mtd_table_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#8 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex lock#4 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &mapping->i_private_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex tk_core.seq.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &base->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &nvmeq->sq_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &folio_wait_table[i] irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mtd_table_mutex (console_sem).lock irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner_lock irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 mtd_table_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 mtd_table_mutex pcpu_alloc_mutex irq_context: 0 mtd_table_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 mtd_table_mutex batched_entropy_u32.lock irq_context: 0 mtd_table_mutex mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex blk_queue_ida.xa_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock &q->unused_hctx_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex &q->sysfs_lock cpu_hotplug_lock irq_context: 0 mtd_table_mutex &q->sysfs_lock cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 mtd_table_mutex &q->sysfs_lock fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_lock &xa->xa_lock#10 irq_context: 0 mtd_table_mutex &obj_hash[i].lock pool_lock irq_context: 0 mtd_table_mutex &set->tag_list_lock irq_context: 0 mtd_table_mutex bio_slab_lock irq_context: 0 mtd_table_mutex percpu_counters_lock irq_context: 0 mtd_table_mutex &sb->s_type->i_lock_key#3 irq_context: 0 mtd_table_mutex &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &xa->xa_lock#9 irq_context: 0 mtd_table_mutex lock &q->queue_lock irq_context: 0 mtd_table_mutex lock &q->queue_lock &blkcg->lock irq_context: 0 mtd_table_mutex &q->mq_freeze_lock irq_context: 0 mtd_table_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex set->srcu irq_context: 0 mtd_table_mutex percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->queue_lock irq_context: 0 mtd_table_mutex &q->queue_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->queue_lock pcpu_lock irq_context: 0 mtd_table_mutex &q->queue_lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex &q->queue_lock percpu_counters_lock irq_context: 0 mtd_table_mutex &q->queue_lock &obj_hash[i].lock pool_lock irq_context: 0 mtd_table_mutex &q->queue_lock &blkcg->lock irq_context: 0 mtd_table_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex &bdev->bd_size_lock irq_context: 0 mtd_table_mutex elv_list_lock irq_context: 0 mtd_table_mutex (work_completion)(&(&q->requeue_work)->work) irq_context: 0 mtd_table_mutex (work_completion)(&(&hctx->run_work)->work) irq_context: 0 mtd_table_mutex &q->debugfs_mutex irq_context: 0 mtd_table_mutex subsys mutex#36 irq_context: 0 mtd_table_mutex subsys mutex#36 &k->k_lock irq_context: 0 mtd_table_mutex dev_hotplug_mutex irq_context: 0 mtd_table_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &c->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &____s->seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock lock kernfs_idr_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pin_fs_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &c->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &____s->seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock percpu_ref_switch_lock rcu_read_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex set->srcu irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex percpu_ref_switch_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &q->mq_freeze_wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex pin_fs_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 mtd_table_mutex &q->sysfs_dir_lock &q->sysfs_lock &stats->lock irq_context: 0 mtd_table_mutex subsys mutex#37 irq_context: 0 mtd_table_mutex subsys mutex#37 &k->k_lock irq_context: 0 mtd_table_mutex cgwb_lock irq_context: 0 mtd_table_mutex bdi_lock irq_context: 0 mtd_table_mutex inode_hash_lock irq_context: 0 mtd_table_mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 rtnl_mutex stack_depot_init_mutex irq_context: 0 rtnl_mutex pcpu_alloc_mutex irq_context: 0 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex kthread_create_lock irq_context: 0 rtnl_mutex &p->pi_lock irq_context: 0 rtnl_mutex &x->wait irq_context: 0 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex wq_pool_mutex irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex irq_context: 0 rtnl_mutex crngs.lock irq_context: 0 rtnl_mutex &xa->xa_lock#4 irq_context: 0 rtnl_mutex net_rwsem irq_context: 0 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex &x->wait#9 irq_context: 0 rtnl_mutex &k->list_lock irq_context: 0 rtnl_mutex gdp_mutex irq_context: 0 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 rtnl_mutex lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex bus_type_sem irq_context: 0 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex &dev->power.lock irq_context: 0 rtnl_mutex dpm_list_mtx irq_context: 0 rtnl_mutex uevent_sock_mutex irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex running_helpers_waitq.lock irq_context: 0 rtnl_mutex subsys mutex#20 irq_context: 0 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 rtnl_mutex &dir->lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex dev_hotplug_mutex irq_context: 0 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 rtnl_mutex dev_base_lock irq_context: 0 rtnl_mutex input_pool.lock irq_context: 0 rtnl_mutex batched_entropy_u32.lock irq_context: 0 rtnl_mutex &tbl->lock irq_context: 0 rtnl_mutex sysctl_lock irq_context: 0 rtnl_mutex nl_table_lock irq_context: 0 rtnl_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock &bond->stats_lock irq_context: 0 rtnl_mutex lweventlist_lock irq_context: 0 rtnl_mutex lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex.wait_lock irq_context: 0 pernet_ops_rwsem &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock pool_lock#2 irq_context: 0 once_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &dir->lock#2 irq_context: 0 once_lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex dev_base_lock irq_context: 0 (inet6addr_validator_chain).rwsem irq_context: 0 (inetaddr_validator_chain).rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex stack_depot_init_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)gid-cache-wq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work) irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events (work_completion)(&w->work) cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->work) pool_lock#2 irq_context: hardirq &ret->b_uptodate_lock irq_context: hardirq &folio_wait_table[i] irq_context: hardirq &folio_wait_table[i] &p->pi_lock irq_context: hardirq &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: hardirq &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#8 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#8 &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#8 &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex (console_sem).lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_owner_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex console_owner irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &s->s_inode_list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex pcpu_alloc_mutex irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &bdev->bd_size_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &x->wait#9 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex ext_devt_ida.xa_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &k->list_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sem->wait_lock irq_context: 0 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex bus_type_sem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &dev->power.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex dpm_list_mtx irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex req_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &x->wait#11 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex subsys mutex#36 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex subsys mutex#36 &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#9 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &xa->xa_lock#9 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex inode_hash_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex purge_vmap_area_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex purge_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#8 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#8 &obj_hash[i].lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#8 pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &sb->s_type->i_lock_key#3 &xa->xa_lock#8 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex lock#5 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &disk->open_mutex &lruvec->lru_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock nvme_ns_chr_minor_ida.xa_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock chrdevs_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#50 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock subsys mutex#50 &k->k_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &dentry->d_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 subsys mutex#51 irq_context: 0 subsys mutex#51 &k->k_lock irq_context: 0 gpio_lookup_lock irq_context: 0 mdio_board_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock quarantine_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&ctrl->scan_work) &ctrl->scan_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nvme-wq (work_completion)(&barr->work) &rq->__lock irq_context: 0 mode_list_lock irq_context: 0 l3mdev_lock irq_context: softirq rcu_callback put_task_map-wait-type-override &base->lock irq_context: softirq rcu_callback put_task_map-wait-type-override &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex init_mm.page_table_lock irq_context: 0 &dev->mutex stack_depot_init_mutex irq_context: 0 &dev->mutex pcpu_alloc_mutex irq_context: 0 &dev->mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &dev->mutex cpu_hotplug_lock irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex wq_pool_mutex irq_context: 0 &dev->mutex wq_pool_mutex &wq->mutex irq_context: 0 &dev->mutex pools_reg_lock irq_context: 0 &dev->mutex pools_reg_lock pools_lock irq_context: 0 &dev->mutex pools_reg_lock fs_reclaim irq_context: 0 &dev->mutex pools_reg_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex pools_reg_lock pool_lock#2 irq_context: 0 &dev->mutex pools_reg_lock lock irq_context: 0 &dev->mutex pools_reg_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex pools_reg_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex pools_reg_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &retval->lock irq_context: 0 &dev->mutex &irq_desc_lock_class tmp_mask_lock irq_context: 0 &dev->mutex &irq_desc_lock_class tmp_mask_lock tmpmask_lock irq_context: 0 &dev->mutex &irq_desc_lock_class tmp_mask_lock &its->lock irq_context: 0 &dev->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock &its->lock irq_context: 0 &dev->mutex rtnl_mutex irq_context: 0 &dev->mutex rtnl_mutex &c->lock irq_context: 0 &dev->mutex rtnl_mutex &____s->seqcount irq_context: 0 &dev->mutex rtnl_mutex pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex rtnl_mutex fs_reclaim irq_context: 0 &dev->mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rtnl_mutex &xa->xa_lock#4 irq_context: 0 &dev->mutex rtnl_mutex net_rwsem irq_context: 0 &dev->mutex rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 &dev->mutex rtnl_mutex &x->wait#9 irq_context: 0 &dev->mutex rtnl_mutex &k->list_lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 &dev->mutex rtnl_mutex lock irq_context: 0 &dev->mutex rtnl_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex rtnl_mutex bus_type_sem irq_context: 0 &dev->mutex rtnl_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex rtnl_mutex &dev->power.lock irq_context: 0 &dev->mutex rtnl_mutex dpm_list_mtx irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex rtnl_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex rtnl_mutex &k->k_lock irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#20 irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 &dev->mutex rtnl_mutex &dir->lock#2 irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rtnl_mutex &rq->__lock irq_context: 0 &dev->mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 &dev->mutex rtnl_mutex dev_hotplug_mutex irq_context: 0 &dev->mutex rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 &dev->mutex rtnl_mutex dev_base_lock irq_context: 0 &dev->mutex rtnl_mutex input_pool.lock irq_context: 0 &dev->mutex rtnl_mutex pcpu_alloc_mutex irq_context: 0 &dev->mutex rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &dev->mutex rtnl_mutex batched_entropy_u32.lock irq_context: 0 &dev->mutex rtnl_mutex &tbl->lock irq_context: 0 &dev->mutex rtnl_mutex sysctl_lock irq_context: 0 &dev->mutex rtnl_mutex nl_table_lock irq_context: 0 &dev->mutex rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)gve irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) irq_context: 0 hnae3_common_lock irq_context: 0 subsys mutex#52 irq_context: 0 subsys mutex#52 &k->k_lock irq_context: 0 compressor_list_lock irq_context: 0 compressor_list_lock pool_lock#2 irq_context: 0 compressor_list_lock &c->lock irq_context: 0 compressor_list_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem hwsim_netgroup_ida.xa_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim &rq->__lock irq_context: 0 hwsim_radio_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim pool_lock#2 irq_context: 0 subsys mutex#53 irq_context: 0 subsys mutex#53 &k->k_lock irq_context: 0 deferred_probe_mutex irq_context: 0 rtnl_mutex param_lock irq_context: 0 rtnl_mutex param_lock rate_ctrl_mutex irq_context: 0 rtnl_mutex (console_sem).lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &k->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex &k->list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex kobj_ns_type_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx bus_type_sem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx sysfs_symlink_target_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &dev->power.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dpm_list_mtx irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx running_helpers_waitq.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &k->k_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 &k->k_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx reg_requests_lock irq_context: 0 rtnl_mutex &base->lock irq_context: 0 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 rfkill_global_mutex irq_context: 0 rfkill_global_mutex fs_reclaim irq_context: 0 rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rfkill_global_mutex pool_lock#2 irq_context: 0 rfkill_global_mutex &k->list_lock irq_context: 0 rfkill_global_mutex lock irq_context: 0 rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rfkill_global_mutex bus_type_sem irq_context: 0 rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 rfkill_global_mutex &c->lock irq_context: 0 rfkill_global_mutex &____s->seqcount irq_context: 0 rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 rfkill_global_mutex &dev->power.lock irq_context: 0 rfkill_global_mutex dpm_list_mtx irq_context: 0 rfkill_global_mutex &rfkill->lock irq_context: 0 rfkill_global_mutex uevent_sock_mutex irq_context: 0 rfkill_global_mutex &obj_hash[i].lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rfkill_global_mutex running_helpers_waitq.lock irq_context: 0 rfkill_global_mutex &k->k_lock irq_context: 0 rfkill_global_mutex subsys mutex#40 irq_context: 0 rfkill_global_mutex subsys mutex#40 &k->k_lock irq_context: 0 rfkill_global_mutex triggers_list_lock irq_context: 0 rfkill_global_mutex leds_list_lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfkill_global_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfkill_global_mutex.wait_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx stack_depot_init_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#4 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx net_rwsem irq_context: 0 rtnl_mutex &rdev->wiphy.mtx net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &x->wait#9 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &k->k_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &dir->lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &c->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &dev->power.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx dev_base_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx input_pool.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx batched_entropy_u32.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &tbl->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx sysctl_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &local->iflist_mtx irq_context: 0 hwsim_radio_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rfkill_global_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex crngs.lock irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 &dev->mutex cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#55 irq_context: 0 &dev->mutex rtnl_mutex subsys mutex#55 &k->k_lock irq_context: 0 &dev->mutex rtnl_mutex stack_depot_init_mutex irq_context: 0 &dev->mutex rtnl_mutex crngs.lock irq_context: 0 &dev->mutex rtnl_mutex &sdata->sec_mtx irq_context: 0 &dev->mutex rtnl_mutex &sdata->sec_mtx &sec->lock irq_context: 0 &dev->mutex rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 &dev->mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex rtnl_mutex &local->iflist_mtx#2 irq_context: 0 &dev->mutex hwsim_phys_lock irq_context: 0 &dev->mutex nl_table_lock irq_context: 0 &dev->mutex nl_table_wait.lock irq_context: 0 &dev->mutex rtnl_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex hwsim_phys_lock fs_reclaim irq_context: 0 &dev->mutex hwsim_phys_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex hwsim_phys_lock pool_lock#2 irq_context: hardirq &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 xdomain_lock irq_context: 0 xdomain_lock fs_reclaim irq_context: 0 xdomain_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 xdomain_lock pool_lock#2 irq_context: 0 ioctl_mutex irq_context: 0 address_handler_list_lock irq_context: 0 card_mutex irq_context: 0 subsys mutex#56 irq_context: 0 subsys mutex#56 &k->k_lock irq_context: 0 &x->wait#18 irq_context: 0 &x->wait#18 &p->pi_lock irq_context: 0 &x->wait#18 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#18 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &txlock irq_context: 0 &txlock &list->lock#3 irq_context: 0 &txlock &txwq irq_context: 0 &iocq[i].lock irq_context: 0 &iocq[i].lock &ktiowq[i] irq_context: 0 &txwq irq_context: 0 &txwq &p->pi_lock irq_context: 0 &txwq &p->pi_lock &rq->__lock irq_context: 0 &txwq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh pool_lock#2 irq_context: 0 subsys mutex#57 irq_context: 0 subsys mutex#57 &k->k_lock irq_context: 0 usb_bus_idr_lock irq_context: 0 usb_bus_idr_lock (usb_notifier_list).rwsem irq_context: 0 table_lock irq_context: 0 table_lock &k->list_lock irq_context: 0 table_lock fs_reclaim irq_context: 0 table_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 table_lock pool_lock#2 irq_context: 0 table_lock lock irq_context: 0 table_lock lock kernfs_idr_lock irq_context: 0 table_lock &root->kernfs_rwsem irq_context: 0 table_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 table_lock &k->k_lock irq_context: 0 table_lock uevent_sock_mutex irq_context: 0 table_lock &obj_hash[i].lock irq_context: 0 table_lock rcu_read_lock &pool->lock irq_context: 0 table_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 table_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 table_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 table_lock running_helpers_waitq.lock irq_context: 0 table_lock (console_sem).lock irq_context: 0 table_lock console_lock console_srcu console_owner_lock irq_context: 0 table_lock console_lock console_srcu console_owner irq_context: 0 table_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 table_lock console_lock console_srcu console_owner console_owner_lock irq_context: softirq lib/debugobjects.c:101 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ipmi_timer) irq_context: softirq (&ipmi_timer) &ipmi_interfaces_srcu irq_context: 0 table_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 table_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 table_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 table_lock &rq->__lock irq_context: 0 table_lock &c->lock irq_context: 0 table_lock &____s->seqcount irq_context: 0 table_lock &obj_hash[i].lock pool_lock irq_context: 0 table_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 table_lock &pcp->lock &zone->lock irq_context: 0 table_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 table_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 table_lock batched_entropy_u8.lock irq_context: 0 table_lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#6/1 &obj_hash[i].lock irq_context: 0 &dev->mutex devtree_lock irq_context: 0 &dev->mutex usb_bus_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem fs_reclaim irq_context: 0 &dev->mutex (usb_notifier_list).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem pin_fs_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &x->wait#9 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &k->list_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &k->list_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex fs_reclaim irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem bus_type_sem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem sysfs_symlink_target_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &c->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &dev->power.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem dpm_list_mtx irq_context: 0 &dev->mutex (usb_notifier_list).rwsem req_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &x->wait#11 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem uevent_sock_mutex irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem running_helpers_waitq.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &k->k_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem subsys mutex#57 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem subsys mutex#57 &k->k_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem mon_lock irq_context: 0 &dev->mutex usb_port_peer_mutex irq_context: 0 &dev->mutex device_state_lock irq_context: 0 &dev->mutex usb_bus_idr_lock mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &rq->__lock irq_context: softirq &bh->lock irq_context: softirq lock#6 irq_context: softirq lock#6 kcov_remote_lock irq_context: softirq &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock (console_sem).lock irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex usb_bus_idr_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock input_pool.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock bus_type_sem irq_context: 0 &dev->mutex usb_bus_idr_lock sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock dpm_list_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock req_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &x->wait#11 irq_context: 0 &dev->mutex usb_bus_idr_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock uevent_sock_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock running_helpers_waitq.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex device_links_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex set_config_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex devtree_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &x->wait#9 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex device_state_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex bus_type_sem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex dpm_list_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &new_driver->dynids.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex device_links_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &dev->devres_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex pinctrl_list_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex pinctrl_maps_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex (console_sem).lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &dum_hcd->dum->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &dum_hcd->dum->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &bh->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &x->wait#19 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &x->wait#9 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &k->list_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex bus_type_sem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dpm_list_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &k->k_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &dev->power.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx pm_qos_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex component_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex device_links_srcu irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx fs_reclaim irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx lock kernfs_idr_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex subsys mutex#58 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &x->wait#9 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex (usb_notifier_list).rwsem irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex deferred_probe_mutex irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &lock->wait_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &lock->wait_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock subsys mutex#58 irq_context: 0 &dev->mutex usb_bus_idr_lock &x->wait#9 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fs_reclaim irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &dum_hcd->dum->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &x->wait#19 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &x->wait#19 &p->pi_lock irq_context: softirq &x->wait#19 &p->pi_lock &rq->__lock irq_context: softirq &x->wait#19 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &x->wait#19 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex (&timer.timer) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex hcd_root_hub_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)usb_hub_wq irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) lock#6 irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &lock->wait_lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &lock->wait_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &rq->__lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock &dev->power.wait_queue irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)usb_hub_wq (work_completion)(&hub->events) &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &hub->irq_urb_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) (&hub->irq_urb_retry) irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &base->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_urb_unlink_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock hcd_urb_list_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &bh->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &p->pi_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) usb_kill_urb_queue.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq usb_kill_urb_queue.lock irq_context: softirq usb_kill_urb_queue.lock &p->pi_lock irq_context: softirq usb_kill_urb_queue.lock &p->pi_lock &rq->__lock irq_context: softirq usb_kill_urb_queue.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) (work_completion)(&hub->tt.clear_work) irq_context: 0 &dev->mutex udc_lock irq_context: 0 &dev->mutex subsys mutex#59 irq_context: 0 &dev->mutex subsys mutex#59 &k->k_lock irq_context: 0 &dev->mutex gadget_id_numbers.xa_lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) kernfs_notify_lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&gadget->work) kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events kernfs_notify_work irq_context: 0 (wq_completion)events kernfs_notify_work kernfs_notify_lock irq_context: 0 (wq_completion)events kernfs_notify_work &root->kernfs_supers_rwsem irq_context: 0 &dev->mutex subsys mutex#60 irq_context: 0 func_lock irq_context: 0 g_tf_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) pool_lock#2 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dum_hcd->dum->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) device_state_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_urb_list_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &vhci_hcd->vhci->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &vhci_hcd->vhci->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &vhci_hcd->vhci->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &vhci_hcd->vhci->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &vhci_hcd->vhci->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 batched_entropy_u8.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 kfence_freelist_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &meta->lock irq_context: 0 &dev->mutex usb_bus_idr_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &x->wait#19 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) fs_reclaim irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &x->wait#19 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) (&timer.timer) irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &c->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &base->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex usb_bus_idr_lock (&timer.timer) irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex dev_pm_qos_sysfs_mtx lock kernfs_idr_lock pool_lock#2 irq_context: softirq usb_kill_urb_queue.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &____s->seqcount irq_context: 0 reading_mutex &rq->__lock irq_context: 0 reading_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &x->wait#14 &p->pi_lock irq_context: hardirq &x->wait#14 &p->pi_lock &rq->__lock irq_context: hardirq &x->wait#14 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex &hub->status_mutex &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock batched_entropy_u8.lock irq_context: 0 &dev->mutex usb_bus_idr_lock kfence_freelist_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &(&priv->bus_notifier)->rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex hcd->bandwidth_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock &____s->seqcount irq_context: hardirq &x->wait#4 irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) hcd_root_hub_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock quarantine_lock irq_context: 0 &dev->mutex usb_bus_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex usb_bus_idr_lock batched_entropy_u8.lock crngs.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &dev->mutex &dev->mutex usb_port_peer_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#2 &c->lock irq_context: 0 &type->i_mutex_dir_key#2 &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (usb_notifier_list).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex usb_bus_idr_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex usb_bus_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex (usb_notifier_list).rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &dev->mutex (usb_notifier_list).rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &pcp->lock &zone->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&hub->init_work)->work) &dev->mutex &hub->status_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &c->lock irq_context: 0 &dev->mutex (usb_notifier_list).rwsem gdp_mutex &____s->seqcount irq_context: 0 &dev->mutex usb_bus_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex usb_bus_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) pool_lock#2 irq_context: 0 &dev->mutex &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex &dev->mutex &lock->wait_lock irq_context: 0 &dev->mutex &dev->mutex &rq->__lock irq_context: 0 &dev->mutex &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &lock->wait_lock irq_context: 0 &dev->mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &dev->mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 input_ida.xa_lock irq_context: 0 input_ida.xa_lock &____s->seqcount irq_context: 0 input_ida.xa_lock pool_lock#2 irq_context: 0 subsys mutex#31 irq_context: 0 subsys mutex#31 &k->k_lock irq_context: 0 misc_mtx lock kernfs_idr_lock &c->lock irq_context: 0 misc_mtx lock kernfs_idr_lock &____s->seqcount irq_context: 0 input_mutex input_ida.xa_lock irq_context: 0 input_mutex fs_reclaim irq_context: 0 input_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 input_mutex pool_lock#2 irq_context: 0 input_mutex &x->wait#9 irq_context: 0 input_mutex &obj_hash[i].lock irq_context: 0 input_mutex &dev->mutex#2 irq_context: 0 input_mutex chrdevs_lock irq_context: 0 input_mutex &k->list_lock irq_context: 0 input_mutex lock irq_context: 0 input_mutex lock kernfs_idr_lock irq_context: 0 input_mutex &root->kernfs_rwsem irq_context: 0 input_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 input_mutex bus_type_sem irq_context: 0 input_mutex sysfs_symlink_target_lock irq_context: 0 input_mutex &root->kernfs_rwsem irq_context: 0 input_mutex &c->lock irq_context: 0 input_mutex &____s->seqcount irq_context: 0 input_mutex &dev->power.lock irq_context: 0 input_mutex dpm_list_mtx irq_context: 0 input_mutex req_lock irq_context: 0 input_mutex &p->pi_lock irq_context: 0 input_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 input_mutex &x->wait#11 irq_context: 0 input_mutex &rq->__lock irq_context: 0 input_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 input_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 input_mutex uevent_sock_mutex irq_context: 0 input_mutex rcu_read_lock &pool->lock irq_context: 0 input_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 input_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 input_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 input_mutex running_helpers_waitq.lock irq_context: 0 input_mutex &k->k_lock irq_context: 0 input_mutex subsys mutex#31 irq_context: 0 input_mutex subsys mutex#31 &k->k_lock irq_context: 0 input_mutex &p->pi_lock &rq->__lock irq_context: 0 input_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 input_mutex rcu_read_lock &rq->__lock irq_context: 0 input_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 serio_event_lock irq_context: 0 serio_event_lock pool_lock#2 irq_context: 0 serio_event_lock rcu_read_lock &pool->lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 serio_event_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 serio_event_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long irq_context: 0 (wq_completion)events_long serio_event_work irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex serio_event_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &k->list_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &k->k_lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_long serio_event_work serio_mutex pool_lock#2 irq_context: 0 &new_driver->dynids.lock irq_context: 0 &dev->mutex (efi_runtime_lock).lock irq_context: 0 &dev->mutex &x->wait#12 irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) &x->wait#12 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) &x->wait#12 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)efi_rts_wq (work_completion)(&efi_rts_work.work) &x->wait#12 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex rtc_ida.xa_lock irq_context: 0 &dev->mutex &rtc->ops_lock irq_context: 0 &dev->mutex &rtc->ops_lock (efi_runtime_lock).lock irq_context: 0 &dev->mutex &rtc->ops_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex &rtc->ops_lock &x->wait#12 irq_context: 0 &dev->mutex &rtc->ops_lock &rq->__lock irq_context: 0 &dev->mutex &rtc->ops_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &rtc->ops_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &rtc->ops_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex chrdevs_lock irq_context: 0 &dev->mutex req_lock irq_context: 0 &dev->mutex &x->wait#11 irq_context: 0 &dev->mutex subsys mutex#27 irq_context: 0 &dev->mutex subsys mutex#27 &k->k_lock irq_context: 0 &dev->mutex subsys mutex#27 fs_reclaim irq_context: 0 &dev->mutex subsys mutex#27 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex subsys mutex#27 pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 &x->wait#9 irq_context: 0 &dev->mutex subsys mutex#27 &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 platform_devid_ida.xa_lock irq_context: 0 &dev->mutex subsys mutex#27 &k->list_lock irq_context: 0 &dev->mutex subsys mutex#27 lock irq_context: 0 &dev->mutex subsys mutex#27 lock kernfs_idr_lock irq_context: 0 &dev->mutex subsys mutex#27 &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex subsys mutex#27 bus_type_sem irq_context: 0 &dev->mutex subsys mutex#27 sysfs_symlink_target_lock irq_context: 0 &dev->mutex subsys mutex#27 &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->power.lock irq_context: 0 &dev->mutex subsys mutex#27 dpm_list_mtx irq_context: 0 &dev->mutex subsys mutex#27 &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex subsys mutex#27 &(&priv->bus_notifier)->rwsem iommu_probe_device_lock irq_context: 0 &dev->mutex subsys mutex#27 &(&priv->bus_notifier)->rwsem iommu_probe_device_lock iommu_device_lock irq_context: 0 &dev->mutex subsys mutex#27 uevent_sock_mutex irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex subsys mutex#27 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex subsys mutex#27 running_helpers_waitq.lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &dev->power.lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &k->list_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &k->k_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex device_links_srcu irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex device_links_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex fs_reclaim irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &dev->devres_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex pinctrl_list_mutex irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex pinctrl_maps_mutex irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex subsys mutex#27 &dev->mutex probe_waitqueue.lock irq_context: 0 &dev->mutex subsys mutex#27 subsys mutex#4 irq_context: 0 &dev->mutex subsys mutex#27 &c->lock irq_context: 0 &dev->mutex subsys mutex#27 &____s->seqcount irq_context: 0 &dev->mutex subsys mutex#27 wakeup_ida.xa_lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &k->list_lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex fs_reclaim irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &____s->seqcount irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex subsys mutex#27 gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex subsys mutex#27 subsys mutex#13 irq_context: 0 &dev->mutex subsys mutex#27 subsys mutex#13 &k->k_lock irq_context: 0 &dev->mutex subsys mutex#27 events_lock irq_context: 0 &dev->mutex subsys mutex#27 rtcdev_lock irq_context: 0 fs_reclaim &rq->__lock irq_context: 0 g_smscore_deviceslock irq_context: 0 g_smscore_deviceslock fs_reclaim irq_context: 0 g_smscore_deviceslock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 g_smscore_deviceslock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock pool_lock#2 irq_context: 0 cx231xx_devlist_mutex irq_context: 0 em28xx_devlist_mutex irq_context: 0 pvr2_context_sync_data.lock irq_context: 0 &dev->mutex core_lock irq_context: 0 &dev->mutex core_lock fs_reclaim irq_context: 0 &dev->mutex core_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex core_lock pool_lock#2 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem fs_reclaim irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem pool_lock#2 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem i2c_dev_list_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &x->wait#9 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem chrdevs_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &k->list_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex &k->list_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex fs_reclaim irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem bus_type_sem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem sysfs_symlink_target_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &c->lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &____s->seqcount irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &dev->power.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem dpm_list_mtx irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem req_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &x->wait#11 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &rq->__lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem uevent_sock_mutex irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem running_helpers_waitq.lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem &k->k_lock irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem subsys mutex#61 irq_context: 0 &dev->mutex &(&priv->bus_notifier)->rwsem subsys mutex#61 &k->k_lock irq_context: 0 &dev->mutex subsys mutex#62 irq_context: 0 &dev->mutex pin_fs_lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &dev->mutex core_lock &k->list_lock irq_context: 0 &dev->mutex core_lock &k->k_lock irq_context: 0 &dev->mutex dvbdev_register_lock irq_context: 0 &dev->mutex dvbdev_register_lock (console_sem).lock irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex dvbdev_register_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex (kmod_concurrent_max).lock irq_context: 0 &dev->mutex &x->wait#17 irq_context: 0 &dev->mutex &dev->mutex &dev->power.lock &dev->power.lock/1 irq_context: 0 &dev->mutex &dev->mutex &dev->power.lock &dev->power.wait_queue irq_context: 0 &dev->mutex &dev->mutex device_links_srcu irq_context: 0 &dev->mutex &dev->mutex fwnode_link_lock irq_context: 0 &dev->mutex &dev->mutex device_links_lock irq_context: 0 &dev->mutex &dev->mutex fs_reclaim irq_context: 0 &dev->mutex &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &dev->mutex &c->lock irq_context: 0 &dev->mutex &dev->mutex &____s->seqcount irq_context: 0 &dev->mutex &dev->mutex pool_lock#2 irq_context: 0 &dev->mutex &dev->mutex &dev->devres_lock irq_context: 0 &dev->mutex &dev->mutex pinctrl_list_mutex irq_context: 0 &dev->mutex &dev->mutex pinctrl_maps_mutex irq_context: 0 &dev->mutex &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 &dev->mutex &dev->mutex &obj_hash[i].lock irq_context: 0 &dev->mutex &dev->mutex &(&priv->bus_notifier)->rwsem irq_context: 0 &dev->mutex &dev->mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex &dev->mutex lock irq_context: 0 &dev->mutex &dev->mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex &dev->mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &dev->mutex deferred_probe_mutex irq_context: 0 &dev->mutex &dev->mutex uevent_sock_mutex irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex &dev->mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex frontend_mutex irq_context: 0 &dev->mutex frontend_mutex fs_reclaim irq_context: 0 &dev->mutex frontend_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex (console_sem).lock irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex frontend_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock fs_reclaim irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock minor_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &xa->xa_lock#12 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &mdev->graph_mutex pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock (console_sem).lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &x->wait#9 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &obj_hash[i].lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &k->list_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex &k->list_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex fs_reclaim irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &c->lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &____s->seqcount irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock bus_type_sem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock sysfs_symlink_target_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &dev->power.lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock dpm_list_mtx irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock req_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &p->pi_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &x->wait#11 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &rq->__lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock uevent_sock_mutex irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock running_helpers_waitq.lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock &k->k_lock irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock subsys mutex#63 irq_context: 0 &dev->mutex frontend_mutex dvbdev_register_lock subsys mutex#63 &k->k_lock irq_context: 0 &dev->mutex &dmxdev->lock irq_context: 0 &dev->mutex dvbdev_register_lock fs_reclaim irq_context: 0 &dev->mutex dvbdev_register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex dvbdev_register_lock pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock minor_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex dvbdev_register_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex dvbdev_register_lock &____s->seqcount irq_context: 0 &dev->mutex dvbdev_register_lock &xa->xa_lock#12 irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex irq_context: 0 &dev->mutex dvbdev_register_lock &xa->xa_lock#12 pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex dvbdev_register_lock &c->lock irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex &c->lock irq_context: 0 &dev->mutex dvbdev_register_lock &mdev->graph_mutex &____s->seqcount irq_context: 0 &dev->mutex dvbdev_register_lock &x->wait#9 irq_context: 0 &dev->mutex dvbdev_register_lock &obj_hash[i].lock irq_context: 0 &dev->mutex dvbdev_register_lock &k->list_lock irq_context: 0 &dev->mutex dvbdev_register_lock gdp_mutex irq_context: 0 &dev->mutex dvbdev_register_lock gdp_mutex &k->list_lock irq_context: 0 &dev->mutex dvbdev_register_lock lock irq_context: 0 &dev->mutex dvbdev_register_lock lock kernfs_idr_lock irq_context: 0 &dev->mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock bus_type_sem irq_context: 0 &dev->mutex dvbdev_register_lock sysfs_symlink_target_lock irq_context: 0 &dev->mutex dvbdev_register_lock &root->kernfs_rwsem irq_context: 0 &dev->mutex dvbdev_register_lock &dev->power.lock irq_context: 0 &dev->mutex dvbdev_register_lock dpm_list_mtx irq_context: 0 &dev->mutex dvbdev_register_lock req_lock irq_context: 0 &dev->mutex dvbdev_register_lock &p->pi_lock irq_context: 0 &dev->mutex dvbdev_register_lock &x->wait#11 irq_context: 0 &dev->mutex dvbdev_register_lock &rq->__lock irq_context: 0 &dev->mutex dvbdev_register_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dvbdev_register_lock uevent_sock_mutex irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex dvbdev_register_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex dvbdev_register_lock running_helpers_waitq.lock irq_context: 0 &dev->mutex dvbdev_register_lock &k->k_lock irq_context: 0 &dev->mutex dvbdev_register_lock subsys mutex#63 irq_context: 0 &dev->mutex dvbdev_register_lock subsys mutex#63 &k->k_lock irq_context: 0 &dev->mutex &dvbdemux->mutex irq_context: 0 &dev->mutex media_devnode_lock irq_context: 0 &dev->mutex subsys mutex#64 irq_context: 0 &dev->mutex videodev_lock irq_context: 0 &dev->mutex subsys mutex#65 irq_context: 0 &dev->mutex subsys mutex#65 &k->k_lock irq_context: 0 &dev->mutex &xa->xa_lock#12 irq_context: 0 &dev->mutex &mdev->graph_mutex irq_context: 0 &dev->mutex &mdev->graph_mutex fs_reclaim irq_context: 0 &dev->mutex &mdev->graph_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &mdev->graph_mutex pool_lock#2 irq_context: 0 &dev->mutex gdp_mutex &____s->seqcount irq_context: 0 &dev->mutex vimc_sensor:396:(&vsensor->hdl)->_lock irq_context: 0 &dev->mutex &v4l2_dev->lock irq_context: 0 &dev->mutex vimc_debayer:581:(&vdebayer->hdl)->_lock irq_context: 0 &dev->mutex vimc_lens:61:(&vlens->hdl)->_lock irq_context: 0 &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex tk_core.seq.seqcount irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1618:(hdl_fb)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1627:(hdl_vbi_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1630:(hdl_radio_rx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1632:(hdl_radio_tx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1634:(hdl_sdr_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1636:(hdl_meta_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1638:(hdl_meta_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1640:(hdl_tch_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &c->lock irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1608:(hdl_user_vid)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock vivid_ctrls:1620:(hdl_vid_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock fs_reclaim irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock pool_lock#2 irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1622:(hdl_vid_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1614:(hdl_sdtv_cap)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1616:(hdl_loop_cap)->_lock vivid_ctrls:1625:(hdl_vbi_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1627:(hdl_vbi_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1627:(hdl_vbi_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1630:(hdl_radio_rx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1630:(hdl_radio_rx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1632:(hdl_radio_tx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1610:(hdl_user_aud)->_lock vivid_ctrls:1632:(hdl_radio_tx)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1634:(hdl_sdr_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1634:(hdl_sdr_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1636:(hdl_meta_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1636:(hdl_meta_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1638:(hdl_meta_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1638:(hdl_meta_out)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock vivid_ctrls:1640:(hdl_tch_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1612:(hdl_streaming)->_lock vivid_ctrls:1640:(hdl_tch_cap)->_lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &zone->lock irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex vivid_ctrls:1606:(hdl_user_gen)->_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &adap->kthread_waitq irq_context: 0 &dev->cec_xfers_slock irq_context: 0 &dev->kthread_waitq_cec irq_context: 0 &dev->mutex cec_devnode_lock irq_context: 0 &dev->mutex subsys mutex#66 irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &dev->mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &dev->mutex &adap->lock irq_context: 0 &dev->mutex &adap->lock tk_core.seq.seqcount irq_context: 0 &dev->mutex &adap->lock &adap->devnode.lock_fhs irq_context: 0 ptp_clocks_map.xa_lock irq_context: 0 subsys mutex#67 irq_context: 0 subsys mutex#67 &k->k_lock irq_context: 0 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pers_lock irq_context: 0 lock pidmap_lock &c->lock irq_context: 0 lock pidmap_lock &____s->seqcount irq_context: 0 _lock irq_context: 0 dm_bufio_clients_lock irq_context: 0 _ps_lock irq_context: 0 _lock#2 irq_context: 0 _lock#3 irq_context: 0 register_lock#2 irq_context: 0 subsys mutex#68 irq_context: 0 subsys mutex#68 &k->k_lock irq_context: 0 bp_lock irq_context: 0 bp_lock irq_context: 0 subsys mutex#69 irq_context: 0 subsys mutex#69 &k->k_lock irq_context: 0 free_vmap_area_lock &obj_hash[i].lock irq_context: 0 free_vmap_area_lock pool_lock#2 irq_context: softirq (&dsp_spl_tl) irq_context: softirq (&dsp_spl_tl) dsp_lock irq_context: softirq (&dsp_spl_tl) dsp_lock iclock_lock irq_context: softirq (&dsp_spl_tl) dsp_lock iclock_lock tk_core.seq.seqcount irq_context: softirq (&dsp_spl_tl) dsp_lock &obj_hash[i].lock irq_context: softirq (&dsp_spl_tl) dsp_lock &base->lock irq_context: softirq (&dsp_spl_tl) dsp_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex lock#7 irq_context: 0 iscsi_transport_lock irq_context: 0 subsys mutex#70 irq_context: 0 subsys mutex#70 &k->k_lock irq_context: 0 &tx_task->waiting irq_context: 0 link_ops_rwsem irq_context: 0 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 disable_lock irq_context: 0 disable_lock fs_reclaim irq_context: 0 disable_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 disable_lock &c->lock irq_context: 0 disable_lock &pcp->lock &zone->lock irq_context: 0 disable_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 disable_lock &____s->seqcount irq_context: 0 disable_lock pool_lock#2 irq_context: 0 disable_lock &x->wait#9 irq_context: 0 disable_lock &obj_hash[i].lock irq_context: 0 disable_lock &k->list_lock irq_context: 0 disable_lock lock irq_context: 0 disable_lock lock kernfs_idr_lock irq_context: 0 disable_lock &root->kernfs_rwsem irq_context: 0 disable_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 disable_lock bus_type_sem irq_context: 0 disable_lock sysfs_symlink_target_lock irq_context: 0 disable_lock &k->k_lock irq_context: 0 disable_lock &root->kernfs_rwsem irq_context: 0 disable_lock &dev->power.lock irq_context: 0 disable_lock dpm_list_mtx irq_context: 0 disable_lock &(&priv->bus_notifier)->rwsem irq_context: 0 disable_lock &(&priv->bus_notifier)->rwsem iommu_probe_device_lock irq_context: 0 disable_lock &(&priv->bus_notifier)->rwsem iommu_probe_device_lock iommu_device_lock irq_context: 0 disable_lock uevent_sock_mutex irq_context: 0 disable_lock rcu_read_lock &pool->lock irq_context: 0 disable_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 disable_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 disable_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 disable_lock running_helpers_waitq.lock irq_context: 0 disable_lock &dev->mutex &dev->power.lock irq_context: 0 disable_lock &dev->mutex &k->list_lock irq_context: 0 disable_lock &dev->mutex &k->k_lock irq_context: 0 disable_lock subsys mutex#4 irq_context: 0 protocol_lock irq_context: 0 protocol_lock pool_lock#2 irq_context: 0 psinfo_lock irq_context: 0 psinfo_lock fs_reclaim irq_context: 0 psinfo_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 psinfo_lock pool_lock#2 irq_context: 0 psinfo_lock free_vmap_area_lock irq_context: 0 psinfo_lock vmap_area_lock irq_context: 0 psinfo_lock &____s->seqcount irq_context: 0 psinfo_lock init_mm.page_table_lock irq_context: 0 psinfo_lock (console_sem).lock irq_context: 0 psinfo_lock console_lock console_srcu console_owner_lock irq_context: 0 psinfo_lock console_lock console_srcu console_owner irq_context: 0 psinfo_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 psinfo_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 psinfo_lock &rq->__lock irq_context: 0 psinfo_lock pstore_sb_lock irq_context: 0 psinfo_lock dump_list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_node_0 irq_context: 0 vsock_register_mutex irq_context: 0 misc_mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 misc_mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 comedi_drivers_list_lock irq_context: 0 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &domain->mutex irq_context: 0 &domain->mutex sparse_irq_lock irq_context: 0 &domain->mutex sparse_irq_lock fs_reclaim irq_context: 0 &domain->mutex sparse_irq_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &domain->mutex sparse_irq_lock pool_lock#2 irq_context: 0 &domain->mutex sparse_irq_lock pcpu_alloc_mutex irq_context: 0 &domain->mutex sparse_irq_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 &domain->mutex sparse_irq_lock &obj_hash[i].lock irq_context: 0 &domain->mutex sparse_irq_lock &c->lock irq_context: 0 &domain->mutex sparse_irq_lock &____s->seqcount irq_context: 0 &domain->mutex sparse_irq_lock lock irq_context: 0 &domain->mutex sparse_irq_lock lock kernfs_idr_lock irq_context: 0 &domain->mutex sparse_irq_lock &root->kernfs_rwsem irq_context: 0 &domain->mutex sparse_irq_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &domain->mutex sparse_irq_lock lock kernfs_idr_lock &c->lock irq_context: 0 &domain->mutex sparse_irq_lock lock kernfs_idr_lock &____s->seqcount irq_context: 0 &domain->mutex sparse_irq_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 &domain->mutex fs_reclaim irq_context: 0 &domain->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &domain->mutex pool_lock#2 irq_context: 0 &domain->mutex &irq_desc_lock_class irq_context: 0 &desc->request_mutex irq_context: 0 &desc->request_mutex &irq_desc_lock_class irq_context: 0 &desc->request_mutex &irq_desc_lock_class irq_controller_lock irq_context: 0 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cscfg_mutex irq_context: 0 cscfg_mutex fs_reclaim irq_context: 0 cscfg_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cscfg_mutex pool_lock#2 irq_context: 0 cscfg_mutex &x->wait#9 irq_context: 0 cscfg_mutex &obj_hash[i].lock irq_context: 0 cscfg_mutex &k->list_lock irq_context: 0 cscfg_mutex lock irq_context: 0 cscfg_mutex lock kernfs_idr_lock irq_context: 0 cscfg_mutex &root->kernfs_rwsem irq_context: 0 cscfg_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cscfg_mutex bus_type_sem irq_context: 0 cscfg_mutex &root->kernfs_rwsem irq_context: 0 cscfg_mutex &dev->power.lock irq_context: 0 cscfg_mutex dpm_list_mtx irq_context: 0 fs_reclaim icc_bw_lock irq_context: 0 subsys mutex#71 irq_context: 0 subsys mutex#71 &rq->__lock irq_context: 0 subsys mutex#71 &k->k_lock irq_context: 0 snd_ctl_layer_rwsem irq_context: 0 snd_card_mutex irq_context: 0 snd_ioctl_rwsem irq_context: 0 strings irq_context: 0 strings fs_reclaim irq_context: 0 strings fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 strings pool_lock#2 irq_context: 0 register_mutex irq_context: 0 sound_mutex irq_context: 0 sound_mutex fs_reclaim irq_context: 0 sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sound_mutex pool_lock#2 irq_context: 0 sound_mutex &k->list_lock irq_context: 0 sound_mutex gdp_mutex irq_context: 0 sound_mutex gdp_mutex &k->list_lock irq_context: 0 sound_mutex lock irq_context: 0 sound_mutex lock kernfs_idr_lock irq_context: 0 sound_mutex &root->kernfs_rwsem irq_context: 0 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sound_mutex bus_type_sem irq_context: 0 sound_mutex sysfs_symlink_target_lock irq_context: 0 sound_mutex &root->kernfs_rwsem irq_context: 0 sound_mutex &dev->power.lock irq_context: 0 sound_mutex dpm_list_mtx irq_context: 0 sound_mutex req_lock irq_context: 0 sound_mutex &p->pi_lock irq_context: 0 sound_mutex &p->pi_lock &rq->__lock irq_context: 0 sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sound_mutex &x->wait#11 irq_context: 0 sound_mutex &rq->__lock irq_context: 0 sound_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sound_mutex &obj_hash[i].lock irq_context: 0 sound_mutex uevent_sock_mutex irq_context: 0 sound_mutex &c->lock irq_context: 0 sound_mutex &____s->seqcount irq_context: 0 sound_mutex rcu_read_lock &pool->lock irq_context: 0 sound_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sound_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sound_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sound_mutex running_helpers_waitq.lock irq_context: 0 sound_mutex subsys mutex#71 irq_context: 0 sound_mutex subsys mutex#71 &k->k_lock irq_context: 0 register_mutex#2 irq_context: 0 register_mutex#3 irq_context: 0 register_mutex#3 fs_reclaim irq_context: 0 register_mutex#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#3 pool_lock#2 irq_context: 0 register_mutex#3 sound_mutex irq_context: 0 register_mutex#3 sound_mutex fs_reclaim irq_context: 0 register_mutex#3 sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#3 sound_mutex pool_lock#2 irq_context: 0 register_mutex#3 sound_mutex &k->list_lock irq_context: 0 register_mutex#3 sound_mutex gdp_mutex irq_context: 0 register_mutex#3 sound_mutex gdp_mutex &k->list_lock irq_context: 0 register_mutex#3 sound_mutex lock irq_context: 0 register_mutex#3 sound_mutex lock kernfs_idr_lock irq_context: 0 register_mutex#3 sound_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 register_mutex#3 sound_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#3 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 register_mutex#3 sound_mutex bus_type_sem irq_context: 0 register_mutex#3 sound_mutex &c->lock irq_context: 0 register_mutex#3 sound_mutex &____s->seqcount irq_context: 0 register_mutex#3 sound_mutex sysfs_symlink_target_lock irq_context: 0 register_mutex#3 sound_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#3 sound_mutex &dev->power.lock irq_context: 0 register_mutex#3 sound_mutex dpm_list_mtx irq_context: 0 register_mutex#3 sound_mutex req_lock irq_context: 0 register_mutex#3 sound_mutex &p->pi_lock irq_context: 0 register_mutex#3 sound_mutex &p->pi_lock &rq->__lock irq_context: 0 register_mutex#3 sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#3 sound_mutex &rq->__lock irq_context: 0 register_mutex#3 sound_mutex &cfs_rq->removed.lock irq_context: 0 register_mutex#3 sound_mutex &obj_hash[i].lock irq_context: 0 register_mutex#3 sound_mutex &x->wait#11 irq_context: 0 register_mutex#3 sound_mutex uevent_sock_mutex irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 register_mutex#3 sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#3 sound_mutex running_helpers_waitq.lock irq_context: 0 register_mutex#3 sound_mutex subsys mutex#71 irq_context: 0 register_mutex#3 sound_mutex subsys mutex#71 &k->k_lock irq_context: 0 register_mutex#3 clients_lock irq_context: 0 &client->ports_mutex irq_context: 0 &client->ports_mutex &client->ports_lock irq_context: 0 register_mutex#4 irq_context: 0 register_mutex#4 fs_reclaim irq_context: 0 register_mutex#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#4 pool_lock#2 irq_context: 0 register_mutex#4 sound_oss_mutex irq_context: 0 register_mutex#4 sound_oss_mutex fs_reclaim irq_context: 0 register_mutex#4 sound_oss_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 register_mutex#4 sound_oss_mutex pool_lock#2 irq_context: 0 register_mutex#4 sound_oss_mutex sound_loader_lock irq_context: 0 register_mutex#4 sound_oss_mutex &x->wait#9 irq_context: 0 register_mutex#4 sound_oss_mutex &obj_hash[i].lock irq_context: 0 register_mutex#4 sound_oss_mutex &k->list_lock irq_context: 0 register_mutex#4 sound_oss_mutex gdp_mutex irq_context: 0 register_mutex#4 sound_oss_mutex gdp_mutex &k->list_lock irq_context: 0 register_mutex#4 sound_oss_mutex &____s->seqcount irq_context: 0 register_mutex#4 sound_oss_mutex lock irq_context: 0 register_mutex#4 sound_oss_mutex lock kernfs_idr_lock irq_context: 0 register_mutex#4 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#4 sound_oss_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 register_mutex#4 sound_oss_mutex bus_type_sem irq_context: 0 register_mutex#4 sound_oss_mutex sysfs_symlink_target_lock irq_context: 0 register_mutex#4 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 register_mutex#4 sound_oss_mutex &c->lock irq_context: 0 register_mutex#4 sound_oss_mutex &dev->power.lock irq_context: 0 register_mutex#4 sound_oss_mutex dpm_list_mtx irq_context: 0 register_mutex#4 sound_oss_mutex req_lock irq_context: 0 register_mutex#4 sound_oss_mutex &p->pi_lock irq_context: 0 register_mutex#4 sound_oss_mutex &p->pi_lock &rq->__lock irq_context: 0 register_mutex#4 sound_oss_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 register_mutex#4 sound_oss_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 register_mutex#4 sound_oss_mutex &rq->__lock irq_context: 0 register_mutex#4 sound_oss_mutex &x->wait#11 irq_context: 0 register_mutex#4 sound_oss_mutex uevent_sock_mutex irq_context: 0 register_mutex#4 sound_oss_mutex &obj_hash[i].lock pool_lock irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 register_mutex#4 sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 register_mutex#4 sound_oss_mutex running_helpers_waitq.lock irq_context: 0 register_mutex#4 sound_oss_mutex subsys mutex#71 irq_context: 0 register_mutex#4 sound_oss_mutex subsys mutex#71 &k->k_lock irq_context: 0 clients_lock irq_context: 0 &client->ports_lock irq_context: 0 &grp->list_mutex/1 irq_context: 0 &grp->list_mutex#2 irq_context: 0 &grp->list_mutex#2 &grp->list_lock irq_context: 0 &grp->list_mutex/1 clients_lock irq_context: 0 &grp->list_mutex/1 &client->ports_lock irq_context: 0 (wq_completion)events async_lookup_work irq_context: 0 (wq_completion)events async_lookup_work fs_reclaim irq_context: 0 (wq_completion)events async_lookup_work fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events async_lookup_work pool_lock#2 irq_context: 0 (wq_completion)events async_lookup_work clients_lock irq_context: 0 (wq_completion)events async_lookup_work &client->ports_lock irq_context: 0 (wq_completion)events async_lookup_work snd_card_mutex irq_context: 0 (wq_completion)events async_lookup_work (kmod_concurrent_max).lock irq_context: 0 (wq_completion)events async_lookup_work &obj_hash[i].lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events async_lookup_work rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events async_lookup_work &x->wait#17 irq_context: 0 (wq_completion)events async_lookup_work sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events async_lookup_work &rq->__lock irq_context: 0 (wq_completion)events async_lookup_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &grp->list_mutex/1 register_lock#3 irq_context: 0 &grp->list_mutex/1 fs_reclaim irq_context: 0 &grp->list_mutex/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &grp->list_mutex/1 pool_lock#2 irq_context: 0 &dev->mutex snd_card_mutex irq_context: 0 &dev->mutex &entry->access irq_context: 0 &dev->mutex info_mutex irq_context: 0 &dev->mutex info_mutex proc_subdir_lock irq_context: 0 &dev->mutex info_mutex fs_reclaim irq_context: 0 &dev->mutex info_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex info_mutex pool_lock#2 irq_context: 0 &dev->mutex info_mutex proc_inum_ida.xa_lock irq_context: 0 &dev->mutex info_mutex proc_subdir_lock irq_context: 0 &dev->mutex &card->controls_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 pool_lock#2 irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &card->controls_rwsem &card->ctl_files_rwlock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 &c->lock irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &sig->wait_chldexit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex lock kernfs_idr_lock &c->lock irq_context: 0 &dev->mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 &dev->mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 (wq_completion)events async_lookup_work &cfs_rq->removed.lock irq_context: 0 (wq_completion)events async_lookup_work running_helpers_waitq.lock irq_context: 0 (wq_completion)events async_lookup_work autoload_work irq_context: 0 (wq_completion)events async_lookup_work &x->wait#10 irq_context: 0 (wq_completion)events async_lookup_work sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events async_lookup_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events async_lookup_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events async_lookup_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events async_lookup_work sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events autoload_work irq_context: 0 (wq_completion)events autoload_work &k->list_lock irq_context: 0 (wq_completion)events autoload_work &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex subsys mutex#71 irq_context: 0 &dev->mutex subsys mutex#71 &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 irq_context: 0 &dev->mutex register_mutex#2 fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_mutex fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 sound_mutex pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &k->list_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex bus_type_sem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &c->lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &dev->power.lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex dpm_list_mtx irq_context: 0 &dev->mutex register_mutex#2 sound_mutex req_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &x->wait#11 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex uevent_sock_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex subsys mutex#71 irq_context: 0 &dev->mutex register_mutex#2 sound_mutex subsys mutex#71 &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 register_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex register_mutex#2 sound_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex register_mutex#2 &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex register_mutex#2 &c->lock irq_context: 0 &dev->mutex register_mutex#2 &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex sound_loader_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &x->wait#9 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &k->list_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex bus_type_sem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &dev->power.lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex dpm_list_mtx irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &c->lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex req_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &x->wait#11 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex uevent_sock_mutex irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex subsys mutex#71 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex subsys mutex#71 &k->k_lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex register_mutex#2 sound_oss_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#2 strings irq_context: 0 &dev->mutex register_mutex#2 strings fs_reclaim irq_context: 0 &dev->mutex register_mutex#2 strings fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#2 strings pool_lock#2 irq_context: 0 &dev->mutex register_mutex#2 &entry->access irq_context: 0 &dev->mutex register_mutex#2 info_mutex irq_context: 0 &dev->mutex sound_mutex irq_context: 0 &dev->mutex sound_mutex fs_reclaim irq_context: 0 &dev->mutex sound_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex sound_mutex &c->lock irq_context: 0 &dev->mutex sound_mutex &____s->seqcount irq_context: 0 &dev->mutex sound_mutex pool_lock#2 irq_context: 0 &dev->mutex sound_mutex &k->list_lock irq_context: 0 &dev->mutex sound_mutex lock irq_context: 0 &dev->mutex sound_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex sound_mutex bus_type_sem irq_context: 0 &dev->mutex sound_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex sound_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_mutex &dev->power.lock irq_context: 0 &dev->mutex sound_mutex dpm_list_mtx irq_context: 0 &dev->mutex sound_mutex req_lock irq_context: 0 &dev->mutex sound_mutex &p->pi_lock irq_context: 0 &dev->mutex sound_mutex &x->wait#11 irq_context: 0 &dev->mutex sound_mutex &rq->__lock irq_context: 0 &dev->mutex sound_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex sound_mutex uevent_sock_mutex irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex sound_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex sound_mutex &k->k_lock irq_context: 0 &dev->mutex sound_mutex subsys mutex#71 irq_context: 0 &dev->mutex sound_mutex subsys mutex#71 &k->k_lock irq_context: 0 &dev->mutex &card->controls_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem snd_ctl_led_mutex irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem fs_reclaim irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem pool_lock#2 irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &x->wait#9 irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &k->list_lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem lock kernfs_idr_lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem bus_type_sem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &c->lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &____s->seqcount irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &root->kernfs_rwsem irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &dev->power.lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem dpm_list_mtx irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem &k->k_lock irq_context: 0 &dev->mutex &card->controls_rwsem snd_ctl_layer_rwsem sysfs_symlink_target_lock irq_context: 0 &dev->mutex info_mutex &c->lock irq_context: 0 &dev->mutex info_mutex &____s->seqcount irq_context: 0 &dev->mutex sound_oss_mutex irq_context: 0 &dev->mutex sound_oss_mutex fs_reclaim irq_context: 0 &dev->mutex sound_oss_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex sound_oss_mutex pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex sound_loader_lock irq_context: 0 &dev->mutex sound_oss_mutex &x->wait#9 irq_context: 0 &dev->mutex sound_oss_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex sound_oss_mutex &k->list_lock irq_context: 0 &dev->mutex sound_oss_mutex lock irq_context: 0 &dev->mutex sound_oss_mutex lock kernfs_idr_lock irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex sound_oss_mutex bus_type_sem irq_context: 0 &dev->mutex sound_oss_mutex sysfs_symlink_target_lock irq_context: 0 &dev->mutex sound_oss_mutex &root->kernfs_rwsem irq_context: 0 &dev->mutex sound_oss_mutex &dev->power.lock irq_context: 0 &dev->mutex sound_oss_mutex dpm_list_mtx irq_context: 0 &dev->mutex sound_oss_mutex &c->lock irq_context: 0 &dev->mutex sound_oss_mutex &____s->seqcount irq_context: 0 &dev->mutex sound_oss_mutex req_lock irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock irq_context: 0 &dev->mutex sound_oss_mutex &x->wait#11 irq_context: 0 &dev->mutex sound_oss_mutex &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_oss_mutex uevent_sock_mutex irq_context: 0 &dev->mutex sound_oss_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex sound_oss_mutex running_helpers_waitq.lock irq_context: 0 &dev->mutex sound_oss_mutex &k->k_lock irq_context: 0 &dev->mutex sound_oss_mutex subsys mutex#71 irq_context: 0 &dev->mutex sound_oss_mutex subsys mutex#71 &k->k_lock irq_context: 0 &dev->mutex strings irq_context: 0 &dev->mutex strings fs_reclaim irq_context: 0 &dev->mutex strings fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex strings pool_lock#2 irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim irq_context: 0 &dev->mutex &card->controls_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &card->controls_rwsem pool_lock#2 irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 &pcp->lock &zone->lock irq_context: 0 &dev->mutex &card->controls_rwsem &xa->xa_lock#13 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex register_mutex#2 &pcp->lock &zone->lock irq_context: 0 &dev->mutex register_mutex#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex sound_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex sound_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex info_mutex &rq->__lock irq_context: 0 &dev->mutex register_mutex#5 irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex register_mutex#3 irq_context: 0 &dev->mutex register_mutex#3 fs_reclaim irq_context: 0 &dev->mutex register_mutex#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex register_mutex#3 pool_lock#2 irq_context: 0 &dev->mutex register_mutex#3 clients_lock irq_context: 0 &dev->mutex clients_lock irq_context: 0 &dev->mutex &client->ports_lock irq_context: 0 &dev->mutex &grp->list_mutex/1 irq_context: 0 &dev->mutex &grp->list_mutex/1 clients_lock irq_context: 0 &dev->mutex &grp->list_mutex/1 &client->ports_lock irq_context: 0 &dev->mutex &client->ports_mutex irq_context: 0 &dev->mutex &client->ports_mutex &client->ports_lock irq_context: 0 &dev->mutex &grp->list_mutex/1 register_lock#3 irq_context: 0 &dev->mutex &grp->list_mutex/1 fs_reclaim irq_context: 0 &dev->mutex &grp->list_mutex/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &grp->list_mutex/1 pool_lock#2 irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex sound_oss_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq allocation_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex sound_oss_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &grp->list_mutex/1 &c->lock irq_context: 0 &dev->mutex &grp->list_mutex/1 &____s->seqcount irq_context: 0 &dev->mutex sound_mutex &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex client_mutex irq_context: 0 &dev->mutex client_mutex fs_reclaim irq_context: 0 &dev->mutex client_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex client_mutex pool_lock#2 irq_context: 0 &dev->mutex client_mutex &dev->devres_lock irq_context: 0 pernet_ops_rwsem &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex failover_lock irq_context: 0 llc_sap_list_lock irq_context: 0 llc_sap_list_lock pool_lock#2 irq_context: 0 act_id_mutex irq_context: 0 act_id_mutex fs_reclaim irq_context: 0 act_id_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 act_id_mutex pool_lock#2 irq_context: 0 act_mod_lock irq_context: 0 act_id_mutex &____s->seqcount irq_context: 0 ife_mod_lock irq_context: 0 cls_mod_lock irq_context: 0 ematch_mod_lock irq_context: 0 sock_diag_table_mutex irq_context: 0 nfnl_subsys_acct irq_context: 0 nfnl_subsys_queue irq_context: 0 nfnl_subsys_ulog irq_context: 0 nf_log_mutex irq_context: 0 nfnl_subsys_osf irq_context: 0 nf_sockopt_mutex irq_context: 0 nfnl_subsys_ctnetlink irq_context: 0 nfnl_subsys_ctnetlink_exp irq_context: 0 pernet_ops_rwsem nf_ct_ecache_mutex irq_context: 0 nfnl_subsys_cttimeout irq_context: 0 nfnl_subsys_cthelper irq_context: 0 nf_ct_helper_mutex irq_context: 0 pernet_ops_rwsem nf_log_mutex irq_context: 0 nf_conntrack_expect_lock irq_context: 0 nf_ct_nat_helpers_mutex irq_context: 0 nfnl_subsys_nftables irq_context: 0 nfnl_subsys_nftcompat irq_context: 0 masq_mutex irq_context: 0 masq_mutex pernet_ops_rwsem irq_context: 0 masq_mutex pernet_ops_rwsem rtnl_mutex irq_context: 0 masq_mutex (inetaddr_chain).rwsem irq_context: 0 masq_mutex inet6addr_chain.lock irq_context: 0 &xt[i].mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock irq_context: 0 subsys mutex#72 irq_context: 0 subsys mutex#72 &k->k_lock irq_context: 0 &xt[i].mutex &rq->__lock irq_context: 0 &xt[i].mutex &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex &obj_hash[i].lock irq_context: 0 &xt[i].mutex pool_lock#2 irq_context: 0 nfnl_subsys_ipset irq_context: 0 ip_set_type_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem &base->lock irq_context: 0 pernet_ops_rwsem &base->lock &obj_hash[i].lock irq_context: 0 ip_vs_sched_mutex irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex pool_lock#2 irq_context: 0 ip_vs_pe_mutex irq_context: 0 tunnel4_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex kfence_freelist_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 xfrm4_protocol_mutex irq_context: 0 &xt[i].mutex fs_reclaim irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem net_generic_ids.xa_lock &c->lock irq_context: 0 pernet_ops_rwsem net_generic_ids.xa_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem net_generic_ids.xa_lock pool_lock#2 irq_context: 0 inet_diag_table_mutex irq_context: 0 xfrm_km_lock irq_context: 0 xfrm6_protocol_mutex irq_context: 0 tunnel6_mutex irq_context: 0 xfrm_if_cb_lock irq_context: 0 inetsw6_lock irq_context: 0 &hashinfo->lock#2 irq_context: 0 pernet_ops_rwsem &hashinfo->lock#2 irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (crypto_chain).rwsem irq_context: 0 (crypto_chain).rwsem fs_reclaim irq_context: 0 (crypto_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (crypto_chain).rwsem pool_lock#2 irq_context: 0 (crypto_chain).rwsem kthread_create_lock irq_context: 0 (crypto_chain).rwsem &p->pi_lock irq_context: 0 (crypto_chain).rwsem &p->pi_lock &rq->__lock irq_context: 0 (crypto_chain).rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (crypto_chain).rwsem &x->wait irq_context: 0 (crypto_chain).rwsem &rq->__lock irq_context: 0 (crypto_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (crypto_chain).rwsem &obj_hash[i].lock irq_context: 0 &x->wait#20 irq_context: 0 &x->wait#20 &p->pi_lock irq_context: 0 &p->alloc_lock &x->wait irq_context: 0 (crypto_chain).rwsem &c->lock irq_context: 0 (crypto_chain).rwsem &____s->seqcount irq_context: 0 rcu_read_lock &base->lock irq_context: 0 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex pool_lock#2 irq_context: 0 stp_proto_mutex irq_context: 0 stp_proto_mutex llc_sap_list_lock irq_context: 0 stp_proto_mutex llc_sap_list_lock pool_lock#2 irq_context: 0 switchdev_notif_chain.lock irq_context: 0 (switchdev_blocking_notif_chain).rwsem irq_context: 0 br_ioctl_mutex irq_context: 0 nf_ct_proto_mutex irq_context: 0 ebt_mutex irq_context: 0 ebt_mutex fs_reclaim irq_context: 0 ebt_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ebt_mutex pool_lock#2 irq_context: 0 dsa_tag_drivers_lock irq_context: 0 rtnl_mutex &tn->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock purge_vmap_area_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &obj_hash[i].lock irq_context: 0 protocol_list_lock irq_context: 0 linkfail_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rq->__lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pool_lock#2 irq_context: 0 rose_neigh_list_lock irq_context: 0 proto_tab_lock#2 irq_context: 0 bt_proto_lock irq_context: 0 bt_proto_lock &c->lock irq_context: 0 bt_proto_lock &____s->seqcount irq_context: 0 bt_proto_lock pool_lock#2 irq_context: 0 bt_proto_lock &dir->lock irq_context: 0 bt_proto_lock &obj_hash[i].lock irq_context: 0 bt_proto_lock chan_list_lock irq_context: 0 bt_proto_lock l2cap_sk_list.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_L2CAP chan_list_lock irq_context: 0 slock-AF_BLUETOOTH-BTPROTO_L2CAP irq_context: 0 rfcomm_wq.lock irq_context: 0 rfcomm_mutex irq_context: 0 auth_domain_lock irq_context: 0 registered_mechs_lock irq_context: 0 atm_dev_notify_chain.lock irq_context: 0 genl_mutex irq_context: 0 proto_tab_lock#3 irq_context: 0 vlan_ioctl_mutex irq_context: 0 pernet_ops_rwsem (console_sem).lock irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner_lock irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner &port_lock_key irq_context: 0 pernet_ops_rwsem console_lock console_srcu console_owner console_owner_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &ndev->lock irq_context: 0 rds_info_lock irq_context: 0 rds_trans_sem irq_context: 0 rds_trans_sem (console_sem).lock irq_context: 0 rds_trans_sem console_lock console_srcu console_owner_lock irq_context: 0 rds_trans_sem console_lock console_srcu console_owner irq_context: 0 rds_trans_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 rds_trans_sem console_lock console_srcu console_owner console_owner_lock irq_context: 0 &id_priv->lock irq_context: 0 lock#7 irq_context: 0 lock#7 fs_reclaim irq_context: 0 lock#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 lock#7 pool_lock#2 irq_context: 0 lock#7 &xa->xa_lock#14 irq_context: 0 lock#7 &xa->xa_lock#14 pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem wq_pool_mutex irq_context: 0 pernet_ops_rwsem wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem pcpu_lock irq_context: 0 pernet_ops_rwsem &list->lock#4 irq_context: 0 pernet_ops_rwsem &dir->lock#2 irq_context: 0 pernet_ops_rwsem ptype_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-clock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex irq_context: 0 smc_wr_rx_hash_lock irq_context: 0 v9fs_trans_lock irq_context: 0 pernet_ops_rwsem &this->receive_lock irq_context: 0 &x->wait#17 &p->pi_lock irq_context: 0 &x->wait#17 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#17 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lowpan_nhc_lock irq_context: 0 ovs_mutex irq_context: 0 pernet_ops_rwsem once_lock irq_context: 0 pernet_ops_rwsem once_lock crngs.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex pool_lock#2 irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 kernfs_idr_lock &obj_hash[i].lock irq_context: 0 kernfs_idr_lock pool_lock#2 irq_context: 0 kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 kernfs_idr_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &root->kernfs_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &root->kernfs_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 uevent_sock_mutex &rq->__lock irq_context: 0 uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 uevent_sock_mutex pool_lock#2 irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock pool_lock#2 irq_context: 0 fs_reclaim &cfs_rq->removed.lock irq_context: 0 fs_reclaim &obj_hash[i].lock irq_context: 0 fs_reclaim pool_lock#2 irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 fs_reclaim &obj_hash[i].lock pool_lock irq_context: 0 fs_reclaim fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 fs_reclaim fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (debug_obj_work).work &____s->seqcount irq_context: 0 (wq_completion)events (debug_obj_work).work pool_lock#2 irq_context: 0 (wq_completion)events (debug_obj_work).work &rq->__lock irq_context: 0 &root->kernfs_rwsem &meta->lock irq_context: 0 &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 misc_mtx &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 key_types_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem fs_reclaim irq_context: 0 key_types_sem asymmetric_key_parsers_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 key_types_sem asymmetric_key_parsers_sem &c->lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &____s->seqcount irq_context: 0 key_types_sem asymmetric_key_parsers_sem pool_lock#2 irq_context: 0 key_types_sem asymmetric_key_parsers_sem crypto_alg_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem &obj_hash[i].lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem crypto_alg_sem irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem fs_reclaim irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem pool_lock#2 irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem kthread_create_lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &p->pi_lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &x->wait irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &rq->__lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &obj_hash[i].lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (crypto_chain).rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &x->wait#20 irq_context: 0 key_types_sem asymmetric_key_parsers_sem &base->lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &base->lock &obj_hash[i].lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &rq->__lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &x->wait#20 &p->pi_lock &rq->__lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem (&timer.timer) irq_context: 0 key_types_sem &type->lock_class irq_context: 0 key_types_sem &type->lock_class fs_reclaim irq_context: 0 key_types_sem &type->lock_class fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 key_types_sem &type->lock_class pool_lock#2 irq_context: 0 key_types_sem &type->lock_class key_user_lock irq_context: 0 key_types_sem &type->lock_class crngs.lock irq_context: 0 key_types_sem &type->lock_class key_serial_lock irq_context: 0 key_types_sem &type->lock_class key_construction_mutex irq_context: 0 key_types_sem &type->lock_class key_construction_mutex &obj_hash[i].lock irq_context: 0 key_types_sem &type->lock_class key_construction_mutex pool_lock#2 irq_context: 0 key_types_sem &type->lock_class ima_keys_lock irq_context: 0 key_types_sem &obj_hash[i].lock irq_context: 0 key_types_sem pool_lock#2 irq_context: 0 slab_mutex lock irq_context: 0 slab_mutex lock kernfs_idr_lock irq_context: 0 slab_mutex &root->kernfs_rwsem irq_context: 0 slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 slab_mutex &k->list_lock irq_context: 0 slab_mutex lock kernfs_idr_lock &c->lock irq_context: 0 slab_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 slab_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 slab_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 slab_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 slab_mutex &obj_hash[i].lock irq_context: 0 slab_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 slab_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 slab_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex crypto_alg_sem irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock fs_reclaim irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock pool_lock#2 irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock free_vmap_area_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock vmap_area_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock init_mm.page_table_lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock &pcp->lock &zone->lock irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cpu_hotplug_lock cpuhp_state_mutex scomp_lock &c->lock irq_context: 0 &mm->page_table_lock irq_context: 0 ptlock_ptr(ptdesc)#2 irq_context: 0 rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex crngs.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &dir->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex kthread_create_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events netstamp_work irq_context: 0 (wq_completion)events netstamp_work cpu_hotplug_lock irq_context: 0 (wq_completion)events netstamp_work cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)events netstamp_work cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 (wq_completion)events netstamp_work cpu_hotplug_lock jump_label_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait#21 irq_context: 0 &x->wait#21 irq_context: 0 &x->wait#21 &p->pi_lock irq_context: 0 &x->wait#21 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#21 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &local->services_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC pool_lock#2 irq_context: 0 pernet_ops_rwsem &rxnet->conn_lock irq_context: 0 pernet_ops_rwsem &call->waitq irq_context: 0 pernet_ops_rwsem &rx->call_lock irq_context: 0 pernet_ops_rwsem &rxnet->call_lock irq_context: 0 bio_slab_lock slab_mutex &root->kernfs_rwsem irq_context: 0 bio_slab_lock slab_mutex &k->list_lock irq_context: 0 bio_slab_lock slab_mutex lock irq_context: 0 bio_slab_lock slab_mutex lock kernfs_idr_lock irq_context: 0 bio_slab_lock slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 bio_slab_lock slab_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 init_user_ns.keyring_sem irq_context: 0 init_user_ns.keyring_sem key_user_lock irq_context: 0 init_user_ns.keyring_sem root_key_user.lock irq_context: 0 init_user_ns.keyring_sem fs_reclaim irq_context: 0 init_user_ns.keyring_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 init_user_ns.keyring_sem pool_lock#2 irq_context: 0 init_user_ns.keyring_sem crngs.lock irq_context: 0 init_user_ns.keyring_sem key_serial_lock irq_context: 0 init_user_ns.keyring_sem key_construction_mutex irq_context: 0 init_user_ns.keyring_sem &type->lock_class irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock fs_reclaim irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock &c->lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock &pcp->lock &zone->lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock &____s->seqcount irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock pool_lock#2 irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock root_key_user.lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex keyring_name_lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex &obj_hash[i].lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex pool_lock#2 irq_context: 0 init_user_ns.keyring_sem keyring_serialise_link_lock irq_context: 0 init_user_ns.keyring_sem key_construction_mutex keyring_name_lock irq_context: 0 init_user_ns.keyring_sem &type->lock_class keyring_serialise_link_lock &obj_hash[i].lock irq_context: 0 template_list irq_context: 0 idr_lock irq_context: softirq (&rxnet->peer_keepalive_timer) irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 ima_extend_list_mutex irq_context: 0 ima_extend_list_mutex fs_reclaim irq_context: 0 ima_extend_list_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ima_extend_list_mutex pool_lock#2 irq_context: 0 (wq_completion)krxrpcd irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &rxnet->peer_hash_lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &obj_hash[i].lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &base->lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) &base->lock &obj_hash[i].lock irq_context: 0 pci_bus_sem irq_context: 0 clk_debug_lock pin_fs_lock irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 clk_debug_lock &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)events_unbound deferred_probe_work irq_context: 0 (wq_completion)events_unbound deferred_probe_work deferred_probe_mutex irq_context: 0 deferred_probe_work irq_context: 0 misc_mtx &pcp->lock &zone->lock irq_context: 0 misc_mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 console_mutex &root->kernfs_rwsem irq_context: 0 console_mutex kernfs_notify_lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 console_mutex kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 console_mutex (console_sem).lock irq_context: 0 console_mutex console_lock console_srcu console_owner_lock irq_context: 0 console_mutex console_lock console_srcu console_owner irq_context: 0 console_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 console_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 k-sk_lock-AF_INET irq_context: 0 k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 k-slock-AF_INET#2 irq_context: 0 k-sk_lock-AF_INET6 irq_context: 0 k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 k-slock-AF_INET6 irq_context: 0 k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 power_off_handler_list.lock irq_context: 0 reg_requests_lock irq_context: 0 (wq_completion)events reg_work irq_context: 0 (wq_completion)events reg_work rtnl_mutex irq_context: 0 (wq_completion)events reg_work rtnl_mutex reg_requests_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events reg_work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events reg_work rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events reg_work rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events reg_work rtnl_mutex reg_pending_beacons_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rdev->wiphy.mtx irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rdev->wiphy.mtx reg_requests_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &base->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &fw_cache.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &fw_cache.lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) async_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) init_task.alloc_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) init_task.alloc_lock init_fs.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) init_task.alloc_lock init_fs.lock &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rcu_read_lock &____s->seqcount#6 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &sb->s_type->i_mutex_key &dentry->d_lock &wq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &dentry->d_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &c->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) (console_sem).lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 detector_work irq_context: 0 &wq->mutex &x->wait#10 irq_context: 0 rcu_read_lock &pool->lock (worker)->lock irq_context: 0 rcu_read_lock &pool->lock (worker)->lock &p->pi_lock irq_context: 0 rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 acpi_gpio_deferred_req_irqs_lock irq_context: 0 console_owner_lock irq_context: 0 console_owner irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem usermodehelper_disabled_waitq.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &x->wait#9 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &c->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem lock kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem bus_type_sem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem sysfs_symlink_target_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &dev->power.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dpm_list_mtx irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem subsys mutex#73 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem subsys mutex#73 &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fw_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem running_helpers_waitq.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &x->wait#22 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &base->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 gpd_list_lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &sb->s_type->i_lock_key#2 irq_context: 0 tomoyo_ss &c->lock irq_context: 0 tomoyo_ss &____s->seqcount irq_context: 0 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 tomoyo_ss &obj_hash[i].lock irq_context: 0 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 tomoyo_ss tomoyo_log_lock irq_context: 0 tomoyo_ss tomoyo_log_wait.lock irq_context: 0 cdev_lock irq_context: 0 tty_mutex (console_sem).lock irq_context: 0 tty_mutex console_lock irq_context: 0 tty_mutex fs_reclaim irq_context: 0 tty_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tty_mutex pool_lock#2 irq_context: 0 tty_mutex tty_ldiscs_lock irq_context: 0 tty_mutex &obj_hash[i].lock irq_context: 0 tty_mutex &k->list_lock irq_context: 0 tty_mutex &k->k_lock irq_context: 0 tty_mutex &tty->legacy_mutex irq_context: 0 tty_mutex &tty->legacy_mutex &tty->read_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->write_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem fs_reclaim irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem pool_lock#2 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem free_vmap_area_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem vmap_area_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &tty->write_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &tty->read_wait irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &tty->termios_rwsem irq_context: 0 &tty->legacy_mutex irq_context: 0 &tty->legacy_mutex &tty->files_lock irq_context: 0 &tty->legacy_mutex &port->lock irq_context: 0 &tty->legacy_mutex &port->mutex irq_context: 0 &tty->legacy_mutex &port->mutex fs_reclaim irq_context: 0 &tty->legacy_mutex &port->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &tty->legacy_mutex &port->mutex &____s->seqcount irq_context: 0 &tty->legacy_mutex &port->mutex &port_lock_key irq_context: 0 &tty->legacy_mutex &port->mutex pool_lock#2 irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class irq_controller_lock irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class mask_lock irq_context: 0 &tty->legacy_mutex &port->mutex &desc->request_mutex &irq_desc_lock_class mask_lock tmp_mask_lock irq_context: 0 &tty->legacy_mutex &port->mutex register_lock irq_context: 0 &tty->legacy_mutex &port->mutex register_lock proc_subdir_lock irq_context: 0 &tty->legacy_mutex &port->mutex register_lock fs_reclaim irq_context: 0 &tty->legacy_mutex &port->mutex register_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &tty->legacy_mutex &port->mutex register_lock pool_lock#2 irq_context: 0 &tty->legacy_mutex &port->mutex register_lock proc_inum_ida.xa_lock irq_context: 0 &tty->legacy_mutex &port->mutex register_lock proc_subdir_lock irq_context: 0 &tty->legacy_mutex &port->mutex register_lock &c->lock irq_context: 0 &tty->legacy_mutex &port->mutex register_lock &____s->seqcount irq_context: 0 &tty->legacy_mutex &port->mutex &irq_desc_lock_class irq_context: 0 &tty->legacy_mutex &port->mutex proc_subdir_lock irq_context: 0 &tty->legacy_mutex &port->mutex proc_inum_ida.xa_lock irq_context: 0 &tty->legacy_mutex &port->mutex proc_subdir_lock irq_context: 0 &tty->legacy_mutex &port_lock_key irq_context: 0 sb_writers#2 irq_context: 0 sb_writers#2 mount_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 rename_lock.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock &obj_hash[i].lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &dentry->d_lock pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &obj_hash[i].lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_log_wait.lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock &____s->seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &s->s_inode_list_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tk_core.seq.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_lock_key#2 &dentry->d_lock irq_context: 0 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key &dentry->d_lock &wq#2 irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key tk_core.seq.seqcount irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 &sb->s_type->i_mutex_key &dentry->d_lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#2 &sb->s_type->i_mutex_key/1 tomoyo_ss &____s->seqcount irq_context: 0 tomoyo_ss file_systems_lock irq_context: 0 tomoyo_ss fs_reclaim irq_context: 0 tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss rcu_read_lock init_fs.seq.seqcount irq_context: 0 &type->s_umount_key#24/1 irq_context: 0 &type->s_umount_key#24/1 fs_reclaim irq_context: 0 &type->s_umount_key#24/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#24/1 pool_lock#2 irq_context: 0 &type->s_umount_key#24/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#24/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#24/1 shrinker_mutex irq_context: 0 &type->s_umount_key#24/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#24/1 sb_lock irq_context: 0 &type->s_umount_key#24/1 inode_hash_lock irq_context: 0 &type->s_umount_key#24/1 inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#24/1 bdev_lock irq_context: 0 &type->s_umount_key#24/1 &disk->open_mutex irq_context: 0 &type->s_umount_key#24/1 &disk->open_mutex bdev_lock irq_context: 0 &type->s_umount_key#24/1 &disk->open_mutex bdev_lock &bdev->bd_holder_lock irq_context: 0 &type->s_umount_key#24/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#24/1 &c->lock irq_context: 0 &type->s_umount_key#24/1 &____s->seqcount irq_context: 0 &type->s_umount_key#24/1 &wq->mutex irq_context: 0 &type->s_umount_key#24/1 &wq->mutex &pool->lock irq_context: 0 &type->s_umount_key#24/1 kthread_create_lock irq_context: 0 &type->s_umount_key#24/1 &p->pi_lock irq_context: 0 &type->s_umount_key#24/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#24/1 &x->wait irq_context: 0 &type->s_umount_key#24/1 &rq->__lock irq_context: 0 &type->s_umount_key#24/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#24/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#24/1 &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#24/1 wq_pool_mutex irq_context: 0 &type->s_umount_key#24/1 wq_pool_mutex &wq->mutex irq_context: 0 &type->s_umount_key#24/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#24/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#24/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#24/1 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &type->s_umount_key#24/1 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#24/1 &xa->xa_lock#8 pool_lock#2 irq_context: 0 &type->s_umount_key#24/1 lock#4 irq_context: 0 &type->s_umount_key#24/1 &mapping->i_private_lock irq_context: 0 &type->s_umount_key#24/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#24/1 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#24/1 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#24/1 bit_wait_table + i irq_context: hardirq bit_wait_table + i irq_context: hardirq bit_wait_table + i &p->pi_lock irq_context: hardirq bit_wait_table + i &p->pi_lock &rq->__lock irq_context: hardirq bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) irq_context: 0 &type->s_umount_key#24/1 &wq->mutex &x->wait#10 irq_context: 0 &type->s_umount_key#24/1 wq_mayday_lock irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock irq_context: 0 &type->s_umount_key#24/1 wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 &type->s_umount_key#24/1 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#24/1 rcu_read_lock &pool->lock (worker)->lock irq_context: 0 &type->s_umount_key#24/1 rcu_read_lock &pool->lock (worker)->lock &p->pi_lock irq_context: 0 &type->s_umount_key#24/1 rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#24/1 &sbi->old_work_lock irq_context: 0 &type->s_umount_key#24/1 (work_completion)(&(&sbi->old_work)->work) irq_context: 0 &type->s_umount_key#24/1 &x->wait#23 irq_context: 0 &disk->open_mutex rcu_node_0 irq_context: 0 &disk->open_mutex &rcu_state.expedited_wq irq_context: 0 &disk->open_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &disk->open_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &disk->open_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex bdev_lock irq_context: 0 &disk->open_mutex bdev_lock &bdev->bd_holder_lock irq_context: 0 &xa->xa_lock#15 irq_context: 0 sb_lock &obj_hash[i].lock irq_context: 0 sb_lock pool_lock#2 irq_context: 0 &type->s_umount_key#25/1 irq_context: 0 &type->s_umount_key#25/1 fs_reclaim irq_context: 0 &type->s_umount_key#25/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#25/1 pool_lock#2 irq_context: 0 &type->s_umount_key#25/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#25/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#25/1 shrinker_mutex irq_context: 0 &type->s_umount_key#25/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#25/1 sb_lock irq_context: 0 &type->s_umount_key#25/1 inode_hash_lock irq_context: 0 &type->s_umount_key#25/1 inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#25/1 bdev_lock irq_context: 0 &type->s_umount_key#25/1 &disk->open_mutex irq_context: 0 &type->s_umount_key#25/1 &disk->open_mutex bdev_lock irq_context: 0 &type->s_umount_key#25/1 &disk->open_mutex bdev_lock &bdev->bd_holder_lock irq_context: 0 &type->s_umount_key#25/1 &c->lock irq_context: 0 &type->s_umount_key#25/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#25/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#25/1 &____s->seqcount irq_context: 0 &type->s_umount_key#25/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#25/1 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#25/1 lock#4 irq_context: 0 &type->s_umount_key#25/1 &mapping->i_private_lock irq_context: 0 &type->s_umount_key#25/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#25/1 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#25/1 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#25/1 bit_wait_table + i irq_context: 0 &type->s_umount_key#25/1 &rq->__lock irq_context: 0 &type->s_umount_key#25/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#25/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &type->s_umount_key#26/1 &journal->j_state_lock irq_context: 0 &type->s_umount_key#26/1 &p->alloc_lock irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 &type->s_umount_key#26/1 cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 &type->s_umount_key#26/1 &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#26/1 wq_pool_mutex irq_context: 0 &type->s_umount_key#26/1 wq_pool_mutex &wq->mutex irq_context: 0 &type->s_umount_key#26/1 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &type->s_umount_key#26/1 &ei->i_es_lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex fs_reclaim irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex &k->list_lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex lock kernfs_idr_lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex &c->lock irq_context: 0 &type->s_umount_key#26/1 ext4_grpinfo_slab_create_mutex slab_mutex &____s->seqcount irq_context: 0 &type->s_umount_key#26/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#26/1 ext4_li_mtx irq_context: 0 &type->s_umount_key#26/1 lock irq_context: 0 &type->s_umount_key#26/1 lock kernfs_idr_lock irq_context: 0 &type->s_umount_key#26/1 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#26/1 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &type->s_umount_key#26/1 lock kernfs_idr_lock pool_lock#2 irq_context: 0 &type->s_umount_key#26/1 (console_sem).lock irq_context: 0 &type->s_umount_key#26/1 console_lock console_srcu console_owner_lock irq_context: 0 &type->s_umount_key#26/1 console_lock console_srcu console_owner irq_context: 0 &type->s_umount_key#26/1 console_lock console_srcu console_owner &port_lock_key irq_context: 0 &type->s_umount_key#26/1 console_lock console_srcu console_owner console_owner_lock irq_context: 0 &type->s_umount_key#26/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#22 irq_context: 0 &type->i_mutex_dir_key#3 irq_context: 0 &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 &xa->xa_lock#8 irq_context: 0 &type->i_mutex_dir_key#3 &xa->xa_lock#8 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 lock#4 irq_context: 0 &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &rsp->gp_wait irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) pcpu_lock irq_context: 0 &type->i_mutex_dir_key#3 inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 &type->i_mutex_dir_key#3 inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 irq_context: 0 &type->i_mutex_dir_key#3 &ei->xattr_sem irq_context: 0 &type->i_mutex_dir_key#3 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq irq_context: 0 tomoyo_ss tomoyo_policy_lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rename_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem &c->lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 namespace_sem mount_lock mount_lock.seqcount pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key namespace_sem &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key namespace_sem &obj_hash[i].lock irq_context: 0 rcu_state.barrier_mutex irq_context: 0 rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: hardirq rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 rcu_state.barrier_mutex &x->wait#24 irq_context: 0 rcu_state.barrier_mutex &rq->__lock irq_context: 0 rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback &x->wait#24 irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#3 &c->lock irq_context: 0 &type->i_mutex_dir_key#3 lock#4 &lruvec->lru_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &retval->lock irq_context: 0 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq#2 irq_context: 0 &mm->mmap_lock fs_reclaim irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock irq_context: 0 &mm->mmap_lock fs_reclaim irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->page_table_lock irq_context: 0 &mm->mmap_lock pool_lock#2 irq_context: 0 &mm->mmap_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &mm->page_table_lock irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &sig->cred_guard_mutex irq_context: 0 &sig->cred_guard_mutex fs_reclaim irq_context: 0 &sig->cred_guard_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &c->lock irq_context: 0 &sig->cred_guard_mutex &____s->seqcount irq_context: 0 &sig->cred_guard_mutex init_fs.lock irq_context: 0 &sig->cred_guard_mutex &p->pi_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#8 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock lock#4 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#8 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &nvmeq->sq_lock irq_context: 0 &sig->cred_guard_mutex &folio_wait_table[i] irq_context: 0 &sig->cred_guard_mutex &rq->__lock irq_context: 0 &sig->cred_guard_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex tomoyo_ss pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &____s->seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock init_fs.seq.seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_log_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_log_wait.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss quarantine_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex binfmt_lock irq_context: 0 &sig->cred_guard_mutex init_binfmt_misc.entries_lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock init_fs.seq.seqcount irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 inode_hash_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 inode_hash_lock &s->s_inode_list_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 lock#4 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->xattr_sem irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->xattr_sem lock#4 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &xa->xa_lock#8 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &xa->xa_lock#8 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &c->lock irq_context: 0 &sig->cred_guard_mutex &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 &dentry->d_lock &wq#2 irq_context: 0 &sig->cred_guard_mutex &dentry->d_lock irq_context: 0 &sig->cred_guard_mutex &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &retval->lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &p->alloc_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &p->alloc_lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &sighand->siglock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &newf->file_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock batched_entropy_u64.lock irq_context: 0 batched_entropy_u16.lock irq_context: 0 batched_entropy_u16.lock crngs.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &____s->seqcount irq_context: 0 &mm->mmap_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 rcu_read_lock ptlock_ptr(ptdesc)#2/1 irq_context: 0 &mm->mmap_lock lock#4 irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock lock#5 irq_context: 0 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &mm->page_table_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem pool_lock#2 irq_context: 0 &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 mapping.invalidate_lock irq_context: 0 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 mapping.invalidate_lock &____s->seqcount irq_context: 0 mapping.invalidate_lock &xa->xa_lock#8 irq_context: 0 mapping.invalidate_lock lock#4 irq_context: 0 mapping.invalidate_lock lock#4 &lruvec->lru_lock irq_context: 0 mapping.invalidate_lock &ei->i_es_lock irq_context: 0 mapping.invalidate_lock pool_lock#2 irq_context: 0 mapping.invalidate_lock &pcp->lock &zone->lock irq_context: 0 mapping.invalidate_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 mapping.invalidate_lock &c->lock irq_context: 0 mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock rcu_read_lock pool_lock#2 irq_context: 0 mapping.invalidate_lock rcu_read_lock &retval->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock rcu_read_lock &nvmeq->sq_lock irq_context: 0 mapping.invalidate_lock &folio_wait_table[i] irq_context: 0 mapping.invalidate_lock &rq->__lock irq_context: 0 mapping.invalidate_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 binfmt_lock irq_context: 0 &fsnotify_mark_srcu irq_context: 0 &xa->xa_lock#8 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 mapping.invalidate_lock rcu_read_lock &c->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &vma->vm_lock->lock fs_reclaim irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &vma->vm_lock->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 &mm->mmap_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock &wq irq_context: 0 rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock &wq#2 irq_context: 0 mapping.invalidate_lock &xa->xa_lock#8 pool_lock#2 irq_context: 0 mapping.invalidate_lock &ei->i_data_sem irq_context: 0 mapping.invalidate_lock &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 mapping.invalidate_lock &ei->i_data_sem pool_lock#2 irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &obj_hash[i].lock pool_lock irq_context: 0 mapping.invalidate_lock &xa->xa_lock#8 &c->lock irq_context: 0 mapping.invalidate_lock &xa->xa_lock#8 &____s->seqcount irq_context: softirq &(&ipvs->defense_work)->timer irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&gc_work->dwork)->timer irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &s->s_inode_list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &ipvs->dropentry_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &ipvs->droppacket_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &ipvs->securetcp_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &base->lock &obj_hash[i].lock irq_context: softirq drivers/char/random.c:251 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 mapping.invalidate_lock batched_entropy_u8.lock irq_context: 0 mapping.invalidate_lock batched_entropy_u8.lock crngs.lock irq_context: 0 mapping.invalidate_lock batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 mapping.invalidate_lock kfence_freelist_lock irq_context: 0 &vma->vm_lock->lock mapping.invalidate_lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &rq->__lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock pool_lock#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: hardirq &rq->__lock pool_lock#2 irq_context: 0 tomoyo_ss batched_entropy_u8.lock irq_context: 0 tomoyo_ss kfence_freelist_lock irq_context: 0 tomoyo_ss &meta->lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &c->lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &____s->seqcount irq_context: 0 mapping.invalidate_lock &ei->i_es_lock key#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 &type->i_mutex_dir_key#3 &obj_hash[i].lock pool_lock irq_context: 0 tomoyo_ss quarantine_lock irq_context: 0 &port->mutex irq_context: 0 &tty->ldisc_sem irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &mm->mmap_lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &port->mutex irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &tty->ldisc_sem &tty->write_wait irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &tty->ldisc_sem &tty->read_wait irq_context: 0 task_group_lock irq_context: 0 &sighand->siglock &p->pi_lock irq_context: 0 &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 lock#4 irq_context: 0 &type->s_umount_key#27/1 irq_context: 0 &type->s_umount_key#27/1 fs_reclaim irq_context: 0 &type->s_umount_key#27/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#27/1 pool_lock#2 irq_context: 0 &type->s_umount_key#27/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#27/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#27/1 shrinker_mutex irq_context: 0 &type->s_umount_key#27/1 &c->lock irq_context: 0 &type->s_umount_key#27/1 &____s->seqcount irq_context: 0 &type->s_umount_key#27/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#27/1 sb_lock irq_context: 0 &type->s_umount_key#27/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#27/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_lock_key#23 irq_context: 0 &type->s_umount_key#27/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#27/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 &type->s_umount_key#27/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_lock_key#23 irq_context: 0 &sb->s_type->i_mutex_key#9 irq_context: 0 &sb->s_type->i_mutex_key#9 rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 proc_subdir_lock irq_context: 0 &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 irq_context: 0 tomoyo_ss rcu_read_lock rename_lock.seqcount irq_context: 0 &p->lock irq_context: 0 &p->lock fs_reclaim irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock pool_lock#2 irq_context: 0 &p->lock &mm->mmap_lock irq_context: 0 &type->s_umount_key#28/1 irq_context: 0 &type->s_umount_key#28/1 fs_reclaim irq_context: 0 &type->s_umount_key#28/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#28/1 pool_lock#2 irq_context: 0 &type->s_umount_key#28/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#28/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#28/1 shrinker_mutex irq_context: 0 &type->s_umount_key#28/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#28/1 sb_lock irq_context: 0 &type->s_umount_key#28/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem &c->lock irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem &____s->seqcount irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#24 irq_context: 0 &type->s_umount_key#28/1 &sb->s_type->i_lock_key#24 irq_context: 0 &type->s_umount_key#28/1 &sb->s_type->i_lock_key#24 &dentry->d_lock irq_context: 0 &type->s_umount_key#28/1 crngs.lock irq_context: 0 &type->s_umount_key#28/1 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#28/1 &dentry->d_lock irq_context: 0 &root->kernfs_iattr_rwsem irq_context: 0 &type->i_mutex_dir_key#4 irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 iattr_mutex irq_context: 0 &type->i_mutex_dir_key#4 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq irq_context: 0 &ent->pde_unload_lock irq_context: 0 &p->lock &c->lock irq_context: 0 &p->lock &pcp->lock &zone->lock irq_context: 0 &p->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &p->lock &____s->seqcount irq_context: 0 &p->lock file_systems_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 namespace_sem mount_lock mount_lock.seqcount pool_lock#2 irq_context: 0 &type->s_umount_key#29 irq_context: 0 &type->s_umount_key#29 &x->wait#23 irq_context: 0 &type->s_umount_key#29 shrinker_mutex irq_context: 0 &type->s_umount_key#29 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#29 pool_lock#2 irq_context: 0 &type->s_umount_key#29 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#29 rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 &type->s_umount_key#29 rename_lock.seqcount irq_context: 0 &type->s_umount_key#29 &dentry->d_lock irq_context: 0 &type->s_umount_key#29 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#29 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock irq_context: 0 &type->s_umount_key#29 &sb->s_type->i_lock_key#23 irq_context: 0 &type->s_umount_key#29 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#29 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#29 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#29 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#29 &dentry->d_lock pool_lock#2 irq_context: 0 &type->s_umount_key#29 &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#29 &dentry->d_lock/1 irq_context: 0 unnamed_dev_ida.xa_lock irq_context: 0 krc.lock irq_context: 0 &x->wait#25 irq_context: 0 &net->unx.table.locks[i] irq_context: 0 &sb->s_type->i_lock_key#8 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#10 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->unx.table.locks[i] irq_context: 0 &sb->s_type->i_mutex_key#10 &u->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &u->lock clock-AF_UNIX irq_context: 0 &sb->s_type->i_mutex_key#10 &u->peer_wait irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_UNIX irq_context: 0 &sb->s_type->i_mutex_key#10 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &dir->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &obj_hash[i].lock irq_context: 0 &dentry->d_lock/1 irq_context: 0 &sig->cred_guard_mutex &fs->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &____s->seqcount#3 irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &____s->seqcount#3 irq_context: 0 &sig->cred_guard_mutex lock#4 irq_context: 0 &sig->cred_guard_mutex &sb->s_type->i_mutex_key#8 irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->alloc_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->alloc_lock cpu_asid_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->wait_chldexit irq_context: 0 tasklist_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock irq_context: 0 &mm->mmap_lock &p->alloc_lock irq_context: 0 &mm->mmap_lock lock#4 irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock lock#5 irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &____s->seqcount irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 &mm->mmap_lock &lruvec->lru_lock irq_context: 0 tasklist_lock &sighand->siglock &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &c->lock irq_context: 0 tasklist_lock &sighand->siglock pool_lock#2 irq_context: 0 rcu_read_lock &____s->seqcount#4 irq_context: 0 &prev->lock irq_context: 0 &(&sig->stats_lock)->lock irq_context: 0 &(&sig->stats_lock)->lock &____s->seqcount#4 irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_data_sem &____s->seqcount irq_context: hardirq bit_wait_table + i &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 irq_context: 0 sb_writers#3 mount_lock irq_context: 0 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 mount_lock irq_context: 0 sb_writers#4 tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_lock_key#23 irq_context: 0 sb_writers#4 &wb->list_lock irq_context: 0 sb_writers#4 &wb->list_lock &sb->s_type->i_lock_key#23 irq_context: 0 &sb->s_type->i_mutex_key#9 &p->alloc_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &pid->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 &p->alloc_lock &fs->lock &dentry->d_lock irq_context: 0 &p->lock namespace_sem irq_context: 0 &p->lock namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &p->lock namespace_sem rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &type->s_umount_key#30 irq_context: 0 &type->s_umount_key#30 &lru->node[i].lock irq_context: 0 &type->s_umount_key#30 &dentry->d_lock irq_context: 0 &type->s_umount_key#30 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#30 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &dentry->d_lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 &sb->s_type->i_lock_key#22 irq_context: 0 &type->s_umount_key#30 &sb->s_type->i_lock_key#22 &lru->node[i].lock irq_context: 0 &type->s_umount_key#30 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem &rsp->gp_wait irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem &x->wait#2 irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback put_task_map-wait-type-override task_group_lock irq_context: softirq rcu_callback &x->wait#2 irq_context: softirq rcu_callback &x->wait#2 &p->pi_lock irq_context: softirq rcu_callback &x->wait#2 &p->pi_lock &rq->__lock irq_context: softirq rcu_callback &x->wait#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &sem->waiters irq_context: 0 &type->s_umount_key#30 &rsp->gp_wait irq_context: 0 &type->s_umount_key#30 &rsp->gp_wait &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &rsp->gp_wait pool_lock#2 irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock irq_context: 0 &type->s_umount_key#30 &p->alloc_lock irq_context: 0 &type->s_umount_key#30 (work_completion)(&sbi->s_sb_upd_work) irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock irq_context: 0 &type->s_umount_key#30 key#3 irq_context: 0 &type->s_umount_key#30 key#4 irq_context: 0 &type->s_umount_key#30 &sbi->s_error_lock irq_context: 0 &type->s_umount_key#30 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#30 pool_lock#2 irq_context: 0 &type->s_umount_key#30 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &base->lock irq_context: 0 &type->s_umount_key#30 &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#30 bit_wait_table + i irq_context: 0 &type->s_umount_key#30 &rq->__lock irq_context: 0 &type->s_umount_key#30 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 ext4_li_mtx irq_context: 0 &type->s_umount_key#30 ext4_li_mtx fs_reclaim irq_context: 0 &type->s_umount_key#30 ext4_li_mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#30 ext4_li_mtx pool_lock#2 irq_context: 0 &type->s_umount_key#30 ext4_li_mtx batched_entropy_u16.lock irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &eli->li_list_mtx irq_context: 0 &type->s_umount_key#30 ext4_li_mtx kthread_create_lock irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &p->pi_lock irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &x->wait irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &rq->__lock irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 ext4_li_mtx &obj_hash[i].lock irq_context: 0 &eli->li_list_mtx irq_context: 0 &type->s_umount_key#30 (console_sem).lock irq_context: 0 &type->s_umount_key#30 console_lock console_srcu console_owner_lock irq_context: 0 &type->s_umount_key#30 console_lock console_srcu console_owner irq_context: 0 &type->s_umount_key#30 console_lock console_srcu console_owner &port_lock_key irq_context: 0 &type->s_umount_key#30 console_lock console_srcu console_owner console_owner_lock irq_context: 0 &type->s_umount_key#30 mount_lock irq_context: 0 &type->s_umount_key#30 mount_lock mount_lock.seqcount irq_context: 0 &type->s_umount_key#30 mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 rcu_read_lock &pid->lock irq_context: 0 rcu_read_lock &sb->s_type->i_lock_key#23 irq_context: 0 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 rename_lock.seqcount irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock irq_context: 0 &pid->lock irq_context: 0 sb_writers#3 tk_core.seq.seqcount irq_context: 0 sb_writers#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &____s->seqcount irq_context: 0 sb_writers#3 &c->lock irq_context: 0 sb_writers#3 pool_lock#2 irq_context: 0 sb_writers#3 &journal->j_state_lock irq_context: 0 sb_writers#3 &journal->j_state_lock irq_context: 0 sb_writers#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 jbd2_handle irq_context: 0 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 jbd2_handle &c->lock irq_context: 0 sb_writers#3 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_writers#3 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &wb->list_lock irq_context: 0 sb_writers#3 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &wb->work_lock irq_context: 0 sb_writers#3 &wb->work_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &wb->work_lock &base->lock irq_context: 0 sb_writers#3 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &xa->xa_lock#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 lock#4 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock &nvmeq->sq_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &meta_group_info[i]->alloc_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle inode_hash_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle batched_entropy_u32.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_es_lock key#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle smack_known_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle smack_known_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle smack_known_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_internal rcu_read_lock init_fs.seq.seqcount irq_context: 0 sb_internal rcu_read_lock rcu_read_lock mount_lock.seqcount irq_context: 0 sb_internal rcu_read_lock rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_internal mmu_notifier_invalidate_range_start irq_context: 0 sb_internal pool_lock#2 irq_context: 0 sb_internal &journal->j_state_lock irq_context: 0 sb_internal jbd2_handle irq_context: 0 sb_internal jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_internal jbd2_handle pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ret->b_state_lock irq_context: 0 sb_internal jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_internal jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_internal jbd2_handle &mapping->i_private_lock irq_context: 0 sb_internal jbd2_handle &journal->j_wait_updates irq_context: 0 sb_internal &obj_hash[i].lock irq_context: 0 &ei->i_data_sem irq_context: 0 &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &sighand->siglock hrtimer_bases.lock irq_context: 0 &sighand->siglock hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 &sighand->siglock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 file_rwsem irq_context: 0 file_rwsem &ctx->flc_lock irq_context: 0 file_rwsem &ctx->flc_lock &fll->lock irq_context: 0 &ctx->flc_lock irq_context: softirq rcu_callback &cfs_rq->removed.lock irq_context: softirq rcu_callback &rq->__lock irq_context: softirq rcu_callback &rsp->gp_wait irq_context: hardirq &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 mount_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex sb_writers#3 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &ret->b_state_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_revoke_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &ei->i_raw_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_wait_updates irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &mapping->i_private_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock batched_entropy_u64.lock crngs.lock irq_context: 0 &mm->mmap_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 jbd2_handle &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &p->alloc_lock irq_context: 0 &type->s_umount_key#31/1 irq_context: 0 &type->s_umount_key#31/1 fs_reclaim irq_context: 0 &type->s_umount_key#31/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#31/1 pool_lock#2 irq_context: 0 &type->s_umount_key#31/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#31/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#31/1 shrinker_mutex irq_context: 0 &type->s_umount_key#31/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#31/1 sb_lock irq_context: 0 &type->s_umount_key#31/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#31/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_lock_key#25 irq_context: 0 &type->s_umount_key#31/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#31/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_lock_key#25 &dentry->d_lock irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 fs_reclaim irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 pool_lock#2 irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 &dentry->d_lock irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 &sb->s_type->i_lock_key#25 irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#31/1 &sb->s_type->i_mutex_key#11 &sb->s_type->i_lock_key#25 &dentry->d_lock irq_context: 0 &type->s_umount_key#31/1 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rename_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &type->s_umount_key/1 fs_reclaim irq_context: 0 &type->s_umount_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key/1 mmu_notifier_invalidate_range_start irq_context: 0 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 &type->s_umount_key#32 irq_context: 0 &type->s_umount_key#32 sb_lock irq_context: 0 &type->s_umount_key#32 fs_reclaim irq_context: 0 &type->s_umount_key#32 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#32 pool_lock#2 irq_context: 0 &type->s_umount_key#32 &dentry->d_lock irq_context: 0 &type->s_umount_key#32 &lru->node[i].lock irq_context: 0 &type->s_umount_key#32 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#32 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rename_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &type->s_umount_key#33 irq_context: 0 &type->s_umount_key#33 sb_lock irq_context: 0 &type->s_umount_key#33 &dentry->d_lock irq_context: 0 &type->s_umount_key#34 irq_context: 0 &type->s_umount_key#34 sb_lock irq_context: 0 &type->s_umount_key#34 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 sysctl_lock irq_context: 0 rcu_read_lock &dentry->d_lock sysctl_lock irq_context: 0 &type->s_umount_key#35/1 irq_context: 0 &type->s_umount_key#35/1 fs_reclaim irq_context: 0 &type->s_umount_key#35/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#35/1 pool_lock#2 irq_context: 0 &type->s_umount_key#35/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#35/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#35/1 shrinker_mutex irq_context: 0 &type->s_umount_key#35/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#35/1 sb_lock irq_context: 0 &type->s_umount_key#35/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#35/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#35/1 &sb->s_type->i_lock_key#26 irq_context: 0 &type->s_umount_key#35/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#35/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#35/1 &sb->s_type->i_lock_key#26 &dentry->d_lock irq_context: 0 &type->s_umount_key#35/1 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount &new_ns->poll irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 namespace_sem mount_lock mount_lock.seqcount &obj_hash[i].lock irq_context: 0 &type->s_umount_key#36 irq_context: 0 &type->s_umount_key#36 sb_lock irq_context: 0 &type->s_umount_key#36 &dentry->d_lock irq_context: 0 redirect_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock fs_reclaim irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock pool_lock#2 irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &mm->mmap_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &tty->write_wait irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &dev->power.lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &rq->__lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &port_lock_key irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &port_lock_key &port->lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &port_lock_key &tty->write_wait irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &port_lock_key &tty->write_wait &p->pi_lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock hrtimer_bases.lock irq_context: 0 (wq_completion)pm (work_completion)(&dev->power.work) &dev->power.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &port->lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &ldata->output_lock &port_lock_key &tty->write_wait irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &port_lock_key irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->termios_rwsem &port_lock_key &dev->power.lock irq_context: 0 &tty->ldisc_sem &tty->atomic_write_lock &tty->files_lock irq_context: 0 &tty->ldisc_sem &tty->write_wait irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &____s->seqcount irq_context: 0 &type->s_umount_key#37/1 irq_context: 0 &type->s_umount_key#37/1 fs_reclaim irq_context: 0 &type->s_umount_key#37/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#37/1 pool_lock#2 irq_context: 0 &type->s_umount_key#37/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#37/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#37/1 shrinker_mutex irq_context: 0 &type->s_umount_key#37/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#37/1 sb_lock irq_context: 0 &type->s_umount_key#37/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#37/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#37/1 &sb->s_type->i_lock_key#27 irq_context: 0 &type->s_umount_key#37/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#37/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#37/1 &sb->s_type->i_lock_key#27 &dentry->d_lock irq_context: 0 &type->s_umount_key#37/1 fuse_mutex irq_context: 0 &type->s_umount_key#37/1 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 &type->s_umount_key#38/1 irq_context: 0 &type->s_umount_key#38/1 fs_reclaim irq_context: 0 &type->s_umount_key#38/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#38/1 pool_lock#2 irq_context: 0 &type->s_umount_key#38/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#38/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#38/1 &____s->seqcount irq_context: 0 &type->s_umount_key#38/1 shrinker_mutex irq_context: 0 &type->s_umount_key#38/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#38/1 sb_lock irq_context: 0 &type->s_umount_key#38/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#38/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_lock_key#28 irq_context: 0 &type->s_umount_key#38/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#38/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_lock_key#28 &dentry->d_lock irq_context: 0 &type->s_umount_key#38/1 pstore_sb_lock irq_context: 0 &type->s_umount_key#38/1 pstore_sb_lock &sb->s_type->i_mutex_key#12 irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 fs_reclaim irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &zone->lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &____s->seqcount irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex (efivars_lock).lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex fs_reclaim irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex pool_lock#2 irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex (efi_runtime_lock).lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex &obj_hash[i].lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex &x->wait#12 irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex &rq->__lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &psinfo->read_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_node_0 irq_context: 0 &type->s_umount_key#38/1 &sb->s_type->i_mutex_key#12 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#38/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#39/1 irq_context: 0 &type->s_umount_key#39/1 fs_reclaim irq_context: 0 &type->s_umount_key#39/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#39/1 pool_lock#2 irq_context: 0 &type->s_umount_key#39/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#39/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#39/1 shrinker_mutex irq_context: 0 &type->s_umount_key#39/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#39/1 sb_lock irq_context: 0 &type->s_umount_key#39/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#39/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#39/1 &sb->s_type->i_lock_key#29 irq_context: 0 &type->s_umount_key#39/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#39/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#39/1 &sb->s_type->i_lock_key#29 &dentry->d_lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock (kmod_concurrent_max).lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock fs_reclaim irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock pool_lock#2 irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock &x->wait#17 irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock &rq->__lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uts_sem irq_context: 0 &type->i_mutex_dir_key#3 &xa->xa_lock#8 &c->lock irq_context: 0 &type->i_mutex_dir_key#3 &xa->xa_lock#8 &____s->seqcount irq_context: softirq (&cb->timer) irq_context: softirq (&cb->timer) &obj_hash[i].lock irq_context: softirq (&cb->timer) &base->lock irq_context: softirq (&cb->timer) &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock key irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock pcpu_lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock percpu_counters_lock irq_context: 0 &type->s_umount_key#39/1 bpf_preload_lock running_helpers_waitq.lock irq_context: 0 &type->s_umount_key#39/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#14 irq_context: 0 &type->s_umount_key#14 sb_lock irq_context: 0 &type->s_umount_key#14 fs_reclaim irq_context: 0 &type->s_umount_key#14 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#14 pool_lock#2 irq_context: 0 &type->s_umount_key#14 &dentry->d_lock irq_context: 0 &type->s_umount_key#14 &lru->node[i].lock irq_context: 0 &type->s_umount_key#14 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#14 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem &c->lock irq_context: 0 &type->i_mutex_dir_key#4 namespace_sem &____s->seqcount irq_context: 0 &sb->s_type->i_lock_key irq_context: 0 &type->i_mutex_dir_key#5 irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &wq irq_context: 0 sb_writers#5 irq_context: 0 sb_writers#5 mount_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sbinfo->stat_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &s->s_inode_list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 batched_entropy_u32.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &xattrs->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &xattrs->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &mm->mmap_lock sb_writers#3 mount_lock irq_context: 0 &mm->mmap_lock sb_writers#3 tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock sb_writers#3 mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock sb_writers#3 pool_lock#2 irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &ei->i_raw_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_wait_updates irq_context: 0 &mm->mmap_lock sb_writers#3 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sbinfo->stat_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &s->s_inode_list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 batched_entropy_u32.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &xattrs->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &xattrs->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers#5 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &wb->list_lock irq_context: 0 sb_writers#5 &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 &sig->cred_guard_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &dentry->d_lock &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &c->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &____s->seqcount irq_context: 0 uts_sem irq_context: 0 uts_sem hostname_poll.wait.lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &sig->cred_guard_mutex quarantine_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &____s->seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &c->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_node_0 irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rcu_read_lock rcu_read_lock mount_lock.seqcount irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &f->f_pos_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mount_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &ei->i_raw_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_wait_updates irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 &f->f_pos_lock &mm->mmap_lock irq_context: 0 &fs->lock &dentry->d_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mm->page_table_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock ptlock_ptr(ptdesc)#2 rcu_read_lock ptlock_ptr(ptdesc)#2/1 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 &rq->__lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#5 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &vma->vm_lock->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#5 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock key irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pcpu_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pool_lock#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#5 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &meta->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pool_lock irq_context: 0 &type->i_mutex_dir_key#5 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &xattrs->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &mm->mmap_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &____s->seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &xa->xa_lock#8 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &sb->s_type->i_lock_key irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &info->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 lock#4 irq_context: 0 &p->alloc_lock &x->wait#25 irq_context: 0 &p->alloc_lock &x->wait#25 &p->pi_lock irq_context: 0 &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock irq_context: 0 &p->alloc_lock &x->wait#25 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sighand->siglock &obj_hash[i].lock irq_context: 0 &sighand->siglock pool_lock#2 irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock &pcp->lock &zone->lock irq_context: 0 &vma->vm_lock->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &net->unx.table.locks[i] irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &net->unx.table.locks[i]/1 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &u->bindlock &bsd_socket_locks[i] irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &____s->seqcount irq_context: 0 &u->iolock irq_context: 0 &u->iolock rlock-AF_UNIX irq_context: 0 &ei->socket.wq.wait irq_context: 0 key#5 irq_context: 0 &p->alloc_lock &x->wait#25 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock quarantine_lock irq_context: 0 tasklist_lock &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_es_lock key#6 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 &bsd_socket_locks[i] irq_context: 0 sb_writers tk_core.seq.seqcount irq_context: 0 sb_writers &sb->s_type->i_lock_key#5 irq_context: 0 sb_writers &wb->list_lock irq_context: 0 sb_writers &wb->list_lock &sb->s_type->i_lock_key#5 irq_context: 0 &u->lock irq_context: 0 &u->lock &u->lock/1 irq_context: 0 &u->lock rlock-AF_UNIX irq_context: 0 rcu_read_lock &ei->socket.wq.wait irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock irq_context: 0 &u->iolock &obj_hash[i].lock irq_context: 0 &u->iolock pool_lock#2 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &wb->list_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 syslog_lock irq_context: 0 &u->lock &u->peer_wait irq_context: 0 &u->iolock &u->peer_wait irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock &rq->__lock irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &xa->xa_lock#8 pool_lock#2 irq_context: 0 &u->iolock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_lock_key#14 irq_context: 0 &sb->s_type->i_lock_key#14 &dentry->d_lock irq_context: 0 &pipe->mutex/1 irq_context: 0 &pipe->rd_wait irq_context: 0 &u->iolock quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &dentry->d_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &____s->seqcount irq_context: 0 cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rq->__lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex &rq->__lock cpu_asid_lock irq_context: 0 &u->iolock &meta->lock irq_context: 0 &u->iolock kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &sem->wait_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &sem->wait_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait irq_context: 0 &pipe->mutex/1 &pipe->wr_wait irq_context: 0 tomoyo_ss key irq_context: 0 tomoyo_ss pcpu_lock irq_context: 0 tomoyo_ss percpu_counters_lock irq_context: 0 &type->i_mutex_dir_key#5 pool_lock#2 irq_context: 0 &u->lock clock-AF_UNIX irq_context: 0 &u->peer_wait irq_context: 0 rlock-AF_UNIX irq_context: 0 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &rq->__lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 tomoyo_ss tomoyo_policy_lock &rq->__lock irq_context: 0 &u->iolock &u->peer_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &rq->__lock irq_context: 0 &pipe->mutex/1 &lock->wait_lock irq_context: 0 &pipe->mutex/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &lock->wait_lock irq_context: 0 &pipe->mutex/1 fs_reclaim irq_context: 0 &pipe->mutex/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pipe->mutex/1 &____s->seqcount irq_context: 0 &pipe->mutex/1 &mm->mmap_lock irq_context: 0 &pipe->rd_wait &p->pi_lock irq_context: 0 &pipe->rd_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#6 tk_core.seq.seqcount irq_context: 0 sb_writers#6 mount_lock irq_context: 0 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock lock#4 &obj_hash[i].lock irq_context: hardirq bit_wait_table + i &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rename_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock sysctl_lock irq_context: 0 sb_writers#4 sysctl_lock irq_context: 0 sb_writers#4 &dentry->d_lock irq_context: 0 sb_writers#4 tomoyo_ss irq_context: 0 sb_writers#4 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#4 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#4 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#4 tomoyo_ss &c->lock irq_context: 0 sb_writers#4 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 sb_writers#4 fs_reclaim irq_context: 0 sb_writers#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 pool_lock#2 irq_context: 0 sb_writers#4 &mm->mmap_lock irq_context: 0 sb_writers#4 &obj_hash[i].lock irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock kfence_freelist_lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem irq_context: softirq (&cb->timer) tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock &wq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock pool_lock#2 irq_context: 0 nl_table_lock pool_lock#2 irq_context: 0 sk_lock-AF_NETLINK irq_context: 0 sk_lock-AF_NETLINK slock-AF_NETLINK irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 slock-AF_NETLINK irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &net->unx.table.locks[i] irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &u->bindlock &bsd_socket_locks[i] irq_context: 0 &u->lock &sk->sk_peer_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle pool_lock#2 irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &ret->b_state_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_revoke_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &sb->s_type->i_lock_key#22 irq_context: 0 &mm->mmap_lock sb_writers#3 &wb->list_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &xa->xa_lock#8 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 lock#4 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq log_wait.lock &p->pi_lock irq_context: hardirq log_wait.lock &p->pi_lock &rq->__lock irq_context: hardirq log_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &xa->xa_lock#8 pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &mapping->i_private_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &ret->b_state_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_revoke_lock irq_context: 0 &type->i_mutex_dir_key#5 &c->lock irq_context: 0 &type->i_mutex_dir_key#5 &____s->seqcount irq_context: hardirq &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 mapping.invalidate_lock &ei->i_es_lock key#6 irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &xattrs->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 tomoyo_ss &n->list_lock irq_context: 0 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &p->alloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#4 oom_adj_mutex irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &p->alloc_lock irq_context: 0 sb_writers#4 oom_adj_mutex &p->alloc_lock irq_context: 0 &mm->mmap_lock &meta->lock irq_context: 0 sb_writers#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_lock_key#15 &dentry->d_lock irq_context: 0 &group->mark_mutex irq_context: 0 &group->mark_mutex &fsnotify_mark_srcu irq_context: 0 &group->mark_mutex fs_reclaim irq_context: 0 &group->mark_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &group->mark_mutex &____s->seqcount irq_context: 0 &group->mark_mutex &c->lock irq_context: 0 &group->mark_mutex pool_lock#2 irq_context: 0 &group->mark_mutex lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock pool_lock#2 irq_context: 0 &group->mark_mutex ucounts_lock irq_context: 0 &group->mark_mutex &mark->lock irq_context: 0 &group->mark_mutex &mark->lock &fsnotify_mark_srcu irq_context: 0 &group->mark_mutex &mark->lock &fsnotify_mark_srcu &conn->lock irq_context: 0 &group->mark_mutex &mark->lock &conn->lock irq_context: 0 &group->mark_mutex &conn->lock irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#22 irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key#22 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock pool_lock#2 irq_context: 0 &sk->sk_peer_lock irq_context: 0 &ep->mtx irq_context: 0 epnested_mutex irq_context: 0 epnested_mutex &ep->mtx irq_context: 0 epnested_mutex &ep->mtx fs_reclaim irq_context: 0 epnested_mutex &ep->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 epnested_mutex &ep->mtx &____s->seqcount irq_context: 0 epnested_mutex &ep->mtx &c->lock irq_context: 0 epnested_mutex &ep->mtx pool_lock#2 irq_context: 0 epnested_mutex &ep->mtx &f->f_lock irq_context: 0 epnested_mutex &ep->mtx &ei->socket.wq.wait irq_context: 0 epnested_mutex &ep->mtx &ep->lock irq_context: 0 epnested_mutex rcu_read_lock &f->f_lock irq_context: 0 &ep->mtx fs_reclaim irq_context: 0 &ep->mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ep->mtx &f->f_lock irq_context: 0 &ep->mtx pool_lock#2 irq_context: 0 &ep->mtx &group->notification_waitq irq_context: 0 &ep->mtx &group->notification_lock irq_context: 0 &ep->mtx &ep->lock irq_context: 0 &ep->mtx &sighand->signalfd_wqh irq_context: 0 &ep->mtx &sighand->siglock irq_context: 0 &ep->mtx &ei->socket.wq.wait irq_context: 0 &ep->lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 remove_cache_srcu irq_context: 0 sb_writers#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 vmap_area_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &dentry->d_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock fs_reclaim irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#4 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 tomoyo_ss tomoyo_policy_lock &n->list_lock irq_context: 0 tomoyo_ss tomoyo_policy_lock &n->list_lock &c->lock irq_context: 0 remove_cache_srcu &c->lock irq_context: 0 remove_cache_srcu &n->list_lock irq_context: 0 remove_cache_srcu &obj_hash[i].lock irq_context: 0 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: hardirq &dev->power.lock hrtimer_bases.lock irq_context: hardirq &dev->power.lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#7 mount_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#7 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#7 &sb->s_type->i_lock_key#24 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#7 &wb->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#7 &wb->list_lock &sb->s_type->i_lock_key#24 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 quarantine_lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_callback &zone->lock irq_context: softirq rcu_callback &zone->lock &____s->seqcount irq_context: softirq &(&krcp->monitor_work)->timer irq_context: softirq &(&tbl->managed_work)->timer irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) krc.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback pool_lock#2 irq_context: 0 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &type->i_mutex_dir_key#4 batched_entropy_u8.lock irq_context: 0 &type->i_mutex_dir_key#4 kfence_freelist_lock irq_context: softirq rcu_callback &pcp->lock &zone->lock irq_context: softirq rcu_callback &pcp->lock &zone->lock &____s->seqcount irq_context: 0 remove_cache_srcu &rq->__lock irq_context: 0 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &obj_hash[i].lock pool_lock irq_context: 0 &vma->vm_lock->lock ptlock_ptr(ptdesc)#2 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock irq_context: 0 swap_lock irq_context: 0 sb_writers#7 irq_context: 0 sb_writers#7 mount_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rename_lock.seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 fs_reclaim irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &dentry->d_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 iattr_mutex irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &obj_hash[i].lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &sb->s_type->i_lock_key#24 &dentry->d_lock &wq#2 irq_context: 0 kn->active fs_reclaim irq_context: 0 kn->active fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active pool_lock#2 irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] pool_lock#2 irq_context: 0 sb_writers#7 fs_reclaim irq_context: 0 sb_writers#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 pool_lock#2 irq_context: 0 sb_writers#7 &mm->mmap_lock irq_context: 0 sb_writers#7 &of->mutex irq_context: 0 sb_writers#7 &of->mutex kn->active &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active &obj_hash[i].lock irq_context: 0 sb_writers#7 &obj_hash[i].lock irq_context: 0 &ep->mtx &mm->mmap_lock irq_context: 0 sb_writers#7 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &____s->seqcount irq_context: 0 &kernfs_locks->open_file_mutex[count] irq_context: 0 &kernfs_locks->open_file_mutex[count] &obj_hash[i].lock irq_context: 0 &kernfs_locks->open_file_mutex[count] pool_lock#2 irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock &obj_hash[i].lock irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock &base->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] krc.lock &base->lock &obj_hash[i].lock irq_context: 0 kn->active#2 fs_reclaim irq_context: 0 kn->active#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#2 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex pool_lock#2 irq_context: 0 rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &nlk->wait irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &obj_hash[i].lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &____s->seqcount irq_context: 0 kn->active#2 &c->lock irq_context: 0 kn->active#2 &____s->seqcount irq_context: 0 &kernfs_locks->open_file_mutex[count] &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &____s->seqcount irq_context: 0 kn->active#2 &n->list_lock irq_context: 0 kn->active#2 &n->list_lock &c->lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &n->list_lock &c->lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 kn->active &c->lock irq_context: 0 kn->active &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &____s->seqcount irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &____s->seqcount irq_context: 0 kn->active &n->list_lock irq_context: 0 kn->active &n->list_lock &c->lock irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#2 quarantine_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &n->list_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &n->list_lock irq_context: 0 sb_writers#7 &n->list_lock &c->lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active remove_cache_srcu irq_context: 0 kn->active remove_cache_srcu quarantine_lock irq_context: 0 kn->active remove_cache_srcu &c->lock irq_context: 0 kn->active remove_cache_srcu &n->list_lock irq_context: 0 kn->active remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#2 &pcp->lock &zone->lock irq_context: 0 kn->active#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &n->list_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &rq->__lock cpu_asid_lock irq_context: 0 &ep->mtx &pcp->lock &zone->lock irq_context: 0 &ep->mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ep->mtx &____s->seqcount irq_context: 0 &ep->mtx &c->lock irq_context: 0 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 kn->active#2 remove_cache_srcu irq_context: 0 kn->active#2 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#2 remove_cache_srcu &c->lock irq_context: 0 kn->active#2 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#2 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#2 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#2 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 iattr_mutex &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 iattr_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &c->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 sb_writers#4 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#4 tomoyo_ss &n->list_lock &c->lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 remove_cache_srcu irq_context: 0 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#7 remove_cache_srcu quarantine_lock irq_context: 0 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) krc.lock &obj_hash[i].lock irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override &c->lock irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#2 remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 kn->active#2 remove_cache_srcu fill_pool_map-wait-type-override &c->lock irq_context: 0 kn->active#2 remove_cache_srcu fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 kn->active#2 remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 &n->list_lock &c->lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex quarantine_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) krc.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &obj_hash[i].lock pool_lock irq_context: 0 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 batched_entropy_u8.lock irq_context: 0 sb_writers#7 kfence_freelist_lock irq_context: 0 sb_writers#7 &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback quarantine_lock irq_context: 0 sb_writers#7 &of->mutex &rq->__lock irq_context: softirq (&vblank->disable_timer) irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock &dev->vblank_time_lock irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock irq_context: softirq (&vblank->disable_timer) &dev->vbl_lock &dev->vblank_time_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 batched_entropy_u8.lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 kfence_freelist_lock irq_context: 0 kn->active#2 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 kn->active#2 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#2 uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: softirq (&q->timeout) irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) irq_context: 0 sb_writers#7 &of->mutex kn->active#2 &rq->__lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu quarantine_lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &c->lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &n->list_lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#2 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active &pcp->lock &zone->lock irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &rq->__lock irq_context: 0 kn->active#2 &rq->__lock irq_context: 0 kn->active#3 fs_reclaim irq_context: 0 kn->active#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#3 pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &____s->seqcount irq_context: 0 kn->active#3 &c->lock irq_context: 0 kn->active#3 &pcp->lock &zone->lock irq_context: 0 kn->active#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#3 &____s->seqcount irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &n->list_lock &c->lock irq_context: 0 kn->active#3 &n->list_lock irq_context: 0 kn->active#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 kn->active#3 remove_cache_srcu irq_context: 0 kn->active#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#3 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 &root->kernfs_rwsem &rq->__lock irq_context: 0 kn->active#3 remove_cache_srcu &c->lock irq_context: 0 kn->active#3 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#3 uevent_sock_mutex &rq->__lock irq_context: 0 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 &mm->mmap_lock &lruvec->lru_lock irq_context: 0 &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &n->list_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &n->list_lock irq_context: 0 &mm->mmap_lock &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_lock_key#24 irq_context: 0 &type->i_mutex_dir_key#4 &rq->__lock irq_context: 0 sb_writers#7 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#7 &wb->list_lock irq_context: 0 sb_writers#7 &wb->list_lock &sb->s_type->i_lock_key#24 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#4 fs_reclaim irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#4 pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &c->lock irq_context: 0 kn->active#4 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#4 &n->list_lock irq_context: 0 kn->active#4 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &device->physical_node_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &____s->seqcount irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 udc_lock irq_context: 0 kn->active#4 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fw_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#4 remove_cache_srcu irq_context: 0 kn->active#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &____s->seqcount irq_context: 0 kn->active#4 remove_cache_srcu &c->lock irq_context: 0 kn->active#4 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 kn->active#4 &____s->seqcount irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &rfkill->lock irq_context: 0 sb_writers#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 kn->active#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &n->list_lock irq_context: 0 &mm->mmap_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu irq_context: 0 &mm->mmap_lock remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &n->list_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &u->bindlock irq_context: 0 &u->bindlock fs_reclaim irq_context: 0 &u->bindlock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &u->bindlock pool_lock#2 irq_context: 0 &u->bindlock batched_entropy_u32.lock irq_context: 0 &u->bindlock &net->unx.table.locks[i] irq_context: 0 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 irq_context: 0 &u->bindlock &net->unx.table.locks[i]/1 irq_context: 0 &u->lock &u->lock/1 &sk->sk_peer_lock irq_context: 0 &u->lock &u->lock/1 &dentry->d_lock irq_context: 0 &u->lock &u->lock/1 &sk->sk_peer_lock &sk->sk_peer_lock/1 irq_context: 0 &u->lock &u->lock/1 &sk->sk_peer_lock/1 irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &dentry->d_lock irq_context: 0 &group->mark_mutex &sb->s_type->i_lock_key &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &group->notification_waitq irq_context: 0 &group->notification_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#5 &rq->__lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#30 sb_writers#3 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 &____s->seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#30 sb_writers#3 lock#4 irq_context: 0 &type->s_umount_key#30 sb_writers#3 pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 &mapping->i_private_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 lock#4 &lruvec->lru_lock irq_context: 0 batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock &base->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 &xa->xa_lock#8 pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 bit_wait_table + i irq_context: 0 &type->s_umount_key#30 sb_writers#3 &rq->__lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &eli->li_list_mtx irq_context: 0 &type->s_umount_key#30 sb_writers#3 &journal->j_state_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &base->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &retval->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &c->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &____s->seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &c->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &____s->seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &n->list_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &x->wait#26 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &__ctx->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_node_0 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &__ctx->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &retval->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock rcu_node_0 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &x->wait#26 irq_context: hardirq &x->wait#26 &p->pi_lock irq_context: hardirq &x->wait#26 &p->pi_lock &rq->__lock irq_context: hardirq &x->wait#26 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem (&timer.timer) irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock &q->requeue_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) &q->requeue_lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) &hctx->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &hctx->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &nvmeq->sq_lock irq_context: hardirq &fq->mq_flush_lock irq_context: hardirq &fq->mq_flush_lock tk_core.seq.seqcount irq_context: hardirq &fq->mq_flush_lock &x->wait#26 irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &journal->j_wait_updates irq_context: 0 &type->s_umount_key#30 sb_writers#3 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock pool_lock#2 irq_context: 0 sb_writers#5 &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &p->pi_lock irq_context: 0 sb_writers#5 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &dentry->d_lock &wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 rcu_node_0 irq_context: softirq &(&ovs_net->masks_rebalance)->timer irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) ovs_mutex irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&ovs_net->masks_rebalance)->work) &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 sb_writers#3 jbd2_handle &meta_group_info[i]->alloc_sem &pcp->lock &zone->lock irq_context: 0 &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#5 fs_reclaim irq_context: 0 kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#5 &c->lock irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex irq_context: 0 &p->lock &of->mutex kn->active#5 param_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rename_lock.seqcount irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem irq_context: 0 sb_writers#7 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &dentry->d_lock irq_context: 0 sb_writers#7 tomoyo_ss irq_context: 0 sb_writers#7 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#7 kn->active#5 fs_reclaim irq_context: 0 sb_writers#7 kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 kn->active#5 &kernfs_locks->open_file_mutex[count] irq_context: 0 sb_writers#7 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 sb_writers#7 kn->active#5 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 tomoyo_ss &c->lock irq_context: 0 sb_writers#7 iattr_mutex irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex &pcp->lock &zone->lock irq_context: 0 &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex &____s->seqcount irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex &c->lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex pool_lock#2 irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex tk_core.seq.seqcount irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &sem->wait_lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &p->pi_lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#5 param_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#5 param_lock disk_events_mutex irq_context: 0 &eli->li_list_mtx &obj_hash[i].lock irq_context: 0 &eli->li_list_mtx pool_lock#2 irq_context: 0 ext4_li_mtx irq_context: 0 ext4_li_mtx &eli->li_list_mtx irq_context: 0 ext4_li_mtx &obj_hash[i].lock irq_context: 0 ext4_li_mtx pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#4 &c->lock irq_context: 0 &p->lock &of->mutex kn->active#4 &____s->seqcount irq_context: 0 &p->lock &of->mutex kn->active#4 pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#4 &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#4 &n->list_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &n->list_lock &c->lock irq_context: 0 &p->lock &of->mutex kn->active#4 &pcp->lock &zone->lock irq_context: 0 &p->lock &n->list_lock irq_context: 0 &p->lock &n->list_lock &c->lock irq_context: 0 &p->lock &of->mutex kn->active#4 &device->physical_node_lock irq_context: 0 &p->lock &mm->mmap_lock fs_reclaim irq_context: 0 &p->lock &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &mm->mmap_lock &____s->seqcount irq_context: 0 &p->lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &p->lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &p->lock remove_cache_srcu irq_context: 0 &p->lock remove_cache_srcu quarantine_lock irq_context: 0 &p->lock remove_cache_srcu &c->lock irq_context: 0 &p->lock remove_cache_srcu &n->list_lock irq_context: 0 &p->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &p->lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &p->lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mapping.invalidate_lock rcu_read_lock &n->list_lock irq_context: 0 mapping.invalidate_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 mapping.invalidate_lock &folio_wait_table[i] &p->pi_lock irq_context: 0 mapping.invalidate_lock &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 mapping.invalidate_lock &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock &rq->__lock irq_context: 0 &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 mapping.invalidate_lock &obj_hash[i].lock irq_context: 0 mapping.invalidate_lock &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 &____s->seqcount#6/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &c->lock irq_context: 0 &p->lock &of->mutex kn->active#4 quarantine_lock irq_context: 0 &rq->__lock &obj_hash[i].lock irq_context: 0 &rq->__lock &base->lock irq_context: 0 &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 kn->active#6 fs_reclaim irq_context: 0 kn->active#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &of->mutex irq_context: 0 kn->active#7 fs_reclaim irq_context: 0 kn->active#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#8 fs_reclaim irq_context: 0 kn->active#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#8 &c->lock irq_context: 0 kn->active#8 &n->list_lock irq_context: 0 kn->active#8 &n->list_lock &c->lock irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#9 fs_reclaim irq_context: 0 kn->active#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#10 fs_reclaim irq_context: 0 kn->active#10 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#11 fs_reclaim irq_context: 0 kn->active#11 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#12 fs_reclaim irq_context: 0 kn->active#12 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock irq_context: 0 &p->lock &of->mutex kn->active#4 udc_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#5 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sem->wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &n->list_lock &c->lock irq_context: 0 mapping.invalidate_lock &n->list_lock irq_context: 0 mapping.invalidate_lock &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock irq_context: 0 &p->lock remove_cache_srcu &rq->__lock irq_context: 0 &p->lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rq->__lock irq_context: 0 kn->active#13 fs_reclaim irq_context: 0 kn->active#13 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#14 fs_reclaim irq_context: 0 kn->active#14 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#14 dev_base_lock irq_context: 0 kn->active#15 fs_reclaim irq_context: 0 kn->active#15 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#15 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#16 fs_reclaim irq_context: 0 kn->active#16 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#16 dev_base_lock irq_context: 0 kn->active#17 fs_reclaim irq_context: 0 kn->active#17 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#17 &c->lock irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#18 fs_reclaim irq_context: 0 kn->active#18 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#19 fs_reclaim irq_context: 0 kn->active#19 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#19 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#20 fs_reclaim irq_context: 0 kn->active#20 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#20 &c->lock irq_context: 0 kn->active#21 fs_reclaim irq_context: 0 kn->active#20 &____s->seqcount irq_context: 0 kn->active#21 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#21 &c->lock irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &p->lock &of->mutex kn->active#20 dev_base_lock irq_context: 0 kn->active#22 fs_reclaim irq_context: 0 kn->active#22 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#22 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#22 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#22 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#22 dev_base_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &c->lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &n->list_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 kn->active#13 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#17 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 &vma->vm_lock->lock &c->lock irq_context: 0 kn->active#23 fs_reclaim irq_context: 0 kn->active#23 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#23 &c->lock irq_context: 0 kn->active#23 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#23 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#23 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#24 fs_reclaim irq_context: 0 kn->active#24 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#4 fw_lock irq_context: 0 kn->active#25 fs_reclaim irq_context: 0 kn->active#25 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#25 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#25 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#25 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#26 fs_reclaim irq_context: 0 kn->active#26 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#27 fs_reclaim irq_context: 0 kn->active#27 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#27 &c->lock irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#28 fs_reclaim irq_context: 0 kn->active#28 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#28 &c->lock irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#28 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &n->list_lock irq_context: 0 kn->active#29 fs_reclaim irq_context: 0 kn->active#29 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#29 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#29 dev_base_lock irq_context: 0 kn->active#30 fs_reclaim irq_context: 0 kn->active#30 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#30 &c->lock irq_context: 0 kn->active#30 &n->list_lock irq_context: 0 kn->active#30 &n->list_lock &c->lock irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#30 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &of->mutex kn->active#30 &dev->power.lock irq_context: 0 &of->mutex kn->active#30 pci_lock irq_context: 0 kn->active#31 fs_reclaim irq_context: 0 kn->active#31 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#31 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#32 fs_reclaim irq_context: 0 kn->active#32 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#32 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#32 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#32 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 slock-AF_INET/1 irq_context: 0 rtnl_mutex devnet_rename_sem irq_context: 0 rtnl_mutex devnet_rename_sem (console_sem).lock irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex devnet_rename_sem console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex devnet_rename_sem fs_reclaim irq_context: 0 rtnl_mutex devnet_rename_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex devnet_rename_sem pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem &k->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem fs_reclaim irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem kernfs_rename_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex devnet_rename_sem kernfs_rename_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex fs_reclaim irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex nl_table_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex nl_table_wait.lock irq_context: 0 rtnl_mutex devnet_rename_sem &obj_hash[i].lock irq_context: 0 rtnl_mutex &x->wait#2 irq_context: 0 kn->active#24 &c->lock irq_context: 0 kn->active#33 fs_reclaim irq_context: 0 kn->active#33 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#13 &c->lock irq_context: 0 kn->active#13 &n->list_lock irq_context: 0 kn->active#13 &n->list_lock &c->lock irq_context: 0 kn->active#34 fs_reclaim irq_context: 0 kn->active#34 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#34 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#34 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#34 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#34 fs_reclaim irq_context: 0 &p->lock &of->mutex kn->active#34 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &p->lock &of->mutex kn->active#34 pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#34 &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#34 &c->lock irq_context: 0 kn->active#15 &c->lock irq_context: 0 kn->active#21 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 &evdev->client_lock irq_context: 0 &evdev->mutex irq_context: 0 &evdev->mutex &dev->mutex#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rcu_read_lock &____s->seqcount#6 irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#15 irq_context: 0 mapping.invalidate_lock#2 irq_context: 0 mapping.invalidate_lock#2 mmu_notifier_invalidate_range_start irq_context: 0 mapping.invalidate_lock#2 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 mapping.invalidate_lock#2 &____s->seqcount irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#8 irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#8 pool_lock#2 irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#8 &c->lock irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#8 &n->list_lock irq_context: 0 mapping.invalidate_lock#2 &xa->xa_lock#8 &n->list_lock &c->lock irq_context: 0 mapping.invalidate_lock#2 lock#4 irq_context: 0 mapping.invalidate_lock#2 pool_lock#2 irq_context: 0 mapping.invalidate_lock#2 tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &nvmeq->sq_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 kn->active#24 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#33 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 mapping.invalidate_lock#2 &pcp->lock &zone->lock irq_context: 0 &ep->mtx &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#34 &n->list_lock irq_context: 0 &p->lock &of->mutex kn->active#34 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 kn->active#34 &c->lock irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 rtnl_mutex sysctl_lock krc.lock irq_context: 0 rtnl_mutex &nft_net->commit_mutex irq_context: 0 rtnl_mutex proc_subdir_lock irq_context: 0 rtnl_mutex &ent->pde_unload_lock irq_context: 0 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 rtnl_mutex sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex proc_subdir_lock irq_context: 0 rtnl_mutex target_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET#2 irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_INET irq_context: 0 &evdev->mutex &dev->mutex#2 &obj_hash[i].lock irq_context: 0 &evdev->mutex &dev->mutex#2 &x->wait#2 irq_context: 0 &evdev->mutex &dev->mutex#2 &rq->__lock irq_context: 0 &evdev->mutex &dev->mutex#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 kn->active#4 &c->lock irq_context: 0 &type->i_mutex_dir_key#2 &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#2 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sem->wait_lock irq_context: 0 sb_writers &p->pi_lock irq_context: 0 sb_writers &p->pi_lock &rq->__lock irq_context: 0 sb_writers &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sem->wait_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#9 &c->lock irq_context: 0 kn->active#10 &c->lock irq_context: 0 sb_writers &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu irq_context: 0 &vma->vm_lock->lock remove_cache_srcu quarantine_lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &c->lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &n->list_lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#2 &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 mapping.invalidate_lock#2 &c->lock irq_context: 0 mapping.invalidate_lock#2 lock#4 &lruvec->lru_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &rq->__lock irq_context: softirq rcu_callback &____s->seqcount irq_context: 0 kn->active#6 &c->lock irq_context: 0 kn->active#6 &n->list_lock irq_context: 0 kn->active#6 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 key#9 irq_context: 0 kn->active#11 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#2 &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss quarantine_lock irq_context: 0 kn->active#7 &c->lock irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#6 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#11 remove_cache_srcu irq_context: 0 kn->active#11 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 kn->active#9 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#12 &c->lock irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 kn->active#12 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#11 &n->list_lock irq_context: 0 kn->active#11 &n->list_lock &c->lock irq_context: 0 kn->active#11 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &c->lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &n->list_lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock pool_lock#2 irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &retval->lock irq_context: 0 lock#4 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#7 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock rename_lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#10 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 kn->active#12 &n->list_lock irq_context: 0 kn->active#12 &n->list_lock &c->lock irq_context: 0 kn->active#7 &n->list_lock irq_context: 0 kn->active#7 &n->list_lock &c->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &group->mark_mutex &pcp->lock &zone->lock irq_context: 0 &group->mark_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 &____s->seqcount#6/1 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rename_lock rename_lock.seqcount &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#8 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#5 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &lruvec->lru_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &info->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &xa->xa_lock#8 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 kn->active#26 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 kn->active#26 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#27 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#35 fs_reclaim irq_context: 0 kn->active#35 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#35 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#35 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#35 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 mapping.invalidate_lock#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#10 &n->list_lock irq_context: 0 kn->active#10 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss quarantine_lock irq_context: 0 &type->i_mutex_dir_key#5 &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &p->pi_lock irq_context: 0 mapping.invalidate_lock#2 batched_entropy_u8.lock irq_context: 0 mapping.invalidate_lock#2 kfence_freelist_lock irq_context: 0 &p->lock rcu_read_lock rcu_node_0 irq_context: 0 &p->lock rcu_read_lock &rq->__lock irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 kn->active#8 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 mapping.invalidate_lock#2 rcu_read_lock &____s->seqcount irq_context: 0 kn->active#7 remove_cache_srcu irq_context: 0 kn->active#7 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#9 &n->list_lock irq_context: 0 kn->active#9 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &n->list_lock &c->lock irq_context: 0 &u->iolock &mm->mmap_lock &rq->__lock irq_context: 0 mapping.invalidate_lock#2 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 kn->active#8 remove_cache_srcu irq_context: 0 kn->active#8 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#8 remove_cache_srcu &c->lock irq_context: 0 kn->active#8 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#8 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#8 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &pipe->rd_wait irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &n->list_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &n->list_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem quarantine_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock quarantine_lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sig->cred_guard_mutex &n->list_lock irq_context: 0 &sig->cred_guard_mutex &n->list_lock &c->lock irq_context: 0 &ep->mtx key#10 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 kn->active#36 fs_reclaim irq_context: 0 kn->active#36 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#36 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#36 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#36 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#37 fs_reclaim irq_context: 0 kn->active#37 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#37 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#2 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#2 &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &sem->wait_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &mapping->i_private_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &sem->wait_lock irq_context: 0 &mm->mmap_lock &p->pi_lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock fs_reclaim &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex &folio_wait_table[i] irq_context: 0 videodev_lock irq_context: 0 fh->state->lock irq_context: 0 &vdev->fh_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_node_0 irq_context: 0 &dev->dev_mutex irq_context: 0 &dev->dev_mutex fs_reclaim irq_context: 0 &dev->dev_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->dev_mutex &c->lock irq_context: 0 &dev_instance->mutex irq_context: 0 &dev_instance->mutex fs_reclaim irq_context: 0 &dev->dev_mutex pool_lock#2 irq_context: 0 &dev_instance->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev_instance->mutex &c->lock irq_context: 0 &dev->dev_mutex vim2m:1183:(hdl)->_lock irq_context: 0 &dev_instance->mutex pool_lock#2 irq_context: 0 &dev_instance->mutex vicodec_core:1851:(hdl)->_lock irq_context: 0 &dev->dev_mutex &obj_hash[i].lock irq_context: 0 &dev->dev_mutex &vdev->fh_lock irq_context: 0 &dev_instance->mutex &vdev->fh_lock irq_context: 0 &mdev->req_queue_mutex irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex irq_context: 0 &mdev->req_queue_mutex &vdev->fh_lock irq_context: 0 &mdev->req_queue_mutex &mdev->graph_mutex irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &m2m_dev->job_spinlock irq_context: 0 &mdev->req_queue_mutex vim2m:1183:(hdl)->_lock irq_context: 0 &mdev->req_queue_mutex vim2m:1183:(hdl)->_lock &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &q->done_wq irq_context: 0 &mdev->req_queue_mutex vim2m:1183:(hdl)->_lock pool_lock#2 irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &q->mmap_lock irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex irq_context: 0 &mdev->req_queue_mutex &dev_instance->mutex pool_lock#2 irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &m2m_dev->job_spinlock irq_context: 0 &mdev->req_queue_mutex vicodec_core:1851:(hdl)->_lock irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &q->done_wq irq_context: 0 &mdev->req_queue_mutex vicodec_core:1851:(hdl)->_lock &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &q->mmap_lock irq_context: 0 &mdev->req_queue_mutex vicodec_core:1851:(hdl)->_lock pool_lock#2 irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex &dev->dev_mutex pool_lock#2 irq_context: 0 &mdev->req_queue_mutex &obj_hash[i].lock irq_context: 0 &mdev->req_queue_mutex pool_lock#2 irq_context: 0 &pipe->rd_wait &ep->lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &sem->wait_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem key irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem percpu_counters_lock irq_context: 0 &dev_instance->mutex &____s->seqcount irq_context: 0 &ep->mtx rcu_read_lock &pipe->rd_wait irq_context: 0 &ep->mtx &obj_hash[i].lock irq_context: 0 &sighand->signalfd_wqh irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock irq_context: 0 kn->active#38 fs_reclaim irq_context: 0 kn->active#38 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 quarantine_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock remove_cache_srcu irq_context: 0 &mm->mmap_lock remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &n->list_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem quarantine_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rq->__lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock cpu_asid_lock irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#38 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem quarantine_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock key irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock irq_context: 0 &mm->mmap_lock rcu_read_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock key irq_context: 0 &mm->mmap_lock pcpu_lock irq_context: 0 &mm->mmap_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 kn->active#38 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &vcapture->lock irq_context: 0 &vcapture->lock &q->done_wq irq_context: 0 &vcapture->lock &q->mmap_lock irq_context: 0 &mdev->graph_mutex irq_context: 0 &sig->cred_guard_mutex &stopper->lock irq_context: 0 &sig->cred_guard_mutex &stop_pi_lock irq_context: 0 &sig->cred_guard_mutex &stop_pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#38 &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &x->wait#8 irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu quarantine_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 kn->active#38 remove_cache_srcu irq_context: 0 kn->active#38 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#38 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#38 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#38 remove_cache_srcu &c->lock irq_context: 0 kn->active#38 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 kn->active#38 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#38 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 &dev->mutex#3 irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 &vdev->fh_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem key irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem pcpu_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem percpu_counters_lock irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 &q->done_wq irq_context: 0 &mdev->req_queue_mutex &dev->mutex#3 &q->mmap_lock irq_context: 0 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#38 &pcp->lock &zone->lock irq_context: 0 kn->active#38 &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock sb_writers#3 &c->lock irq_context: 0 &mm->mmap_lock sb_writers#3 &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock sb_writers#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock sb_writers#3 &____s->seqcount irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 &ep->mtx &mm->mmap_lock &rq->__lock irq_context: 0 sb_writers#3 remove_cache_srcu &rq->__lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu quarantine_lock irq_context: 0 &p->lock remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &base->lock &obj_hash[i].lock irq_context: 0 kn->active#4 remove_cache_srcu &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &lo->lo_mutex irq_context: 0 &disk->open_mutex &lo->lo_mutex irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &group->mark_mutex remove_cache_srcu irq_context: 0 &group->mark_mutex remove_cache_srcu quarantine_lock irq_context: 0 &disk->open_mutex nbd_index_mutex irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock mmu_notifier_invalidate_range_start irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock pool_lock#2 irq_context: 0 &disk->open_mutex &nbd->config_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &bdev->bd_size_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &q->queue_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 &disk->open_mutex &nbd->config_lock set->srcu irq_context: 0 &disk->open_mutex &nbd->config_lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &disk->open_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex &nbd->config_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &disk->open_mutex &nbd->config_lock &x->wait#3 irq_context: 0 &disk->open_mutex &nbd->config_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 &disk->open_mutex &nbd->config_lock set->srcu irq_context: 0 &disk->open_mutex &nbd->config_lock pool_lock#2 irq_context: 0 &disk->open_mutex &nbd->config_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &c->lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &rq->__lock irq_context: 0 sb_writers#5 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&ssp->srcu_sup->work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &disk->open_mutex &nbd->config_lock rcu_read_lock &rq->__lock irq_context: 0 &disk->open_mutex &nbd->config_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp &rq->__lock irq_context: 0 (wq_completion)rcu_gp &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &type->i_mutex_dir_key#5 &rq->__lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss &base->lock irq_context: 0 tomoyo_ss &base->lock &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#4 &rfkill->lock irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 mapping.invalidate_lock &xa->xa_lock#8 &n->list_lock irq_context: 0 mapping.invalidate_lock &xa->xa_lock#8 &n->list_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_node_0 irq_context: softirq mm/vmstat.c:2022 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &rq->__lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu quarantine_lock irq_context: 0 tomoyo_ss rcu_read_lock rename_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &c->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &n->list_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#14 &c->lock irq_context: 0 kn->active#14 &pcp->lock &zone->lock irq_context: 0 kn->active#14 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#14 &____s->seqcount irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 &disk->open_mutex &new->lock irq_context: 0 &disk->open_mutex &new->lock &mtdblk->cache_mutex irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock irq_context: 0 tomoyo_ss rename_lock irq_context: 0 kn->active#39 fs_reclaim irq_context: 0 kn->active#39 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#39 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &mtd->master.chrdev_lock irq_context: 0 &mtd->master.chrdev_lock &mm->mmap_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&krcp->monitor_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &c->lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &n->list_lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#5 sched_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 kfence_freelist_lock irq_context: 0 kn->active#4 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&wb->dwork)->timer irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &p->sequence irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) key#11 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 &p->lock &of->mutex &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#2 remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key#2 remove_cache_srcu &c->lock irq_context: 0 &type->i_mutex_dir_key#2 remove_cache_srcu &n->list_lock irq_context: 0 &type->i_mutex_dir_key#2 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#2 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &type->i_mutex_dir_key#2 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &base->lock irq_context: 0 &p->lock &of->mutex kn->active#4 &base->lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 rename_lock irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &____s->seqcount#6 &____s->seqcount#6/1 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock &dentry->d_lock/2 &dentry->d_lock/3 &wq irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock/2 irq_context: 0 &type->i_mutex_dir_key#4 rename_lock rename_lock.seqcount &dentry->d_lock/2 &dentry->d_lock/3 irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock pool_lock#2 irq_context: 0 kn->active#40 fs_reclaim irq_context: 0 kn->active#40 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#40 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#40 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#40 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dentry->d_lock &lru->node[i].lock irq_context: 0 kn->active#40 &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 tk_core.seq.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &group->notification_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &group->notification_waitq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &group->notification_waitq &p->pi_lock irq_context: 0 sb_writers#5 &s->s_inode_list_lock irq_context: 0 sb_writers#5 &info->lock irq_context: 0 sb_writers#5 &obj_hash[i].lock irq_context: 0 sb_writers#5 pool_lock#2 irq_context: 0 sb_writers#5 &sbinfo->stat_lock irq_context: 0 sb_writers#5 &xa->xa_lock#8 irq_context: 0 sb_writers#5 &fsnotify_mark_srcu irq_context: 0 &mark->lock irq_context: 0 &group->inotify_data.idr_lock irq_context: 0 &group->inotify_data.idr_lock &obj_hash[i].lock irq_context: 0 &group->inotify_data.idr_lock pool_lock#2 irq_context: 0 &conn->lock irq_context: 0 destroy_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work irq_context: 0 (wq_completion)events_unbound connector_reaper_work destroy_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)events_unbound connector_reaper_work &fsnotify_mark_srcu irq_context: 0 (wq_completion)events_unbound connector_reaper_work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work &x->wait#3 irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fs/notify/mark.c:89 irq_context: 0 (wq_completion)events_unbound connector_reaper_work pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work irq_context: 0 (wq_completion)events_unbound (reaper_work).work destroy_lock irq_context: 0 (reaper_work).work irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(sdp, lock) irq_context: 0 (wq_completion)events_unbound (reaper_work).work &fsnotify_mark_srcu irq_context: 0 &x->wait#10 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work &x->wait#3 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_state_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (&journal->j_commit_timer) irq_context: 0 &journal->j_checkpoint_mutex irq_context: 0 &journal->j_checkpoint_mutex mmu_notifier_invalidate_range_start irq_context: 0 &journal->j_checkpoint_mutex pool_lock#2 irq_context: 0 &journal->j_checkpoint_mutex tk_core.seq.seqcount irq_context: 0 &journal->j_checkpoint_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 &journal->j_checkpoint_mutex rcu_read_lock &nvmeq->sq_lock irq_context: 0 &journal->j_checkpoint_mutex bit_wait_table + i irq_context: 0 &journal->j_checkpoint_mutex &rq->__lock irq_context: 0 &journal->j_checkpoint_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_checkpoint_mutex &journal->j_state_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_updates irq_context: 0 &journal->j_list_lock irq_context: 0 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &ei->i_es_lock irq_context: 0 &mapping->i_private_lock irq_context: 0 &ret->b_state_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock irq_context: 0 &ei->i_es_lock key#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock &base->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &journal->j_state_lock irq_context: 0 &journal->j_state_lock &journal->j_list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle bit_wait_table + i irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rq->__lock irq_context: 0 rcu_read_lock &retval->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fq->mq_flush_lock &q->requeue_lock irq_context: 0 &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) &__ctx->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &__ctx->lock irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &rq->__lock irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &memcg->move_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &xa->xa_lock#8 irq_context: 0 &ret->b_state_lock &journal->j_list_lock &sb->s_type->i_lock_key#3 irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->list_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->list_lock &sb->s_type->i_lock_key#3 irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &xa->xa_lock#8 key#11 irq_context: 0 &journal->j_state_lock &journal->j_list_lock irq_context: 0 &sbi->s_md_lock irq_context: 0 &journal->j_fc_wait irq_context: 0 &journal->j_history_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &c->lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &n->list_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &ei->i_data_sem &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &n->list_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex rlock-AF_NETLINK irq_context: 0 rtnl_mutex (inetaddr_validator_chain).rwsem irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fib_info_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &dir->lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem nl_table_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rhashtable_bucket irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_lock irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_wait.lock irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_NETLINK irq_context: 0 &sb->s_type->i_mutex_key#10 &nlk->wait irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem hwsim_radio_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex batched_entropy_u32.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex rhashtable_bucket irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex rhashtable_bucket rhashtable_bucket/1 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &n->list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &n->list_lock &c->lock irq_context: 0 rtnl_mutex _xmit_LOOPBACK irq_context: 0 rtnl_mutex netpoll_srcu irq_context: 0 rtnl_mutex &n->list_lock irq_context: 0 rtnl_mutex &n->list_lock &c->lock irq_context: 0 rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex &im->lock irq_context: 0 rtnl_mutex fib_info_lock irq_context: 0 rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &dir->lock#2 irq_context: 0 rtnl_mutex cbs_list_lock irq_context: 0 rtnl_mutex &ndev->lock irq_context: 0 rtnl_mutex &idev->mc_lock irq_context: 0 rtnl_mutex (inet6addr_validator_chain).rwsem irq_context: 0 rtnl_mutex rcu_read_lock &net->ipv6.addrconf_hash_lock irq_context: 0 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 rtnl_mutex &ifa->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex &tb->tb6_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &____s->seqcount irq_context: 0 rtnl_mutex &tb->tb6_lock &c->lock irq_context: 0 rtnl_mutex &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &ndev->lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock irq_context: softirq rcu_callback &dir->lock irq_context: softirq rcu_callback &dir->lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 mapping.invalidate_lock &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_INET6 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &xa->xa_lock#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 lock#4 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &nvmeq->sq_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 inode_hash_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 sb_writers#5 tomoyo_ss irq_context: 0 sb_writers#5 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#5 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#5 &xattrs->lock irq_context: 0 &u->lock/1 irq_context: 0 userns_state_mutex irq_context: 0 mapping.invalidate_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &mapping->i_mmap_rwsem irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &sb->s_type->i_lock_key &xa->xa_lock#8 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 lock#5 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &lruvec->lru_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#5 irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 &mm->mmap_lock irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 &xa->xa_lock#8 irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 &sb->s_type->i_lock_key irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 &info->lock irq_context: 0 &f->f_pos_lock sb_writers#5 &sb->s_type->i_mutex_key#13 lock#4 irq_context: 0 &sb->s_type->i_lock_key#4 irq_context: 0 &sb->s_type->i_lock_key#4 &dentry->d_lock irq_context: 0 sk_lock-AF_UNIX irq_context: 0 sk_lock-AF_UNIX slock-AF_UNIX irq_context: 0 slock-AF_UNIX irq_context: 0 cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem key irq_context: 0 cgroup_threadgroup_rwsem pcpu_lock irq_context: 0 cgroup_threadgroup_rwsem percpu_counters_lock irq_context: 0 cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 cgroup_threadgroup_rwsem &cfs_rq->removed.lock irq_context: 0 vmap_purge_lock &rq->__lock irq_context: 0 sk_lock-AF_INET irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 irq_context: 0 slock-AF_INET#2 irq_context: 0 sk_lock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 slock-AF_INET6 irq_context: 0 slock-AF_INET6 irq_context: 0 sk_lock-AF_INET &table->hash[i].lock irq_context: 0 sk_lock-AF_INET &table->hash[i].lock clock-AF_INET irq_context: 0 sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 sk_lock-AF_INET6 &table->hash[i].lock clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 vmap_purge_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_data_sem &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 vmap_purge_lock vmap_purge_lock.wait_lock irq_context: 0 vmap_purge_lock rcu_read_lock &rq->__lock irq_context: 0 vmap_purge_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vmap_purge_lock.wait_lock irq_context: 0 mapping.invalidate_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_NETLINK &mm->mmap_lock irq_context: 0 sk_lock-AF_NETLINK fs_reclaim irq_context: 0 sk_lock-AF_NETLINK fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_NETLINK pool_lock#2 irq_context: 0 sk_lock-AF_NETLINK free_vmap_area_lock irq_context: 0 sk_lock-AF_NETLINK free_vmap_area_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK free_vmap_area_lock pool_lock#2 irq_context: 0 sk_lock-AF_NETLINK vmap_area_lock irq_context: 0 sk_lock-AF_NETLINK &____s->seqcount irq_context: 0 sk_lock-AF_NETLINK &c->lock irq_context: 0 sk_lock-AF_NETLINK pcpu_alloc_mutex irq_context: 0 sk_lock-AF_NETLINK pcpu_alloc_mutex pcpu_lock irq_context: 0 sk_lock-AF_NETLINK &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK batched_entropy_u32.lock irq_context: 0 sk_lock-AF_NETLINK vmap_purge_lock irq_context: 0 sk_lock-AF_NETLINK vmap_purge_lock purge_vmap_area_lock irq_context: 0 sk_lock-AF_NETLINK &fp->aux->used_maps_mutex irq_context: 0 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#41 fs_reclaim irq_context: 0 kn->active#41 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#41 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#41 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#41 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &ndev->lock &ifa->lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq#2 irq_context: 0 vlan_ioctl_mutex &mm->mmap_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex irq_context: 0 cb_lock irq_context: 0 cb_lock genl_mutex irq_context: 0 cb_lock genl_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex &c->lock irq_context: 0 cb_lock genl_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex &obj_hash[i].lock irq_context: 0 cb_lock fs_reclaim irq_context: 0 cb_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock pool_lock#2 irq_context: 0 cb_lock &c->lock irq_context: 0 cb_lock rlock-AF_NETLINK irq_context: 0 cb_lock rtnl_mutex irq_context: 0 cb_lock &obj_hash[i].lock irq_context: 0 cb_lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &n->list_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#4 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 dev_addr_sem irq_context: 0 cb_lock &n->list_lock irq_context: 0 cb_lock &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx irq_context: 0 cb_lock &rdev->wiphy.mtx irq_context: 0 cb_lock &rdev->wiphy.mtx fs_reclaim irq_context: 0 cb_lock &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock &rdev->wiphy.mtx pool_lock#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &c->lock irq_context: 0 cb_lock &rdev->wiphy.mtx &____s->seqcount irq_context: 0 cb_lock &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx rlock-AF_NETLINK irq_context: 0 cb_lock nlk_cb_mutex-GENERIC irq_context: 0 cb_lock nlk_cb_mutex-GENERIC fs_reclaim irq_context: 0 cb_lock nlk_cb_mutex-GENERIC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock nlk_cb_mutex-GENERIC pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &pcp->lock &zone->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &____s->seqcount irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rtnl_mutex &rdev->wiphy.mtx irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx pool_lock#2 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rdev->wiphy.mtx &rdev->bss_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC rlock-AF_NETLINK irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex &n->list_lock &c->lock irq_context: 0 cb_lock &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->alloc_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 lock#4 &lruvec->lru_lock irq_context: 0 &mm->mmap_lock lock#4 &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 sb_writers#5 fs_reclaim irq_context: 0 sb_writers#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_lock_key &xa->xa_lock#8 irq_context: 0 sb_writers#5 lock#4 irq_context: 0 sb_writers#5 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#5 lock#5 irq_context: 0 sb_writers#5 &lruvec->lru_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 &vma->vm_lock->lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 vmap_area_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 mount_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 &sb->s_type->i_lock_key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 &wb->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 sb_writers#5 &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 &pipe->mutex/1 &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem ptlock_ptr(ptdesc)#2 irq_context: softirq (&net->sctp.addr_wq_timer) irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#5 irq_context: 0 &sig->cred_guard_mutex &p->alloc_lock &x->wait#25 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &meta->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &obj_hash[i].lock pool_lock irq_context: 0 lock pidmap_lock &n->list_lock irq_context: 0 lock pidmap_lock &n->list_lock &c->lock irq_context: 0 tasklist_lock &sighand->siglock batched_entropy_u8.lock irq_context: 0 tasklist_lock &sighand->siglock kfence_freelist_lock irq_context: 0 &sighand->siglock &meta->lock irq_context: 0 &sighand->siglock kfence_freelist_lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key irq_context: 0 rtnl_mutex &dev_addr_list_lock_key pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &pnettable->lock irq_context: 0 rtnl_mutex smc_ib_devices.mutex irq_context: 0 rtnl_mutex napi_hash_lock irq_context: 0 rtnl_mutex lapb_list_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&aux->work) irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock purge_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock free_vmap_area_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock free_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pcpu_lock irq_context: 0 rtnl_mutex x25_neigh_list_lock irq_context: 0 rtnl_mutex console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex console_lock console_srcu console_owner irq_context: 0 rtnl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &u->lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex _xmit_ETHER irq_context: 0 rtnl_mutex &tb->tb6_lock rlock-AF_NETLINK irq_context: 0 rtnl_mutex _xmit_SLIP irq_context: hardirq log_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &sem->wait_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&eql->timer) irq_context: softirq (&eql->timer) &eql->queue.lock irq_context: softirq (&eql->timer) &obj_hash[i].lock irq_context: softirq (&eql->timer) &base->lock irq_context: softirq (&eql->timer) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex quarantine_lock irq_context: 0 rtnl_mutex remove_cache_srcu irq_context: 0 rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rq->__lock irq_context: 0 rtnl_mutex free_vmap_area_lock irq_context: 0 rtnl_mutex vmap_area_lock irq_context: 0 rtnl_mutex init_mm.page_table_lock irq_context: 0 rtnl_mutex &cma->lock irq_context: 0 rtnl_mutex cma_mutex irq_context: 0 rtnl_mutex cma_mutex &zone->lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cma_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex cma_mutex lock#2 irq_context: 0 rtnl_mutex cma_mutex lock#2 &obj_hash[i].lock irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cma_mutex lock#2 &rq->__lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(work) lock#4 &obj_hash[i].lock irq_context: 0 rtnl_mutex cma_mutex lock#2 (work_completion)(work) irq_context: 0 rtnl_mutex cma_mutex lock#2 rcu_read_lock (wq_completion)mm_percpu_wq irq_context: 0 rtnl_mutex cma_mutex lock#2 &x->wait#10 irq_context: 0 rtnl_mutex cma_mutex lock#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex cma_mutex &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex fs_reclaim irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex pool_lock#2 irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex jump_label_mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex jump_label_mutex patch_lock irq_context: 0 rtnl_mutex &zone->lock irq_context: 0 rtnl_mutex cma_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_node_0 irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) (console_sem).lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) console_lock console_srcu console_owner irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) lweventlist_lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_node_0 irq_context: hardirq log_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gve (work_completion)(&priv->service_task) &obj_hash[i].lock irq_context: 0 rtnl_mutex batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 rtnl_mutex &sem->wait_lock irq_context: 0 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex cbs_list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock rlock-AF_NETLINK irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rlock-AF_NETLINK irq_context: 0 rtnl_mutex &rfkill->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &base->lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &data->mutex irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 rtnl_mutex _xmit_ETHER &rdev->wiphy_work_lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &local->filter_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#14 &n->list_lock irq_context: 0 kn->active#14 &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rq->__lock irq_context: 0 rtnl_mutex _xmit_VOID irq_context: 0 &u->iolock &u->lock irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &u->iolock &mm->mmap_lock fs_reclaim irq_context: 0 &u->iolock &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &u->iolock &mm->mmap_lock &____s->seqcount irq_context: 0 &u->iolock &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override pool_lock irq_context: 0 mapping.invalidate_lock key irq_context: 0 mapping.invalidate_lock pcpu_lock irq_context: 0 mapping.invalidate_lock percpu_counters_lock irq_context: 0 rtnl_mutex _xmit_X25 irq_context: 0 rtnl_mutex lapb_list_lock irq_context: 0 rtnl_mutex lapb_list_lock pool_lock#2 irq_context: 0 rtnl_mutex lapb_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex lapb_list_lock &base->lock irq_context: 0 rtnl_mutex lapb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapbeth->up_lock irq_context: 0 rtnl_mutex &lapb->lock irq_context: 0 rtnl_mutex &lapb->lock pool_lock#2 irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rtnl_mutex &lapb->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapb->lock &base->lock irq_context: 0 rtnl_mutex &lapb->lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 batched_entropy_u32.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &ul->lock irq_context: 0 sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 pool_lock#2 irq_context: 0 sk_lock-AF_INET6 batched_entropy_u16.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &table->hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &table->hash[i].lock &table->hash2[i].lock irq_context: softirq rcu_callback &ul->lock irq_context: 0 rtnl_mutex lapb_list_lock &c->lock irq_context: 0 rtnl_mutex &lapb->lock &c->lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 rtnl_mutex &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rename_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &p->alloc_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &c->lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock clock-AF_INET irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock clock-AF_INET irq_context: 0 sk_lock-AF_INET &h->lhash2[i].lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->xattr_sem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &mm->mmap_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 fs_reclaim irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#8 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 lock#4 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &mapping->i_private_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &memcg->move_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &xa->xa_lock#8 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &wb->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock_bh &base->lock irq_context: 0 rtnl_mutex rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rlock-AF_NETLINK irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 &tty->legacy_mutex &tty->ldisc_sem irq_context: 0 &tty->legacy_mutex &f->f_lock irq_context: 0 &tty->legacy_mutex &f->f_lock fasync_lock irq_context: 0 &tty->legacy_mutex &obj_hash[i].lock irq_context: 0 &tty->legacy_mutex pool_lock#2 irq_context: 0 &tty->legacy_mutex tasklist_lock irq_context: 0 &tty->legacy_mutex tasklist_lock &sighand->siglock irq_context: 0 &tty->legacy_mutex tasklist_lock &sighand->siglock &tty->ctrl.lock irq_context: 0 rcu_read_lock &tty->ctrl.lock irq_context: 0 &tty->ctrl.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &dentry->d_lock &wq#2 irq_context: 0 &port_lock_key irq_context: 0 &buf->lock irq_context: 0 &tty->ldisc_sem rcu_read_lock &tty->ctrl.lock irq_context: 0 &tty->ldisc_sem &port_lock_key irq_context: 0 &tty->ldisc_sem &port->lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &tty->ldisc_sem &tty->flow.lock irq_context: softirq rcu_callback rcu_read_lock rt6_exception_lock irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &tty->termios_rwsem &tty->read_wait irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock (work_completion)(&buf->work) irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &rq->__lock irq_context: 0 &tty->ldisc_sem &ldata->atomic_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &c->lock irq_context: 0 cb_lock genl_mutex &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &tx->clean_lock irq_context: softirq &tx->clean_lock &obj_hash[i].lock irq_context: softirq &tx->clean_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock &mm->page_table_lock irq_context: softirq &(&idev->mc_dad_work)->timer irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&idev->mc_dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &meta->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock kfence_freelist_lock irq_context: 0 rtnl_mutex &lapb->lock &n->list_lock irq_context: 0 rtnl_mutex &lapb->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex class irq_context: 0 rtnl_mutex (&tbl->proxy_timer) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &rdev->wiphy_work_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock &obj_hash[i].lock irq_context: softirq &(&idev->mc_ifc_work)->timer irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq rcu_callback &ul->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_INET6 fs_reclaim irq_context: 0 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock irq_context: 0 sk_lock-AF_INET6 once_lock irq_context: 0 sk_lock-AF_INET6 once_lock crngs.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &tbl->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &n->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#8 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq net/core/link_watch.c:31 irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock _xmit_ETHER#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock noop_qdisc.q.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &sch->q.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex class irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex (&tbl->proxy_timer) irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rlock-AF_NETLINK irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock krc.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 &pipe->mutex/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &net->ipv6.addrconf_hash_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock krc.lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &rdev->wiphy_work_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &dir->lock irq_context: 0 rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &dir->lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock krc.lock irq_context: 0 rtnl_mutex &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rt6_exception_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq &(&ifa->dad_work)->timer irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ul->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: softirq &tx->clean_lock &meta->lock irq_context: softirq &tx->clean_lock kfence_freelist_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 &net->packet.sklist_lock irq_context: 0 sk_lock-AF_PACKET irq_context: 0 sk_lock-AF_PACKET slock-AF_PACKET irq_context: 0 sk_lock-AF_PACKET &po->bind_lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock ptype_lock irq_context: 0 sk_lock-AF_PACKET &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &x->wait#2 irq_context: 0 sk_lock-AF_PACKET &rq->__lock irq_context: 0 sk_lock-AF_PACKET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock kfence_freelist_lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &dir->lock#2 irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock ptype_lock irq_context: 0 slock-AF_PACKET irq_context: 0 sk_lock-AF_PACKET &mm->mmap_lock irq_context: 0 sk_lock-AF_PACKET fs_reclaim irq_context: 0 sk_lock-AF_PACKET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_PACKET pool_lock#2 irq_context: 0 sk_lock-AF_PACKET free_vmap_area_lock irq_context: 0 sk_lock-AF_PACKET free_vmap_area_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET vmap_area_lock irq_context: 0 sk_lock-AF_PACKET &____s->seqcount irq_context: 0 sk_lock-AF_PACKET pcpu_alloc_mutex irq_context: 0 sk_lock-AF_PACKET pcpu_alloc_mutex pcpu_lock irq_context: 0 sk_lock-AF_PACKET &c->lock irq_context: 0 sk_lock-AF_PACKET batched_entropy_u32.lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock purge_vmap_area_lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock free_vmap_area_lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock free_vmap_area_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock free_vmap_area_lock pool_lock#2 irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock irq_context: 0 sk_lock-AF_PACKET vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &fp->aux->used_maps_mutex irq_context: 0 rlock-AF_PACKET irq_context: 0 wlock-AF_PACKET irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: softirq rcu_callback &x->wait#2 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET batched_entropy_u32.lock crngs.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->packet.sklist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock ptype_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &po->bind_lock &dir->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PACKET slock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 fanout_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 &x->wait#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_PACKET irq_context: 0 &sb->s_type->i_mutex_key#10 pcpu_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem irq_context: 0 &sb->s_type->i_mutex_key#10 key irq_context: 0 &sb->s_type->i_mutex_key#10 percpu_counters_lock irq_context: softirq &(&idev->mc_ifc_work)->timer rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq &(&tbl->managed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock krc.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &ul->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq &tx->clean_lock rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rlock-AF_PACKET irq_context: softirq rcu_read_lock rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock once_lock irq_context: softirq rcu_read_lock rcu_read_lock once_lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &sighand->siglock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock kfence_freelist_lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock pool_lock#2 irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx rcu_read_lock &sighand->signalfd_wqh irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx rcu_read_lock &ei->socket.wq.wait irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &c->lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &____s->seqcount irq_context: 0 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock key irq_context: 0 rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &sem->wait_lock irq_context: 0 &ep->mtx &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex batched_entropy_u32.lock crngs.lock irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock key irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock irq_context: 0 &mm->mmap_lock rcu_read_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: softirq drivers/base/dd.c:321 irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq drivers/base/dd.c:321 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work device_links_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work device_links_lock &k->list_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work device_links_lock &k->k_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work deferred_probe_mutex irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work &rq->__lock irq_context: 0 (wq_completion)events (deferred_probe_timeout_work).work deferred_probe_work irq_context: softirq &(&wb->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &type->s_umount_key#30 &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &wb->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &wb->list_lock &type->s_umount_key#40 &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 batched_entropy_u8.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 kfence_freelist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 lock#5 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 key#11 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 key#12 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock key#11 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &wb->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &wb->list_lock &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem remove_cache_srcu irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem lock#5 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &pcp->lock &zone->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &mapping->i_private_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &journal->j_revoke_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &pa->pa_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex &lg->lg_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#8 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#8 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &nvmeq->sq_lock irq_context: hardirq &ei->i_completed_io_lock irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq rcu_read_lock &memcg->move_lock irq_context: hardirq rcu_read_lock &xa->xa_lock#8 irq_context: hardirq rcu_read_lock &xa->xa_lock#8 &wb->work_lock irq_context: hardirq rcu_read_lock &xa->xa_lock#8 &wb->work_lock &obj_hash[i].lock irq_context: hardirq rcu_read_lock &xa->xa_lock#8 &wb->work_lock &base->lock irq_context: hardirq rcu_read_lock &xa->xa_lock#8 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: hardirq rcu_read_lock &xa->xa_lock#8 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)ext4-rsv-conversion irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &ei->i_completed_io_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &journal->j_state_lock &journal->j_wait_reserved irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_es_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_raw_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &journal->j_wait_updates irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &ext4__ioend_wq[i] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &ret->b_uptodate_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &memcg->move_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 &wb->work_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 &s->s_inode_wblist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &base->lock &obj_hash[i].lock irq_context: softirq &(&wb->bw_dwork)->timer irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->bw_dwork)->work) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->bw_dwork)->work) &wb->list_lock irq_context: softirq rcu_read_lock rcu_read_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock once_lock irq_context: softirq rcu_read_lock once_lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: softirq (&journal->j_commit_timer) &p->pi_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->work_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->work_lock &obj_hash[i].lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->work_lock &base->lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock key#13 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) irq_context: softirq (&lapb->t1timer) &lapb->lock irq_context: softirq (&lapb->t1timer) &lapb->lock batched_entropy_u8.lock irq_context: softirq (&lapb->t1timer) &lapb->lock kfence_freelist_lock irq_context: softirq (&lapb->t1timer) &lapb->lock pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh &meta->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh kfence_freelist_lock irq_context: softirq (&lapb->t1timer) &lapb->lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock &base->lock irq_context: softirq (&lapb->t1timer) &lapb->lock &base->lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &meta->lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 kfence_freelist_lock irq_context: 0 sk_lock-AF_INET6 &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&lapb->t1timer) &lapb->lock &c->lock irq_context: softirq (&dev->watchdog_timer) irq_context: softirq (&dom->period_timer) irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock irq_context: softirq (&dom->period_timer) key#14 irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock &obj_hash[i].lock irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock &base->lock irq_context: softirq (&dom->period_timer) &p->sequence irq_context: softirq (&dev->watchdog_timer) &dev->tx_global_lock &base->lock &obj_hash[i].lock irq_context: softirq (&dom->period_timer) &obj_hash[i].lock irq_context: softirq (&dom->period_timer) &base->lock irq_context: softirq (&dom->period_timer) &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET &n->list_lock irq_context: 0 sk_lock-AF_PACKET &n->list_lock &c->lock irq_context: 0 rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rlock-AF_NETLINK irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &dir->lock#2 irq_context: softirq &(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#5 &dentry->d_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 hostname_poll.wait.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &pl->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &pl->lock key#12 irq_context: 0 rcu_read_lock_bh quarantine_lock irq_context: softirq &(&tbl->gc_work)->timer irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &journal->j_list_lock &obj_hash[i].lock irq_context: 0 &journal->j_list_lock pool_lock#2 irq_context: 0 &xa->xa_lock#8 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock once_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock once_lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock batched_entropy_u32.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &n->lock irq_context: softirq rcu_read_lock &n->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &n->lock &base->lock irq_context: softirq rcu_read_lock &n->lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &n->lock &(&n->ha_lock)->lock irq_context: softirq rcu_read_lock &n->lock &(&n->ha_lock)->lock &____s->seqcount#8 irq_context: softirq rcu_read_lock rcu_read_lock &n->lock irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock lock#8 irq_context: softirq rcu_read_lock rcu_read_lock id_table_lock irq_context: softirq rcu_read_lock &n->lock irq_context: softirq rcu_read_lock &n->lock &____s->seqcount#8 irq_context: softirq rcu_read_lock nl_table_lock irq_context: softirq rcu_read_lock rlock-AF_NETLINK irq_context: softirq rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock &dir->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 batched_entropy_u16.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 (&req->rsk_timer) irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 &icsk->icsk_accept_queue.rskq_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &icsk->icsk_accept_queue.rskq_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET clock-AF_INET irq_context: 0 &vma->vm_lock->lock &lruvec->lru_lock irq_context: 0 &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &base->lock irq_context: 0 sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET fs_reclaim irq_context: 0 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET &____s->seqcount irq_context: 0 sk_lock-AF_INET &c->lock irq_context: 0 sk_lock-AF_INET pool_lock#2 irq_context: 0 sk_lock-AF_INET &mm->mmap_lock irq_context: 0 sk_lock-AF_INET tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &sd->defer_lock irq_context: softirq &sd->defer_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &base->lock &obj_hash[i].lock irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: softirq (&icsk->icsk_retransmit_timer) irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 irq_context: softirq (&icsk->icsk_delack_timer) irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 batched_entropy_u8.lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 kfence_freelist_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 pool_lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 &u->iolock &dir->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sk_lock-AF_INET &mm->mmap_lock fs_reclaim irq_context: 0 sk_lock-AF_INET &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET &mm->mmap_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: softirq rcu_callback uidhash_lock irq_context: softirq rcu_callback percpu_counters_lock irq_context: softirq rcu_callback ucounts_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET clock-AF_INET irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock tcp_metrics_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock tcp_metrics_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &hashinfo->ehash_locks[i] irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 &pipe->rd_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &pipe->rd_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &cfs_rq->removed.lock irq_context: 0 &u->iolock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock rcu_read_lock rcu_node_0 irq_context: 0 &u->iolock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &ul->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pipe->wr_wait irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &ei->xattr_sem &mapping->i_private_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock fs_reclaim irq_context: 0 &pipe->mutex/1 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &____s->seqcount irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &pipe->mutex/1 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 pool_lock#2 irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 &pipe->wr_wait &p->pi_lock irq_context: 0 &pipe->wr_wait &p->pi_lock &rq->__lock irq_context: 0 &pipe->wr_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET kfence_freelist_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu irq_context: 0 sk_lock-AF_INET remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &c->lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#8 &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#8 pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &xa->xa_lock#8 key#11 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &(ei->i_block_reservation_lock) key#15 irq_context: 0 sk_lock-AF_INET &n->list_lock irq_context: 0 sk_lock-AF_INET &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &meta->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 kfence_freelist_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: softirq &(&ipvs->defense_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem mmu_notifier_invalidate_range_start irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &meta->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem kfence_freelist_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 batched_entropy_u8.lock irq_context: 0 sb_writers#3 kfence_freelist_lock irq_context: 0 sb_writers#3 &meta->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 &c->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &ul->lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 &____s->seqcount irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock quarantine_lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &mm->mmap_lock quarantine_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 jbd2_handle &mapping->i_private_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex kfence_freelist_lock irq_context: 0 sk_lock-AF_PACKET pool_lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#8 &n->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#8 &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 kfence_freelist_lock irq_context: 0 rcu_read_lock &stopper->lock irq_context: 0 rcu_read_lock &stop_pi_lock irq_context: 0 rcu_read_lock &stop_pi_lock &rq->__lock irq_context: 0 rcu_read_lock &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &rq->__lock rcu_read_lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->wr_wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET &rq->__lock irq_context: 0 &vma->vm_lock->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &c->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &n->list_lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &mm->mmap_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stopper->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stop_pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stop_pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &stop_pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &____s->seqcount irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#6 &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 quarantine_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock irq_context: 0 &pipe->mutex/1 rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock batched_entropy_u8.lock irq_context: 0 &vma->vm_lock->lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 tomoyo_ss irq_context: 0 sb_writers#3 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 tomoyo_ss &c->lock irq_context: 0 sb_writers#3 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#3 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#3 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#3 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#3 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 batched_entropy_u8.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_list_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: softirq (&lapb->t1timer) &lapb->lock &____s->seqcount irq_context: softirq (&journal->j_commit_timer) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 lock#5 irq_context: softirq (&wq_watchdog_timer) &obj_hash[i].lock irq_context: softirq (&wq_watchdog_timer) &base->lock irq_context: softirq (&wq_watchdog_timer) &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock batched_entropy_u8.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET#2 batched_entropy_u16.lock crngs.lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&cache_cleaner)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#42 fs_reclaim irq_context: 0 kn->active#42 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#42 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#42 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#42 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#42 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 rcu_read_lock &sighand->siglock pool_lock#2 irq_context: 0 rcu_read_lock &sighand->siglock &p->pi_lock irq_context: 0 &futex_queues[i].lock irq_context: 0 &vma->vm_lock->lock rcu_node_0 irq_context: 0 rcu_read_lock &sighand->siglock &c->lock irq_context: 0 rcu_read_lock &sighand->siglock &____s->seqcount irq_context: 0 &ep->mtx &ep->lock &ep->wq irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &p->lock irq_context: 0 &f->f_pos_lock &p->lock fs_reclaim irq_context: 0 &f->f_pos_lock &p->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 clock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx sysctl_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rq->__lock irq_context: 0 &f->f_pos_lock sysctl_lock irq_context: 0 &f->f_pos_lock fs_reclaim irq_context: 0 &f->f_pos_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &zone->lock irq_context: 0 &f->f_pos_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &ep->mtx kn->active#5 fs_reclaim irq_context: 0 &ep->mtx kn->active#5 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ep->mtx kn->active#5 pool_lock#2 irq_context: 0 &ep->mtx kn->active#5 &on->poll irq_context: 0 &f->f_pos_lock &p->lock &c->lock irq_context: 0 &f->f_pos_lock &p->lock &of->mutex irq_context: 0 &f->f_pos_lock &p->lock &of->mutex kn->active#5 param_lock irq_context: 0 &ep->mtx rcu_read_lock &on->poll irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock fs_reclaim irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock &____s->seqcount irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &p->lock &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &p->lock pool_lock#2 irq_context: 0 &f->f_pos_lock &p->lock module_mutex irq_context: 0 sk_lock-AF_INET once_mutex irq_context: 0 sk_lock-AF_INET once_mutex crngs.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &hashinfo->ehash_locks[i] irq_context: 0 sk_lock-AF_INET batched_entropy_u32.lock irq_context: 0 sk_lock-AF_INET batched_entropy_u16.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &sk->sk_lock.wq irq_context: 0 slock-AF_INET#2 &sk->sk_lock.wq irq_context: 0 slock-AF_INET#2 &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_INET#2 &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: softirq slock-AF_INET#2 tk_core.seq.seqcount irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq slock-AF_INET#2 &obj_hash[i].lock irq_context: softirq slock-AF_INET#2 &base->lock irq_context: softirq slock-AF_INET#2 &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&dm_bufio_cleanup_old_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)dm_bufio_cache irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) dm_bufio_clients_lock irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) &base->lock irq_context: 0 (wq_completion)dm_bufio_cache (work_completion)(&(&dm_bufio_cleanup_old_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ep->mtx &pipe->wr_wait irq_context: 0 rcu_read_lock tasklist_lock irq_context: 0 &ep->mtx rcu_read_lock &pipe->wr_wait irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 sysctl_lock irq_context: 0 sb_writers#4 &p->pi_lock irq_context: 0 sb_writers#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &s->s_inode_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->xattr_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle inode_hash_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u32.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &xa->xa_lock#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &xa->xa_lock#8 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem lock#4 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle lock#4 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &ei->i_es_lock key#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &c->lock irq_context: 0 &type->s_umount_key#41/1 irq_context: 0 &type->s_umount_key#41/1 fs_reclaim irq_context: 0 &type->s_umount_key#41/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#41/1 pool_lock#2 irq_context: 0 &type->s_umount_key#41/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#41/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#41/1 shrinker_mutex irq_context: 0 &type->s_umount_key#41/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#41/1 sb_lock irq_context: 0 &type->s_umount_key#41/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 &type->s_umount_key#41/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#41/1 &____s->seqcount irq_context: 0 &type->s_umount_key#41/1 &sb->s_type->i_lock_key#30 irq_context: 0 &type->s_umount_key#41/1 &sb->s_type->i_lock_key#30 &dentry->d_lock irq_context: 0 &type->s_umount_key#41/1 crngs.lock irq_context: 0 &type->s_umount_key#41/1 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#41/1 &dentry->d_lock irq_context: 0 sb_writers#8 irq_context: 0 sb_writers#8 mount_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex pool_lock#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_iattr_rwsem iattr_mutex tk_core.seq.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock &wq#2 irq_context: 0 kn->active#43 fs_reclaim irq_context: 0 kn->active#43 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#43 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#43 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#43 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#43 pool_lock#2 irq_context: 0 sb_writers#8 fs_reclaim irq_context: 0 sb_writers#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &mm->mmap_lock irq_context: 0 sb_writers#8 &of->mutex irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex css_set_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock css_set_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &obj_hash[i].lock irq_context: 0 cgroup_mutex &pcp->lock &zone->lock irq_context: 0 cgroup_mutex cpu_hotplug_lock css_set_lock irq_context: 0 cgroup_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 cgroup_mutex css_set_lock cgroup_file_kn_lock irq_context: 0 &type->s_umount_key#42/1 irq_context: 0 &type->s_umount_key#42/1 fs_reclaim irq_context: 0 &type->s_umount_key#42/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#42/1 pool_lock#2 irq_context: 0 &type->s_umount_key#42/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#42/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#42/1 shrinker_mutex irq_context: 0 &type->s_umount_key#42/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#42/1 sb_lock irq_context: 0 &type->s_umount_key#42/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#42/1 &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#42/1 &sb->s_type->i_lock_key#31 &dentry->d_lock irq_context: 0 &type->s_umount_key#42/1 crngs.lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#42/1 &dentry->d_lock irq_context: 0 rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#43 irq_context: 0 &type->s_umount_key#43 &x->wait#23 irq_context: 0 &type->s_umount_key#43 shrinker_mutex irq_context: 0 &type->s_umount_key#43 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#43 percpu_ref_switch_lock irq_context: 0 &type->s_umount_key#43 percpu_ref_switch_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#43 percpu_ref_switch_lock pool_lock#2 irq_context: 0 &type->s_umount_key#43 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#43 rename_lock.seqcount irq_context: 0 &type->s_umount_key#43 &dentry->d_lock irq_context: 0 &type->s_umount_key#43 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#43 &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#43 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#43 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#43 inode_hash_lock irq_context: 0 &type->s_umount_key#43 inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 &type->s_umount_key#43 pool_lock#2 irq_context: 0 &type->s_umount_key#43 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#43 &dentry->d_lock/1 irq_context: 0 cgroup_mutex &n->list_lock irq_context: 0 cgroup_mutex &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#42/1 &c->lock irq_context: 0 &type->s_umount_key#42/1 &n->list_lock irq_context: 0 &type->s_umount_key#42/1 &n->list_lock &c->lock irq_context: 0 rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &type->s_umount_key#43 &obj_hash[i].lock pool_lock irq_context: 0 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 cgroup_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 cgroup_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 cgroup_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 cgroup_mutex &x->wait#2 irq_context: 0 cgroup_mutex &rq->__lock irq_context: 0 cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback percpu_ref_switch_waitq.lock irq_context: softirq rcu_callback rcu_read_lock &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_callback rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex cgroup_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 cgroup_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) percpu_ref_switch_lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex rcu_node_0 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex &rq->__lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex cgroup_rstat_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex cgroup_rstat_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex css_set_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) pool_lock#2 irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem &c->lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem &____s->seqcount irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) percpu_ref_switch_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &cgrp->pidlist_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) (wq_completion)cgroup_pidlist_destroy irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &wq->mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &wq->mutex &pool->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) (work_completion)(&cgrp->release_agent_work) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex css_set_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock css_set_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex css_set_lock &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex css_set_lock pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_rstat_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_rstat_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) pcpu_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) krc.lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &x->wait#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &rq->__lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_mutex cgroup_mutex.wait_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex fs_reclaim irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex pool_lock#2 irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &c->lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex.wait_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &p->pi_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_mutex batched_entropy_u8.lock irq_context: 0 cgroup_mutex kfence_freelist_lock irq_context: 0 sb_writers#9 irq_context: 0 sb_writers#9 mount_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tomoyo_ss tomoyo_policy_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 tk_core.seq.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_iattr_rwsem iattr_mutex tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_es_lock key#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 cgroup_mutex cpuset_mutex irq_context: 0 cgroup_mutex cpuset_mutex callback_lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#42/1 &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cgroup_mutex lock kernfs_idr_lock &c->lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 cgroup_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 cgroup_mutex &dom->lock irq_context: 0 cgroup_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&cgrp->bpf.release_work) cgroup_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) cgroup_mutex.wait_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&css->destroy_work) &p->pi_lock irq_context: softirq drivers/regulator/core.c:6335 irq_context: softirq drivers/regulator/core.c:6335 rcu_read_lock &pool->lock irq_context: softirq drivers/regulator/core.c:6335 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq drivers/regulator/core.c:6335 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq drivers/regulator/core.c:6335 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq drivers/regulator/core.c:6335 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (regulator_init_complete_work).work irq_context: 0 (wq_completion)events (regulator_init_complete_work).work &k->list_lock irq_context: 0 (wq_completion)events (regulator_init_complete_work).work &k->k_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpuset_mutex irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex cpuset_mutex callback_lock irq_context: 0 (wq_completion)cgroup_destroy (work_completion)(&(&css->destroy_rwork)->work) cgroup_mutex &dom->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 rename_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &dentry->d_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock &wq#2 irq_context: 0 kn->active#44 fs_reclaim irq_context: 0 kn->active#44 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#44 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#44 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#44 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 fs_reclaim irq_context: 0 sb_writers#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &mm->mmap_lock irq_context: 0 sb_writers#9 &of->mutex irq_context: 0 sb_writers#9 &obj_hash[i].lock irq_context: 0 kn->active#45 fs_reclaim irq_context: 0 kn->active#45 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#45 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#45 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#45 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &of->mutex kn->active#45 cpu_hotplug_lock irq_context: 0 sb_writers#9 &of->mutex kn->active#45 cpu_hotplug_lock cpuset_mutex irq_context: 0 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#10 irq_context: 0 sb_writers#10 &mm->mmap_lock irq_context: 0 sb_writers#10 &attr->mutex irq_context: 0 sb_writers#10 &attr->mutex &mm->mmap_lock irq_context: 0 &type->s_umount_key#44 irq_context: 0 &type->s_umount_key#44 sb_lock irq_context: 0 &type->s_umount_key#44 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#16 irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem mount_lock rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem mount_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#16 namespace_sem mount_lock pool_lock#2 irq_context: 0 &sb->s_type->i_lock_key#26 irq_context: 0 sb_writers#11 irq_context: 0 sb_writers#11 fs_reclaim irq_context: 0 sb_writers#11 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 pool_lock#2 irq_context: 0 sb_writers#11 &mm->mmap_lock irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 rename_lock.seqcount irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 fs_reclaim irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 pool_lock#2 irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &dentry->d_lock irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &dentry->d_lock &wq irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &sb->s_type->i_lock_key#26 irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &s->s_inode_list_lock irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 tk_core.seq.seqcount irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 &sb->s_type->i_lock_key#26 &dentry->d_lock irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 init_binfmt_misc.entries_lock irq_context: 0 sb_writers#11 &sb->s_type->i_mutex_key#16 rcu_read_lock &dentry->d_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &wpan_dev->association_lock irq_context: 0 rtnl_mutex dev_addr_sem irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &tn->lock irq_context: 0 rtnl_mutex dev_addr_sem &sdata->sec_mtx irq_context: 0 rtnl_mutex dev_addr_sem &sdata->sec_mtx &sec->lock irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex dev_addr_sem pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem nl_table_lock irq_context: 0 rtnl_mutex dev_addr_sem rlock-AF_NETLINK irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem nl_table_wait.lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock irq_context: 0 rtnl_mutex dev_addr_sem &pn->hash_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem input_pool.lock irq_context: 0 rtnl_mutex _xmit_IEEE802154 irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &xa->xa_lock#16 irq_context: 0 &sb->s_type->i_mutex_key#10 genl_sk_destructing_waitq.lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rdev->beacon_registrations_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rdev->mgmt_registrations_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &wdev->pmsr_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem reg_indoor_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 fs_reclaim irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &dentry->d_lock &wq#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem key#15 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem key#3 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &k->list_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &k->k_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &genl_data->genl_data_mutex irq_context: 0 (wq_completion)events (work_completion)(&w->w) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_lock nl_table_wait.lock irq_context: 0 nl_table_wait.lock &p->pi_lock irq_context: 0 nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock &xa->xa_lock#8 &c->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &base->lock irq_context: 0 &sig->cred_guard_mutex mapping.invalidate_lock rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 &wb->list_lock irq_context: 0 &sbi->s_writepages_rwsem irq_context: 0 &sbi->s_writepages_rwsem &xa->xa_lock#8 irq_context: 0 &sbi->s_writepages_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem lock#4 irq_context: 0 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 &sbi->s_writepages_rwsem lock#5 irq_context: 0 &sbi->s_writepages_rwsem &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem &pcp->lock &zone->lock irq_context: 0 &sbi->s_writepages_rwsem &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem &journal->j_state_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle lock#4 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle lock#5 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &memcg->move_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#8 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#8 &s->s_inode_wblist_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &journal->j_wait_updates irq_context: 0 &sbi->s_writepages_rwsem tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem &base->lock irq_context: 0 &sbi->s_writepages_rwsem &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &nvmeq->sq_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 &base->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 &wb->work_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 &wb->work_lock &base->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 jbd2_handle irq_context: 0 &journal->j_wait_commit irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle &ret->b_state_lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle key#3 irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle key#4 irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle &sbi->s_error_lock irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) jbd2_handle &journal->j_wait_updates irq_context: 0 (wq_completion)events (work_completion)(&sbi->s_sb_upd_work) &obj_hash[i].lock irq_context: 0 &journal->j_wait_done_commit &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#8 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 free_vmap_area_lock irq_context: 0 &sb->s_type->i_mutex_key#8 vmap_area_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 init_mm.page_table_lock irq_context: 0 &sb->s_type->i_mutex_key#8 pcpu_alloc_mutex irq_context: 0 &sb->s_type->i_mutex_key#8 pcpu_alloc_mutex pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#8 batched_entropy_u32.lock irq_context: 0 &sb->s_type->i_mutex_key#8 free_vmap_area_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 free_vmap_area_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 swap_cgroup_mutex &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 &sb->s_type->i_mutex_key#8 tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &q->requeue_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 &x->wait#26 irq_context: 0 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#8 &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock irq_context: hardirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock &rq->__lock irq_context: hardirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &fq->mq_flush_lock &x->wait#26 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 key irq_context: 0 &sb->s_type->i_mutex_key#8 pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#8 percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#8 (&timer.timer) irq_context: 0 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#8 &((cluster_info + ci)->lock)/1 irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex swap_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex swap_lock &p->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex percpu_ref_switch_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex swap_lock &p->lock#2 swap_avail_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex (console_sem).lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &sb->s_type->i_mutex_key#8 swapon_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &sb->s_type->i_mutex_key#8 proc_poll_wait.lock irq_context: 0 swap_slots_cache_enable_mutex irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-down irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex cpuhp_state-up irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &x->wait#6 irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &rq->__lock irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cpu_hotplug_lock cpuhp_state-up swap_slots_cache_mutex irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock irq_context: 0 swap_slots_cache_enable_mutex cpu_hotplug_lock cpuhp_state_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 swap_slots_cache_enable_mutex swap_lock irq_context: 0 &____s->seqcount#4 irq_context: 0 rcu_read_lock &sighand->siglock batched_entropy_u8.lock irq_context: 0 rcu_read_lock &sighand->siglock kfence_freelist_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock pool_lock irq_context: 0 rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 sb_writers#5 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#5 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sb->s_type->i_lock_key#22 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock irq_context: 0 sb_writers#5 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 sb_writers#5 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock irq_context: 0 sb_writers#5 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &ep->mtx &lock->wait_lock irq_context: 0 &ep->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx rcu_read_lock &rq->__lock irq_context: 0 &ep->mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock rcu_node_0 irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss tomoyo_policy_lock tomoyo_policy_lock.wait_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 &fsnotify_mark_srcu &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock sb_writers#5 mount_lock irq_context: 0 &mm->mmap_lock sb_writers#5 tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock sb_writers#5 &sb->s_type->i_lock_key irq_context: 0 &mm->mmap_lock sb_writers#5 &wb->list_lock irq_context: 0 &mm->mmap_lock sb_writers#5 &wb->list_lock &sb->s_type->i_lock_key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &newf->file_lock &newf->resize_wait irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 &ep->mtx rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock irq_context: 0 &kcov->lock irq_context: 0 &mm->mmap_lock &kcov->lock irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 &kcov->lock kcov_remote_lock irq_context: 0 &kcov->lock kcov_remote_lock pool_lock#2 irq_context: 0 pid_caches_mutex irq_context: 0 pid_caches_mutex slab_mutex irq_context: 0 pid_caches_mutex slab_mutex fs_reclaim irq_context: 0 pid_caches_mutex slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pid_caches_mutex slab_mutex pool_lock#2 irq_context: 0 pid_caches_mutex slab_mutex pcpu_alloc_mutex irq_context: 0 pid_caches_mutex slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pid_caches_mutex slab_mutex &root->kernfs_rwsem irq_context: 0 pid_caches_mutex slab_mutex &k->list_lock irq_context: 0 pid_caches_mutex slab_mutex lock irq_context: 0 pid_caches_mutex slab_mutex lock kernfs_idr_lock irq_context: 0 pid_caches_mutex slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pid_caches_mutex slab_mutex &c->lock irq_context: 0 &type->s_umount_key#45 irq_context: 0 &type->s_umount_key#45 sb_lock irq_context: 0 &type->s_umount_key#45 &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rcu_read_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rename_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rename_lock rename_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem rename_lock rename_lock.seqcount &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem mount_lock &obj_hash[i].lock irq_context: 0 bt_proto_lock &sk->sk_peer_lock irq_context: 0 bt_proto_lock hci_sk_list.lock irq_context: 0 misc_mtx &base->lock irq_context: 0 misc_mtx &base->lock &obj_hash[i].lock irq_context: 0 (work_completion)(&(&data->open_timeout)->work) irq_context: 0 &data->open_mutex irq_context: 0 &data->open_mutex fs_reclaim irq_context: 0 &data->open_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex pool_lock#2 irq_context: 0 &data->open_mutex &c->lock irq_context: 0 &data->open_mutex &pcp->lock &zone->lock irq_context: 0 &data->open_mutex &____s->seqcount irq_context: 0 &data->open_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex &obj_hash[i].lock pool_lock irq_context: 0 &data->open_mutex &x->wait#9 irq_context: 0 &data->open_mutex hci_index_ida.xa_lock irq_context: 0 &data->open_mutex pcpu_alloc_mutex irq_context: 0 &data->open_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &data->open_mutex cpu_hotplug_lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex wq_pool_mutex irq_context: 0 &data->open_mutex wq_pool_mutex &wq->mutex irq_context: 0 &data->open_mutex pin_fs_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &data->open_mutex &k->list_lock irq_context: 0 &data->open_mutex gdp_mutex irq_context: 0 &data->open_mutex gdp_mutex &k->list_lock irq_context: 0 &data->open_mutex gdp_mutex fs_reclaim irq_context: 0 &data->open_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex gdp_mutex pool_lock#2 irq_context: 0 &data->open_mutex gdp_mutex lock irq_context: 0 &data->open_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 &data->open_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &data->open_mutex lock irq_context: 0 &data->open_mutex lock kernfs_idr_lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &data->open_mutex bus_type_sem irq_context: 0 &data->open_mutex sysfs_symlink_target_lock irq_context: 0 &data->open_mutex &n->list_lock irq_context: 0 &data->open_mutex &n->list_lock &c->lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex &dev->power.lock irq_context: 0 &data->open_mutex dpm_list_mtx irq_context: 0 &data->open_mutex uevent_sock_mutex irq_context: 0 &data->open_mutex uevent_sock_mutex fs_reclaim irq_context: 0 &data->open_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &data->open_mutex uevent_sock_mutex &c->lock irq_context: 0 &data->open_mutex uevent_sock_mutex nl_table_lock irq_context: 0 &data->open_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &data->open_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex subsys mutex#74 irq_context: 0 &data->open_mutex subsys mutex#74 &k->k_lock irq_context: 0 &data->open_mutex &dev->devres_lock irq_context: 0 &data->open_mutex triggers_list_lock irq_context: 0 &data->open_mutex leds_list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex irq_context: 0 &data->open_mutex rfkill_global_mutex fs_reclaim irq_context: 0 &data->open_mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex rfkill_global_mutex pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex &k->list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex lock irq_context: 0 &data->open_mutex rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &data->open_mutex rfkill_global_mutex bus_type_sem irq_context: 0 &data->open_mutex rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &____s->seqcount irq_context: 0 &data->open_mutex rfkill_global_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 &data->open_mutex rfkill_global_mutex &dev->power.lock irq_context: 0 &data->open_mutex rfkill_global_mutex dpm_list_mtx irq_context: 0 &data->open_mutex rfkill_global_mutex &rfkill->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex nl_table_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex &k->k_lock irq_context: 0 &data->open_mutex rfkill_global_mutex subsys mutex#40 irq_context: 0 &data->open_mutex rfkill_global_mutex subsys mutex#40 &k->k_lock irq_context: 0 &data->open_mutex rfkill_global_mutex triggers_list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex leds_list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex.wait_lock irq_context: 0 &data->open_mutex &p->pi_lock irq_context: 0 &data->open_mutex &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex &rq->__lock irq_context: 0 &data->open_mutex &rfkill->lock irq_context: 0 &data->open_mutex hci_dev_list_lock irq_context: 0 &data->open_mutex tk_core.seq.seqcount irq_context: 0 &data->open_mutex hci_sk_list.lock irq_context: 0 &data->open_mutex (pm_chain_head).rwsem irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &data->open_mutex &list->lock#5 irq_context: 0 &data->open_mutex &data->read_wait irq_context: 0 &list->lock#5 irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start stock_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 irq_context: 0 tty_mutex &tty->legacy_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock console_srcu &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &obj_hash[i].lock pool_lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu rcu_node_0 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI sock_cookie_ida.xa_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &p->alloc_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI pool_lock#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &c->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &n->list_lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &n->list_lock &c->lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI tk_core.seq.seqcount irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_sk_list.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &obj_hash[i].lock irq_context: 0 slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 hci_dev_list_lock irq_context: 0 &data->read_wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 tty_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem quarantine_lock irq_context: 0 &type->lock_class/1 irq_context: 0 &mm->mmap_lock &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rcu_read_lock rcu_read_lock stock_lock irq_context: 0 &list->lock#7 irq_context: 0 rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work &rq->__lock irq_context: 0 key_types_sem &type->lock_class#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc17_nci_cmd_wq#14 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 rcu_node_0 irq_context: 0 sk_lock-AF_INET6 stock_lock irq_context: 0 sk_lock-AF_INET6 mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_INET6 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &sb->s_type->i_lock_key#8 irq_context: 0 sk_lock-AF_INET6 &dir->lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 pool_lock#2 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &dir->lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 fs_reclaim irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &c->lock irq_context: 0 sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 k-clock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 &xa->xa_lock#8 irq_context: 0 sk_lock-AF_INET6 &fsnotify_mark_srcu irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-clock-AF_INET6 irq_context: 0 &pool->lock &x->wait#10 irq_context: 0 &pool->lock &x->wait#10 &p->pi_lock irq_context: 0 &pool->lock &x->wait#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pool->lock &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 &pool->lock &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem &c->lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 &knet->mutex irq_context: 0 &mux->lock irq_context: 0 &mux->rx_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM slock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_KCM clock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 &mux->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (work_completion)(&kcm->tx_work) irq_context: 0 &sb->s_type->i_mutex_key#10 &mux->rx_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &mux->rx_lock rlock-AF_KCM irq_context: 0 &sb->s_type->i_mutex_key#10 &knet->mutex irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events key_gc_work sched_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)events key_gc_work sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock &base->lock irq_context: 0 rtnl_mutex dev_addr_sem stock_lock irq_context: 0 rtnl_mutex dev_addr_sem key irq_context: 0 rtnl_mutex dev_addr_sem pcpu_lock irq_context: 0 rtnl_mutex dev_addr_sem percpu_counters_lock irq_context: 0 rtnl_mutex dev_addr_sem pcpu_lock stock_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock &base->lock &obj_hash[i].lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &data->open_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc17_nci_rx_wq#14 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu irq_context: 0 &hdev->req_lock irq_context: 0 &hdev->req_lock pool_lock#2 irq_context: 0 &hdev->req_lock &pcp->lock &zone->lock irq_context: 0 &hdev->req_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &hdev->req_lock &____s->seqcount irq_context: 0 &hdev->req_lock &list->lock#6 irq_context: 0 &hdev->req_lock &list->lock#7 irq_context: 0 &hdev->req_lock &rq->__lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem &sighand->siglock irq_context: 0 cgroup_threadgroup_rwsem &sighand->siglock &rq->__lock irq_context: 0 &hdev->req_lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 krc.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &dir->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 stock_lock irq_context: 0 sk_lock-AF_INET6 &n->list_lock irq_context: 0 sk_lock-AF_INET6 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &n->list_lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &rq->__lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu rcu_read_lock &obj_hash[i].lock irq_context: 0 remove_cache_srcu rcu_read_lock pool_lock#2 irq_context: 0 remove_cache_srcu rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET6 &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#9 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 key#26 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 key#26 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 cb_lock genl_mutex nbd_index_mutex irq_context: 0 key_types_sem &type->lock_class#2 &obj_hash[i].lock pool_lock irq_context: 0 key_types_sem &type->lock_class#2 &rq->__lock irq_context: 0 &fsnotify_mark_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex stock_lock irq_context: 0 cb_lock genl_mutex key irq_context: 0 cb_lock genl_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex percpu_counters_lock irq_context: 0 cb_lock genl_mutex pcpu_lock stock_lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_sk_list.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI sock_cookie_ida.xa_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI hci_sk_list.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI clock-AF_BLUETOOTH irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_BLUETOOTH-BTPROTO_HCI irq_context: 0 &sb->s_type->i_mutex_key#10 hci_dev_list_lock irq_context: 0 namespace_sem mnt_id_ida.xa_lock pool_lock#2 irq_context: 0 &type->s_umount_key#21/1 &n->list_lock irq_context: 0 &type->s_umount_key#21/1 &n->list_lock &c->lock irq_context: 0 rcu_read_lock &undo_list->lock irq_context: 0 sb_writers#4 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &n->list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &____s->seqcount#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock quarantine_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) rcu_node_0 irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &n->list_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &ei->socket.wq.wait irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &n->list_lock &c->lock irq_context: 0 rtnl_mutex &nr_netdev_addr_lock_key irq_context: 0 tty_mutex &tty->legacy_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex listen_lock irq_context: 0 tty_mutex &tty->legacy_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET6 pool_lock#2 irq_context: 0 tty_mutex &tty->legacy_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex br_ioctl_mutex.wait_lock irq_context: 0 br_ioctl_mutex rtnl_mutex.wait_lock irq_context: 0 br_ioctl_mutex &p->pi_lock irq_context: 0 &u->iolock &base->lock irq_context: 0 &u->iolock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex &ndev->lock &____s->seqcount#2 irq_context: 0 rtnl_mutex dev_addr_sem &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem irq_context: 0 pernet_ops_rwsem stack_depot_init_mutex irq_context: 0 pernet_ops_rwsem crngs.lock irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock irq_context: 0 pernet_ops_rwsem fs_reclaim irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 pernet_ops_rwsem proc_subdir_lock irq_context: 0 pernet_ops_rwsem &c->lock irq_context: 0 pernet_ops_rwsem &____s->seqcount irq_context: 0 pernet_ops_rwsem sysctl_lock irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem &dir->lock irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-slock-AF_NETLINK irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem nl_table_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nl_table_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem nl_table_wait.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex irq_context: 0 pernet_ops_rwsem nl_table_lock irq_context: 0 pernet_ops_rwsem &net->rules_mod_lock irq_context: 0 pernet_ops_rwsem percpu_counters_lock irq_context: 0 pernet_ops_rwsem batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem k-slock-AF_INET/1 irq_context: 0 pernet_ops_rwsem &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem cache_list_lock irq_context: 0 pernet_ops_rwsem tk_core.seq.seqcount irq_context: 0 pernet_ops_rwsem &k->list_lock irq_context: 0 pernet_ops_rwsem lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem &n->list_lock irq_context: 0 pernet_ops_rwsem &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem uevent_sock_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 &data->open_mutex uevent_sock_mutex &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_es_lock key#6 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex.wait_lock irq_context: 0 pernet_ops_rwsem &p->pi_lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &sn->pipefs_sb_lock irq_context: 0 pernet_ops_rwsem &s->s_inode_list_lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_ecache_mutex irq_context: 0 pernet_ops_rwsem nf_log_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &c->lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem &base->lock irq_context: 0 pernet_ops_rwsem &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem &hashinfo->lock#2 irq_context: 0 pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock irq_context: 0 pernet_ops_rwsem proc_inum_ida.xa_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem &this->receive_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &x->wait#9 irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex bus_type_sem irq_context: 0 pernet_ops_rwsem rtnl_mutex sysfs_symlink_target_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem rtnl_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dpm_list_mtx irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &dir->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex dev_base_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex input_pool.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &tbl->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex sysctl_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fsnotify_mark_srcu &obj_hash[i].lock irq_context: 0 &xt[i].mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &tbl->lock &n->lock irq_context: 0 rtnl_mutex &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 rtnl_mutex &tbl->lock nl_table_lock irq_context: 0 rtnl_mutex &tbl->lock nl_table_wait.lock irq_context: 0 rtnl_mutex &tbl->lock rcu_read_lock lock#8 irq_context: 0 rtnl_mutex &tbl->lock rcu_read_lock id_table_lock irq_context: 0 rtnl_mutex &tbl->lock &dir->lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 pernet_ops_rwsem rdma_nets.xa_lock irq_context: 0 pernet_ops_rwsem devices_rwsem irq_context: 0 pernet_ops_rwsem hwsim_netgroup_ida.xa_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex stack_depot_init_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 rtnl_mutex &tbl->lock krc.lock irq_context: 0 &xt[i].mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 &xt[i].mutex &meta->lock irq_context: 0 rtnl_mutex dev_hotplug_mutex &rq->__lock irq_context: 0 sb_writers#3 &fsnotify_mark_srcu &rq->__lock irq_context: 0 sb_writers#3 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &net->ipv6.addrconf_hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &rq->__lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem remove_cache_srcu pool_lock#2 irq_context: 0 misc_mtx fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex failover_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem &tn->lock &rq->__lock irq_context: 0 &type->s_umount_key#50 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &____s->seqcount irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 (kmod_concurrent_max).lock irq_context: 0 sk_lock-AF_INET6 &x->wait#17 irq_context: 0 &tsk->futex_exit_mutex &rq->__lock cpu_asid_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &meta->lock irq_context: 0 &hdev->req_lock &hdev->lock rcu_read_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_INET6 running_helpers_waitq.lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock console_owner irq_context: 0 sk_lock-AF_INET6 &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock key irq_context: 0 rcu_read_lock &vma->vm_lock->lock key irq_context: 0 rcu_read_lock &vma->vm_lock->lock pcpu_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock percpu_counters_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &u->bindlock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &u->bindlock &net->unx.table.locks[i] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock (console_sem).lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &u->bindlock &net->unx.table.locks[i] &net->unx.table.locks[i]/1 &dentry->d_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock sb_writers#5 &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &cfs_rq->removed.lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex sched_map-wait-type-override &pool->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &u->bindlock &net->unx.table.locks[i]/1 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &u->bindlock &bsd_socket_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &n->list_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem batched_entropy_u8.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sta->lock irq_context: 0 sb_internal &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &meta->lock irq_context: 0 &mm->mmap_lock &c->lock batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock &c->lock kfence_freelist_lock irq_context: 0 namespace_sem pcpu_alloc_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 kfence_freelist_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &dentry->d_lock sysctl_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock sysctl_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 &data->open_mutex remove_cache_srcu irq_context: 0 &data->open_mutex remove_cache_srcu quarantine_lock irq_context: 0 &data->open_mutex remove_cache_srcu &c->lock irq_context: 0 &data->open_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 &data->open_mutex remove_cache_srcu &obj_hash[i].lock irq_context: softirq (&pool->idle_timer) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem stock_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->work_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &type->lock_class keyring_serialise_link_lock &rq->__lock irq_context: 0 &data->open_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &xa->xa_lock#8 &n->list_lock irq_context: 0 &xa->xa_lock#8 &n->list_lock &c->lock irq_context: 0 &data->open_mutex remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &pcp->lock &zone->lock irq_context: 0 cgroup_threadgroup_rwsem stock_lock irq_context: 0 cgroup_threadgroup_rwsem pcpu_lock stock_lock irq_context: 0 sb_writers#12 irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex gdp_mutex.wait_lock irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &____s->seqcount#2 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &rq->__lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &sem->wait_lock irq_context: 0 &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &c->lock irq_context: 0 &hdev->req_lock &hdev->req_wait_q irq_context: 0 &hdev->req_lock &base->lock irq_context: 0 &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex.wait_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 &c->lock irq_context: 0 pernet_ops_rwsem rdma_nets.xa_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &c->lock irq_context: 0 &hdev->req_lock (&timer.timer) irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex irq_context: 0 nfc_devlist_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &nf_conntrack_locks[i] irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount#2 irq_context: 0 rtnl_mutex remove_cache_srcu &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 pcpu_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 pcpu_lock stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &rq->__lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &type->lock_class keyring_serialise_link_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dir->lock#2 &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex device_links_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 kn->active#51 &n->list_lock irq_context: 0 kn->active#51 &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &root->kernfs_rwsem &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex &____s->seqcount#2 irq_context: 0 misc_mtx nfc_devlist_mutex &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 rcu_node_0 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_node_0 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &n->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex (console_sem).lock irq_context: 0 &root->deactivate_waitq irq_context: 0 &root->deactivate_waitq &p->pi_lock irq_context: 0 &root->deactivate_waitq &p->pi_lock &rq->__lock irq_context: 0 &root->deactivate_waitq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 &data->open_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &data->open_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &data->open_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &n->list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &n->list_lock &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &rq->__lock irq_context: 0 br_ioctl_mutex &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss batched_entropy_u8.lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.changes_list_lock &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.changes_list_lock kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex dpm_list_mtx &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex stock_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#46/1 &rq->__lock irq_context: 0 rtnl_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock console_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock console_owner_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex.wait_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex key irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex pcpu_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex percpu_counters_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex pcpu_lock stock_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock krc.lock &base->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 hrtimer_bases.lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_owner_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 batched_entropy_u8.lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 kfence_freelist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_owner irq_context: 0 rtnl_mutex devnet_rename_sem rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex &wq->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events free_ipc_work pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER &____s->seqcount#2 irq_context: 0 &pool->lock &x->wait#10 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem &____s->seqcount irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 tty_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER &pcp->lock &zone->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex devnet_rename_sem &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 &wb->work_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex devnet_rename_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &data->open_mutex rfkill_global_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#2 stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &cfs_rq->removed.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex console_owner_lock irq_context: 0 cb_lock genl_mutex console_owner irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx rcu_read_lock &pool->lock (worker)->lock &p->pi_lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &____s->seqcount#2 irq_context: 0 sb_writers#3 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rnp->exp_wq[0] irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock key irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#3 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex &mm->mmap_lock stock_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 crypto_alg_sem &rq->__lock irq_context: 0 &hdev->lock quarantine_lock irq_context: 0 crypto_alg_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 oom_adj_mutex &rcu_state.expedited_wq irq_context: 0 &dir->lock#2 kfence_freelist_lock irq_context: softirq (&peer->timer_persistent_keepalive) batched_entropy_u8.lock irq_context: softirq (&peer->timer_persistent_keepalive) kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 oom_adj_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &base->lock irq_context: 0 sb_writers#4 oom_adj_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock quarantine_lock irq_context: 0 sb_writers#4 oom_adj_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock kfence_freelist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &meta->lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex device_links_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex sysctl_lock krc.lock &base->lock irq_context: 0 rtnl_mutex sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events key_gc_work sched_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock stock_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &bsd_socket_locks[i] irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rcu_state.expedited_wq irq_context: 0 rtnl_mutex devnet_rename_sem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock irq_context: 0 rtnl_mutex devnet_rename_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock kfence_freelist_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &wb->work_lock &obj_hash[i].lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock &rq->__lock cpu_asid_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->work_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock nsim_bus_dev_list_lock.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex subsys mutex#39 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &base->lock irq_context: 0 callback_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &u->iolock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &disk->open_mutex inode_hash_lock irq_context: 0 &disk->open_mutex inode_hash_lock &sb->s_type->i_lock_key#3 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 rcu_read_lock &sighand->siglock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 key#9 irq_context: 0 &bdev->bd_holder_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex _xmit_IPGRE &n->list_lock irq_context: 0 rtnl_mutex _xmit_IPGRE &n->list_lock &c->lock irq_context: 0 &q->mq_freeze_lock percpu_ref_switch_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock &cfs_rq->removed.lock irq_context: 0 &q->mq_freeze_lock percpu_ref_switch_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ul->lock#2 irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem wq_pool_mutex irq_context: 0 pernet_ops_rwsem wq_pool_mutex &wq->mutex irq_context: 0 pernet_ops_rwsem pcpu_lock irq_context: 0 pernet_ops_rwsem &list->lock#4 irq_context: 0 pernet_ops_rwsem &dir->lock#2 irq_context: 0 pernet_ops_rwsem ptype_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rhashtable_bucket irq_context: 0 pernet_ops_rwsem k-clock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-slock-AF_TIPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-slock-AF_RXRPC irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex crngs.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &sb->s_type->i_lock_key#8 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &dir->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-slock-AF_INET6 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex cpu_hotplug_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex kthread_create_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &x->wait#21 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &local->services_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC pool_lock#2 irq_context: 0 pernet_ops_rwsem &rxnet->conn_lock irq_context: 0 pernet_ops_rwsem &call->waitq irq_context: 0 pernet_ops_rwsem &rx->call_lock irq_context: 0 pernet_ops_rwsem &rxnet->call_lock irq_context: 0 pernet_ops_rwsem net_rwsem irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock sysctl_lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &____s->seqcount#10 irq_context: 0 sb_writers#4 &(&net->ipv4.ping_group_range.lock)->lock irq_context: 0 sb_writers#4 &(&net->ipv4.ping_group_range.lock)->lock &____s->seqcount#10 irq_context: 0 misc_mtx &dir->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex &r->consumer_lock irq_context: 0 rtnl_mutex &r->consumer_lock &r->producer_lock irq_context: 0 rtnl_mutex failover_lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex &mm->mmap_lock irq_context: 0 rtnl_mutex &n->lock irq_context: 0 rtnl_mutex &n->lock &(&n->ha_lock)->lock irq_context: 0 rtnl_mutex &n->lock &(&n->ha_lock)->lock &____s->seqcount#8 irq_context: 0 rtnl_mutex &tbl->lock &n->lock irq_context: 0 rtnl_mutex rcu_read_lock lock#8 irq_context: 0 rtnl_mutex rcu_read_lock id_table_lock irq_context: 0 rtnl_mutex &n->lock irq_context: 0 rtnl_mutex &n->lock &____s->seqcount#8 irq_context: 0 rtnl_mutex &tbl->lock &c->lock irq_context: 0 rtnl_mutex &tbl->lock pool_lock#2 irq_context: 0 rtnl_mutex &tbl->lock batched_entropy_u32.lock irq_context: 0 rtnl_mutex &tbl->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex dev_addr_sem &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock nl_table_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock nl_table_wait.lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock rcu_read_lock lock#8 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock rcu_read_lock id_table_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &dir->lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &dir->lock#2 irq_context: 0 rtnl_mutex &ndev->lock pcpu_lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 rtnl_mutex &ndev->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock &c->lock irq_context: 0 rtnl_mutex &ndev->lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &ndev->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex pcpu_lock irq_context: 0 rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key pool_lock#2 irq_context: 0 rtnl_mutex (switchdev_blocking_notif_chain).rwsem irq_context: 0 rtnl_mutex &br->hash_lock irq_context: 0 rtnl_mutex &br->hash_lock &____s->seqcount irq_context: 0 rtnl_mutex &br->hash_lock &c->lock irq_context: 0 rtnl_mutex &br->hash_lock pool_lock#2 irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex &br->hash_lock nl_table_lock irq_context: 0 rtnl_mutex &br->hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->hash_lock nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex nf_hook_mutex irq_context: 0 rtnl_mutex nf_hook_mutex fs_reclaim irq_context: 0 rtnl_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex nf_hook_mutex pool_lock#2 irq_context: 0 rtnl_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 rtnl_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 rtnl_mutex j1939_netdev_lock irq_context: 0 rtnl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex key irq_context: 0 rtnl_mutex percpu_counters_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &n->list_lock &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem nl_table_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &c->lock irq_context: 0 rtnl_mutex net_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem batched_entropy_u32.lock crngs.lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock irq_context: 0 rtnl_mutex &ndev->lock &n->list_lock irq_context: 0 rtnl_mutex &ndev->lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock pool_lock#2 irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &c->lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.container_list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key pool_lock#2 irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock pool_lock#2 irq_context: 0 rtnl_mutex key#16 irq_context: 0 rtnl_mutex &bat_priv->tt.changes_list_lock irq_context: 0 rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock batched_entropy_u32.lock crngs.lock irq_context: softirq &(&bat_priv->nc.work)->timer irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) key#17 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) key#18 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_node_0 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex &idev->mc_lock kfence_freelist_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 rtnl_mutex kernfs_idr_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex noop_qdisc.q.lock irq_context: 0 rtnl_mutex tk_core.seq.seqcount irq_context: 0 rtnl_mutex &wq->mutex irq_context: 0 rtnl_mutex &wq->mutex &pool->lock irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 rtnl_mutex init_lock irq_context: 0 rtnl_mutex init_lock slab_mutex irq_context: 0 rtnl_mutex init_lock slab_mutex fs_reclaim irq_context: 0 rtnl_mutex init_lock slab_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex init_lock slab_mutex pool_lock#2 irq_context: 0 rtnl_mutex init_lock slab_mutex &c->lock irq_context: 0 rtnl_mutex init_lock slab_mutex &____s->seqcount irq_context: 0 rtnl_mutex init_lock slab_mutex pcpu_alloc_mutex irq_context: 0 rtnl_mutex init_lock slab_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 rtnl_mutex init_lock slab_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex init_lock slab_mutex &k->list_lock irq_context: 0 rtnl_mutex init_lock slab_mutex lock irq_context: 0 rtnl_mutex init_lock slab_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex init_lock slab_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex init_lock slab_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex init_lock fs_reclaim irq_context: 0 rtnl_mutex init_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex init_lock &zone->lock irq_context: 0 rtnl_mutex init_lock &____s->seqcount irq_context: 0 rtnl_mutex init_lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex init_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex init_lock &base->lock irq_context: 0 rtnl_mutex init_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex init_lock crngs.lock irq_context: 0 rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex nf_hook_mutex &c->lock irq_context: 0 rtnl_mutex nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex uevent_sock_mutex quarantine_lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key &____s->seqcount irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex deferred_lock irq_context: 0 (wq_completion)events deferred_process_work irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock _xmit_ETHER irq_context: 0 rtnl_mutex &br->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->lock &br->hash_lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock nl_table_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock pool_lock#2 irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex &pn->hash_lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex deferred_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex (switchdev_blocking_notif_chain).rwsem irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex lweventlist_lock &c->lock irq_context: 0 rtnl_mutex lweventlist_lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &r->producer_lock irq_context: 0 rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 sk_lock-AF_INET6 &sighand->siglock irq_context: 0 sk_lock-AF_INET6 &sighand->siglock stock_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &base->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rcu_state.expedited_wq irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 rtnl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: softirq &(&slave->notify_work)->timer irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex net_rwsem &cfs_rq->removed.lock irq_context: 0 rtnl_mutex net_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex net_rwsem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 cb_lock &cfs_rq->removed.lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex hrtimer_bases.lock irq_context: 0 rtnl_mutex hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: softirq (&ndev->rs_timer) batched_entropy_u8.lock irq_context: softirq (&ndev->rs_timer) kfence_freelist_lock irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock kfence_freelist_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock cpu_asid_lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &rq->__lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex net_rwsem key irq_context: 0 rtnl_mutex net_rwsem pcpu_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &cfs_rq->removed.lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&timer) rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: softirq (&timer) rcu_read_lock &pcp->lock &zone->lock irq_context: softirq (&timer) &txwq &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&q->timeout) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 remove_cache_srcu irq_context: 0 sb_writers &type->i_mutex_dir_key#2 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->lock_class keyring_serialise_link_lock &c->lock irq_context: 0 &type->lock_class keyring_serialise_link_lock root_key_user.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->lock_class keyring_serialise_link_lock &n->list_lock irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->xfrm.xfrm_state_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &bond->stats_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pmc->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&mp->timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&br->mcast_gc_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 loop_validate_mutex loop_validate_mutex.wait_lock irq_context: 0 sb_writers#7 remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_LOOPBACK irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &base->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &table->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &list->lock#14 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 key_types_sem fs_reclaim irq_context: 0 key_types_sem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 key_types_sem &type->lock_class root_key_user.lock irq_context: 0 proto_tab_lock &c->lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle bit_wait_table + i irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events reg_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events reg_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex stock_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex key irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex pcpu_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex percpu_counters_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex pcpu_lock stock_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex pool_lock#2 irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 cb_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock &blkcg->lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &disk->open_mutex &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &lock->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ul->lock#2 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ul->lock#2 &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ul->lock#2 &c->lock irq_context: 0 rtnl_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex crngs.lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex pool_lock#2 irq_context: 0 rtnl_mutex ptype_lock irq_context: 0 syslog_lock &rq->__lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 slock-AF_AX25 irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &c->lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &____s->seqcount irq_context: 0 rtnl_mutex batched_entropy_u8.lock irq_context: 0 rtnl_mutex kfence_freelist_lock irq_context: softirq &(&bat_priv->mcast.work)->timer irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock key#16 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &bat_priv->tt.changes_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &bat_priv->tvlv.container_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_NONE irq_context: 0 rtnl_mutex lock#9 irq_context: 0 rtnl_mutex _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER &____s->seqcount irq_context: 0 sb_writers#6 &cfs_rq->removed.lock irq_context: 0 sb_writers#6 &obj_hash[i].lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock lock kernfs_idr_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 subsys mutex#37 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 clock-AF_AX25 irq_context: 0 tty_mutex &tty->legacy_mutex console_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM slock-AF_NETROM irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->lock_class keyring_serialise_link_lock &n->list_lock &c->lock irq_context: 0 &type->lock_class keyring_serialise_link_lock remove_cache_srcu irq_context: 0 sb_writers#4 &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_CAIF &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex devnet_rename_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->lock_class keyring_serialise_link_lock remove_cache_srcu quarantine_lock irq_context: 0 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 &root->kernfs_rwsem &____s->seqcount irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (inetaddr_chain).rwsem (inetaddr_chain).rwsem.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (inetaddr_chain).rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (inetaddr_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem.wait_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex bpf_devs_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu irq_context: 0 &q->sysfs_dir_lock rcu_node_0 irq_context: 0 rtnl_mutex &tb->tb6_lock &n->list_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#11 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 sb_writers#12 mount_lock irq_context: 0 &xt[i].mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock rcu_node_0 irq_context: 0 nfc_devlist_mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 pool_lock irq_context: 0 rtnl_mutex (inetaddr_validator_chain).rwsem &cfs_rq->removed.lock irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) crngs.lock base_crng.lock irq_context: 0 (crypto_chain).rwsem &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_validator_chain).rwsem &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex bus_type_sem &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_validator_chain).rwsem pool_lock#2 irq_context: 0 (crypto_chain).rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &q->mq_freeze_wq irq_context: 0 key_types_sem &type->lock_class key_construction_mutex root_key_user.lock irq_context: 0 key_types_sem &type->lock_class key_construction_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock pool_lock irq_context: 0 (&bdi->laptop_mode_wb_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex &hsr->list_lock irq_context: 0 rtnl_mutex pin_fs_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 (console_sem).lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex rcu_read_lock mount_lock irq_context: 0 rtnl_mutex rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex mount_lock irq_context: 0 rtnl_mutex mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) pool_lock#2 irq_context: softirq &(&bat_priv->orig_work)->timer irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq|softirq allocation_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) key#19 irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work irq_context: 0 (wq_completion)events_power_efficient (gc_work).work tk_core.seq.seqcount irq_context: 0 (wq_completion)events_power_efficient (gc_work).work "ratelimiter_table_lock" irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &base->lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 rtnl_mutex gdp_mutex lock irq_context: 0 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 rtnl_mutex &k->k_lock irq_context: 0 key_types_sem &type->lock_class key_construction_mutex &rq->__lock irq_context: 0 key_types_sem &type->lock_class key_construction_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem &type->lock_class &c->lock irq_context: 0 (wq_completion)events key_gc_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events key_gc_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &meta->lock irq_context: 0 key_types_sem &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 rtnl_mutex &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6 &sighand->siglock &____s->seqcount irq_context: softirq rcu_read_lock &br->hash_lock &base->lock irq_context: softirq rcu_read_lock &br->hash_lock &base->lock &obj_hash[i].lock irq_context: 0 key_types_sem &type->lock_class &rq->__lock irq_context: 0 key_types_sem &type->lock_class &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->lock_class keyring_serialise_link_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events key_gc_work pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wg->socket_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &x->wait#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rq->__lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock &base->lock irq_context: 0 sb_writers#7 fs_reclaim &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock init_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 quarantine_lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem quarantine_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock quarantine_lock irq_context: 0 br_ioctl_mutex rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &sighand->siglock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock stock_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem stock_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem key irq_context: 0 &dev->mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 fs_reclaim rcu_node_0 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &journal->j_list_lock key#13 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock kfence_freelist_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &type->lock_class keyring_serialise_link_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &meta->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &base->lock &obj_hash[i].lock irq_context: softirq &(&slave->notify_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key pool_lock#2 irq_context: softirq (&app->join_timer) irq_context: softirq (&app->join_timer) &app->lock irq_context: softirq (&app->join_timer) &list->lock#10 irq_context: softirq (&app->join_timer) batched_entropy_u32.lock irq_context: softirq (&app->join_timer) &obj_hash[i].lock irq_context: softirq (&app->join_timer) &base->lock irq_context: softirq (&app->join_timer) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &____s->seqcount irq_context: softirq (&app->join_timer)#2 irq_context: softirq (&app->join_timer)#2 &app->lock#2 irq_context: softirq (&app->join_timer)#2 &list->lock#11 irq_context: softirq (&app->join_timer)#2 &app->lock#2 batched_entropy_u32.lock irq_context: softirq (&app->join_timer)#2 &app->lock#2 &obj_hash[i].lock irq_context: softirq (&app->join_timer)#2 &app->lock#2 &base->lock irq_context: softirq (&app->join_timer)#2 &app->lock#2 &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 key_types_sem &type->lock_class &n->list_lock irq_context: 0 key_types_sem &type->lock_class &n->list_lock &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 pool_lock#2 irq_context: 0 rtnl_mutex &xa->xa_lock#18 irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) pool_lock#2 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#3/1 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &c->lock irq_context: 0 rtnl_mutex nf_hook_mutex remove_cache_srcu irq_context: 0 rtnl_mutex nf_hook_mutex remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex nf_hook_mutex remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex nf_hook_mutex remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex nf_hook_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex nf_hook_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex nf_hook_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &____s->seqcount irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock &tap_major->minor_lock irq_context: 0 rtnl_mutex rcu_read_lock &tap_major->minor_lock pool_lock#2 irq_context: 0 rtnl_mutex req_lock irq_context: 0 rtnl_mutex &x->wait#11 irq_context: 0 rtnl_mutex subsys mutex#75 irq_context: 0 rtnl_mutex subsys mutex#75 &k->k_lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 &____s->seqcount irq_context: 0 rtnl_mutex &xa->xa_lock#4 pool_lock#2 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &sem->wait_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &p->pi_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#46 fs_reclaim irq_context: 0 kn->active#46 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 kn->active#47 fs_reclaim irq_context: 0 kn->active#47 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock nsim_bus_dev_ids.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock dpm_list_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex device_links_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex fwnode_link_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex device_links_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &dev->devres_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex pinctrl_list_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex pinctrl_maps_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex pinctrl_list_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex pinctrl_list_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &____s->seqcount#2 irq_context: 0 fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 fill_pool_map-wait-type-override &rcu_state.expedited_wq irq_context: 0 key#22 irq_context: 0 &fsnotify_mark_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex.wait_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &sig->cred_guard_mutex &fs->lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &mm->mmap_lock &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock irq_context: 0 fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx pool_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &n->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 rtnl_mutex &tb->tb6_lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex &tb->tb6_lock kfence_freelist_lock irq_context: 0 rtnl_mutex &tb->tb6_lock &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss key irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &c->lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 kn->active#46 &c->lock irq_context: 0 kn->active#46 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock nsim_bus_dev_list_lock.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss percpu_counters_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rnp->exp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_event_queue_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex.wait_lock irq_context: 0 sk_lock-AF_PACKET &p->pi_lock irq_context: 0 sk_lock-AF_PACKET &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &c->lock irq_context: 0 rtnl_mutex gdp_mutex &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &xa->xa_lock#4 &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex bus_type_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &n->list_lock irq_context: 0 &fsnotify_mark_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#4 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &mapping->i_private_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 &xt[i].mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock quarantine_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock key irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock pcpu_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock percpu_counters_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock &meta->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 sb_writers#7 &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 fs_reclaim &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &batadv_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_INET rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &mapping->i_private_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &c->lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &fsnotify_mark_srcu irq_context: 0 &type->i_mutex_dir_key/1 &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key/1 &sbinfo->stat_lock irq_context: 0 &type->i_mutex_dir_key/1 &xa->xa_lock#8 irq_context: 0 &type->i_mutex_dir_key/1 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->i_mutex_dir_key/1 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &meta->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 stock_lock irq_context: 0 sb_writers#5 key irq_context: 0 sb_writers#5 pcpu_lock irq_context: 0 sb_writers#5 percpu_counters_lock irq_context: 0 sb_writers#5 pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: hardirq rcu_read_lock &xa->xa_lock#8 &pl->lock irq_context: hardirq rcu_read_lock &xa->xa_lock#8 &pl->lock key#12 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 loop_validate_mutex.wait_lock irq_context: 0 proto_tab_lock &n->list_lock irq_context: 0 proto_tab_lock &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu irq_context: 0 &sig->cred_guard_mutex tomoyo_ss batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss kfence_freelist_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &meta->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex deferred_probe_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock (&timer.timer) irq_context: 0 rtnl_mutex uevent_sock_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex probe_waitqueue.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock subsys mutex#76 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock pool_lock#2 irq_context: 0 key_types_sem &type->lock_class &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &type->lock_class keyring_serialise_link_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events key_gc_work &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &hdev->req_lock (work_completion)(&hdev->tx_work) &rq->__lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 _xmit_ETHER &n->list_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 &wb->work_lock &base->lock irq_context: 0 sb_writers#4 oom_adj_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#4 oom_adj_mutex &obj_hash[i].lock irq_context: 0 sb_writers#4 oom_adj_mutex pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &base->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock key irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &meta->lock irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &data->open_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 &data->open_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#790 irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &____s->seqcount irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock pcpu_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_CAIF rcu_read_lock &rq->__lock irq_context: 0 key_types_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex devnet_rename_sem &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &obj_hash[i].lock irq_context: softirq (&app->join_timer)#2 &app->lock#2 batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &rq->__lock irq_context: 0 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 rtnl_mutex &devlink_port->type_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock irq_context: 0 &hdev->lock remove_cache_srcu &rq->__lock irq_context: 0 &hdev->lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem &rq->__lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 &c->lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &base->lock &obj_hash[i].lock irq_context: softirq &(&hwstats->traffic_dw)->timer irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &hwstats->hwsdev_list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_lock_key#23 bit_wait_table + i irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAIF stock_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle bit_wait_table + i irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock (work_completion)(&hdev->tx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAIF &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_CAIF rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock sb_writers#3 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock percpu_counters_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock kfence_freelist_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 misc_mtx &dev->mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock pcpu_lock stock_lock irq_context: 0 ebt_mutex stock_lock irq_context: 0 ebt_mutex key irq_context: 0 ebt_mutex pcpu_lock irq_context: 0 ebt_mutex percpu_counters_lock irq_context: 0 ebt_mutex pcpu_lock stock_lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 &base->lock irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &rq->__lock cpu_asid_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 br_ioctl_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (work_completion)(&local->timeout_work) &rq->__lock irq_context: 0 (work_completion)(&local->timeout_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem nl_table_wait.lock &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem quarantine_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 ebt_mutex &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &base->lock irq_context: 0 rtnl_mutex devnet_rename_sem &n->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem &n->list_lock &c->lock irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq &(&hdev->cmd_timer)->timer irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &x->wait#27 &p->pi_lock irq_context: 0 &x->wait#27 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &x->wait#27 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#27 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#8 &n->list_lock irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &c->lock irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#8 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 nf_sockopt_mutex rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pcpu_alloc_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock pcpu_alloc_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &table->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &peer->endpoint_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &n->list_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rq->__lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 namespace_sem &rq->__lock irq_context: 0 proto_tab_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle stock_lock irq_context: 0 namespace_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle pcpu_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem kfence_freelist_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &data->fib_event_queue_lock irq_context: 0 namespace_sem &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &n->list_lock &c->lock irq_context: 0 namespace_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle percpu_counters_lock irq_context: 0 rtnl_mutex dev_addr_sem &pcp->lock &zone->lock irq_context: 0 rtnl_mutex dev_addr_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle pcpu_lock stock_lock irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 rtnl_mutex _xmit_SIT irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_node_0 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&w->w) &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) kfence_freelist_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &n->list_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock percpu_counters_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &c->lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rq->__lock cpu_asid_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem rcu_read_lock rcu_node_0 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem rcu_read_lock &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle key irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex devnet_rename_sem &meta->lock irq_context: 0 rtnl_mutex devnet_rename_sem kfence_freelist_lock irq_context: 0 rtnl_mutex pool_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&barr->work) irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->mcast.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock pcpu_lock stock_lock irq_context: 0 sk_lock-AF_CAIF key irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock key#24 irq_context: 0 rtnl_mutex dev_addr_sem &br->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &wg->static_identity.lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock nl_table_lock irq_context: 0 &sighand->siglock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock nl_table_wait.lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &br->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock &base->lock irq_context: 0 rtnl_mutex &br->lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAIF pcpu_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock percpu_counters_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 pcpu_alloc_mutex rcu_node_0 irq_context: 0 pcpu_alloc_mutex &rcu_state.expedited_wq irq_context: 0 pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pcpu_alloc_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAIF percpu_counters_lock irq_context: 0 sk_lock-AF_CAIF pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle pcpu_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &c->lock irq_context: 0 rtnl_mutex _xmit_TUNNEL irq_context: 0 rtnl_mutex _xmit_IPGRE irq_context: 0 rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_CAIF pool_lock#2 irq_context: 0 sk_lock-AF_CAIF &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 cb_lock remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 tty_mutex console_lock console_srcu &rq->__lock irq_context: 0 tty_mutex console_lock console_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 sb_internal jbd2_handle &rq->__lock irq_context: 0 sb_internal jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: softirq (&brmctx->ip6_own_query.timer) irq_context: softirq (&brmctx->ip6_own_query.timer) &br->multicast_lock irq_context: softirq (&brmctx->ip4_own_query.timer) irq_context: softirq (&brmctx->ip4_own_query.timer) &br->multicast_lock irq_context: softirq (&in_dev->mr_ifc_timer) irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &dir->lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &ul->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount#7 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &dir->lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &tbl->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &n->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount#8 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) batched_entropy_u32.lock irq_context: softirq (&in_dev->mr_ifc_timer) &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) &base->lock irq_context: softirq (&in_dev->mr_ifc_timer) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (work_completion)(&local->rx_work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rt6_exception_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (work_completion)(&local->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem rcu_read_lock &rq->__lock irq_context: 0 key_types_sem root_key_user.cons_lock &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 sb_internal rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock irq_context: 0 key_types_sem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 key_types_sem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 key_types_sem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 sk_lock-AF_INET6 vmap_area_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq irq_context: 0 &xt[i].mutex &mm->mmap_lock key irq_context: 0 &xt[i].mutex &mm->mmap_lock pcpu_lock irq_context: 0 sk_lock-AF_X25 x25_list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx (console_sem).lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx (console_sem).lock &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pool_lock#2 irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 kn->active#4 irq_context: 0 kn->active#4 &root->deactivate_waitq irq_context: 0 bt_proto_lock &____s->seqcount#2 irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 ax25_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu kfence_freelist_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#169 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &meta->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem kfence_freelist_lock irq_context: 0 sb_writers#3 sb_internal rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock &rq->__lock cpu_asid_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &____s->seqcount irq_context: 0 &type->i_mutex_dir_key/1 &n->list_lock irq_context: 0 &type->i_mutex_dir_key/1 &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#66 &rq->__lock irq_context: 0 &type->s_umount_key#66 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#4 oom_adj_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &obj_hash[i].lock pool_lock irq_context: 0 &p->lock remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#13 irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &rq->__lock irq_context: softirq &(&br->gc_work)->timer irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&br->gc_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex _xmit_TUNNEL6 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx &dev->mutex fs_reclaim &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex (console_sem).lock irq_context: 0 alg_types_sem irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rq->__lock irq_context: 0 sk_lock-AF_ALG irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 pool_lock#2 irq_context: 0 sk_lock-AF_ALG slock-AF_ALG irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock fs_reclaim irq_context: 0 tty_mutex &tty->legacy_mutex console_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 tty_mutex &tty->legacy_mutex console_lock pool_lock#2 irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &obj_hash[i].lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &____s->seqcount irq_context: 0 tty_mutex &tty->legacy_mutex console_lock kbd_event_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock kbd_event_lock led_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock genl_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock (worker)->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock (worker)->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_mayday_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &x->wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &r->consumer_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->forw_bcast_list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 &rq->__lock cpu_asid_lock irq_context: 0 nfc_devlist_mutex dpm_list_mtx &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 nfc_devlist_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &pcp->lock &zone->lock irq_context: softirq &(&hwstats->traffic_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &n->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &c->lock irq_context: 0 sk_lock-AF_X25 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex devnet_rename_sem &____s->seqcount irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &____s->seqcount irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock irq_context: 0 sk_lock-AF_INET &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx sched_map-wait-type-override &rq->__lock irq_context: 0 misc_mtx sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 &n->list_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &n->list_lock &c->lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &bridge_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->gw.list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->bat_v.ogm_wq)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->bat_v.ogm_buff_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->bat_v.ogm_buff_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->bat_v.ogm_buff_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tvlv.container_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tvlv.handler_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->nc.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#17 irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&barr->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#18 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tvlv.container_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tvlv.container_list_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->dat.work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &hash->list_locks[i] irq_context: 0 kn->active#4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->bla.work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#20 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#20 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#20 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#20 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->mcast.work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->tt.work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#16 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#16 &bat_priv->softif_vlan_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#16 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#16 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#16 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#21 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.req_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.changes_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &bat_priv->tt.roam_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&bat_priv->orig_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#19 irq_context: 0 &p->lock remove_cache_srcu &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &____s->seqcount#2 irq_context: 0 rtnl_mutex dev_addr_sem &____s->seqcount#2 irq_context: 0 rtnl_mutex &tb->tb6_lock quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_mayday_lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 &c->lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 &n->list_lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &hn->hn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &x->wait#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)inet_frag_wq irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock irq_context: 0 &group->mark_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &xt[i].mutex rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#18 fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &this->info_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pnettable->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pnetids_ndev->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 rlock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6/1 &list->lock#19 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &sctp_ep_hashtable[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock k-slock-AF_INET6/1 k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET6 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events fqdir_free_work &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem &n->list_lock &c->lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock crngs.lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: 0 tomoyo_ss remove_cache_srcu &base->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: softirq rcu_read_lock &br->hash_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_node_0 irq_context: 0 kn->active#47 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock stock_lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem key_user_lock irq_context: 0 key_types_sem root_key_user.cons_lock irq_context: 0 key_types_sem root_key_user.cons_lock key_user_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &genl_data->genl_data_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &genl_data->genl_data_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem root_key_user.cons_lock root_key_user.lock irq_context: 0 key_types_sem root_key_user.cons_lock fs_reclaim irq_context: 0 sk_lock-AF_INET6 &sighand->siglock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &n->list_lock irq_context: 0 key_types_sem &type->lock_class#2 irq_context: 0 key_types_sem root_key_user.cons_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: softirq (&app->periodic_timer) irq_context: softirq (&app->periodic_timer) &app->lock#2 irq_context: softirq (&app->periodic_timer) &app->lock#2 &obj_hash[i].lock irq_context: softirq (&app->periodic_timer) &app->lock#2 &base->lock irq_context: softirq (&app->periodic_timer) &app->lock#2 &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->hash_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu pool_lock#2 irq_context: 0 &mm->mmap_lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bond->stats_lock/1 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 key_types_sem root_key_user.cons_lock pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_INET6 &sighand->siglock &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &____s->seqcount irq_context: 0 tty_mutex tty_mutex.wait_lock irq_context: 0 tty_mutex.wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->ipv4.ra_mutex irq_context: 0 &sb->s_type->i_mutex_key#10 &hashinfo->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pool_lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &n->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &n->list_lock &c->lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pcpu_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock fs_reclaim pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &vma->vm_lock->lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &____s->seqcount irq_context: 0 tomoyo_ss remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&app->join_timer) batched_entropy_u32.lock crngs.lock irq_context: softirq rcu_read_lock &br->hash_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&br->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock batched_entropy_u8.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock rcu_read_lock kfence_freelist_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount#9 irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock quarantine_lock irq_context: 0 misc_mtx pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_CAN &rq->__lock irq_context: 0 sk_lock-AF_CAN &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock &sighand->siglock stock_lock irq_context: 0 &p->lock &mm->mmap_lock &rq->__lock irq_context: 0 &p->lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 tasklist_lock &sighand->siglock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 tomoyo_ss &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[1] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_node_0 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 key irq_context: 0 key_types_sem root_key_user.cons_lock crngs.lock irq_context: 0 key_types_sem root_key_user.cons_lock key_serial_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &rq->__lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim pool_lock#2 irq_context: 0 &type->s_umount_key#30 stock_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 key irq_context: 0 &type->s_umount_key#30 pcpu_lock irq_context: 0 &type->s_umount_key#30 percpu_counters_lock irq_context: 0 &type->s_umount_key#30 pcpu_lock stock_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock quarantine_lock irq_context: 0 tty_mutex &c->lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex irq_context: 0 syslog_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 rcu_node_0 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock quarantine_lock irq_context: 0 &ep->mtx &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 &n->list_lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 &n->list_lock &c->lock irq_context: 0 loop_validate_mutex stock_lock irq_context: 0 loop_validate_mutex &obj_hash[i].lock irq_context: 0 loop_validate_mutex key irq_context: 0 loop_validate_mutex pcpu_lock irq_context: 0 loop_validate_mutex percpu_counters_lock irq_context: 0 loop_validate_mutex pcpu_lock stock_lock irq_context: 0 loop_validate_mutex pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 sb_writers#3 key irq_context: 0 sb_writers#3 pcpu_lock irq_context: 0 sb_writers#3 percpu_counters_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock key irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pcpu_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex &rq->__lock irq_context: 0 &pipe->mutex/1 &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock pool_lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu quarantine_lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem stock_lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex fs_reclaim irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &kernfs_locks->open_file_mutex[count] &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem &rq->__lock cpu_asid_lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER &n->list_lock &c->lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 stock_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 key irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock kfence_freelist_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 pcpu_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 percpu_counters_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &rq->__lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 irq_context: 0 rtnl_mutex &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 stock_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex (inet6addr_validator_chain).rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->ipv6.addrconf_hash_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ifa->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock deferred_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock (console_sem).lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: softirq (&pmctx->ip6_own_query.timer) irq_context: softirq (&pmctx->ip6_own_query.timer) &br->multicast_lock irq_context: softirq (&pmctx->ip4_own_query.timer) irq_context: softirq (&pmctx->ip4_own_query.timer) &br->multicast_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock deferred_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq &list->lock#12 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq rcu_read_lock &br->hash_lock irq_context: softirq rcu_read_lock &br->hash_lock pool_lock#2 irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock &br->hash_lock nl_table_lock irq_context: softirq rcu_read_lock &br->hash_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->hash_lock nl_table_wait.lock irq_context: softirq rcu_read_lock &br->multicast_lock irq_context: softirq rcu_read_lock &br->multicast_lock pool_lock#2 irq_context: softirq rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: softirq rcu_read_lock &br->multicast_lock &c->lock irq_context: softirq rcu_read_lock &br->multicast_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock &br->multicast_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock &br->multicast_lock &____s->seqcount irq_context: softirq rcu_read_lock &br->multicast_lock &dir->lock#2 irq_context: softirq rcu_read_lock &br->multicast_lock deferred_lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &br->multicast_lock nl_table_lock irq_context: softirq rcu_read_lock &br->multicast_lock nl_table_wait.lock irq_context: softirq rcu_read_lock &br->multicast_lock &base->lock irq_context: softirq rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events key_gc_work &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &obj_hash[i].lock pool_lock irq_context: 0 kn->active#4 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex devnet_rename_sem &pcp->lock &zone->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &pcp->lock &zone->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount#2 irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 proto_tab_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &pcp->lock &zone->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 (wq_completion)events key_gc_work &base->lock irq_context: 0 (wq_completion)events free_ipc_work quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &pcp->lock &zone->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem quarantine_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 irq_context: 0 rtnl_mutex _xmit_ETHER/1 _xmit_ETHER irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem key#16 &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events key_gc_work &base->lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &hsr->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: softirq rcu_read_lock pcpu_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 rcu_read_lock _xmit_ETHER &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &obj_hash[i].lock pool_lock irq_context: softirq (&hsr->announce_timer) irq_context: softirq (&hsr->announce_timer) rcu_read_lock pool_lock#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &obj_hash[i].lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock pool_lock#2 irq_context: softirq (&hsr->announce_timer) rcu_read_lock &obj_hash[i].lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &base->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock &br->multicast_lock &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex (switchdev_blocking_notif_chain).rwsem &rq->__lock irq_context: 0 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &q->queue_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &q->queue_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &p->alloc_lock &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex root_key_user.lock irq_context: 0 rtnl_mutex rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock remove_cache_srcu fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock quarantine_lock irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mount_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &xa->xa_lock#19 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex klist_remove_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq rcu_read_lock &br->multicast_lock &n->list_lock irq_context: softirq rcu_read_lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &c->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &____s->seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &nn->netlink_tap_lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bat_priv->tt.changes_list_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock key#16 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex j1939_netdev_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex &vlan_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER (console_sem).lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex (inet6addr_validator_chain).rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &n->list_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 rcu_read_lock &n->list_lock irq_context: 0 rcu_read_lock &n->list_lock &c->lock irq_context: 0 rcu_read_lock key irq_context: 0 rcu_read_lock percpu_counters_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER &pcp->lock &zone->lock irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem _xmit_ETHER &____s->seqcount irq_context: softirq rcu_read_lock &meta->lock irq_context: softirq rcu_read_lock kfence_freelist_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: softirq (&in_dev->mr_ifc_timer) batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex _xmit_ETHER (console_sem).lock irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock pool_lock#2 irq_context: softirq rcu_read_lock &list->lock#13 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &list->lock#13 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) pool_lock#2 irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock &n->list_lock irq_context: 0 rtnl_mutex &bat_priv->softif_vlan_list_lock &n->list_lock &c->lock irq_context: 0 &p->lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &pcp->lock &zone->lock irq_context: 0 rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock pool_lock irq_context: 0 rtnl_mutex _xmit_ETHER &pcp->lock &zone->lock irq_context: 0 rtnl_mutex _xmit_ETHER &pcp->lock &zone->lock &____s->seqcount irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &c->lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &____s->seqcount irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 rtnl_mutex dev_addr_sem &hard_iface->bat_iv.ogm_buff_mutex irq_context: 0 rtnl_mutex key#20 irq_context: 0 rtnl_mutex &bat_priv->tt.commit_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex lweventlist_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock (console_sem).lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle &rq->__lock irq_context: 0 rtnl_mutex lweventlist_lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex lweventlist_lock kfence_freelist_lock irq_context: 0 rtnl_mutex mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#8 irq_context: 0 rtnl_mutex &dir->lock irq_context: 0 rtnl_mutex k-slock-AF_INET/1 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET irq_context: 0 rtnl_mutex k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 rtnl_mutex k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex k-slock-AF_INET#2 irq_context: 0 rtnl_mutex rcu_read_lock (console_sem).lock irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex rcu_read_lock console_lock console_srcu console_owner console_owner_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 rtnl_mutex k-slock-AF_INET6 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock crngs.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh key#20 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock irq_context: 0 rtnl_mutex &wg->device_update_lock fs_reclaim irq_context: 0 rtnl_mutex &wg->device_update_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &wg->device_update_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &wg->device_update_lock &sb->s_type->i_lock_key#8 irq_context: 0 rtnl_mutex &wg->device_update_lock &dir->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-slock-AF_INET/1 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET &table->hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET &table->hash[i].lock k-clock-AF_INET irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-slock-AF_INET#2 irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock jump_label_mutex irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-slock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 &table->hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 &table->hash[i].lock k-clock-AF_INET6 irq_context: 0 rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock &wg->socket_update_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock &list->lock#14 irq_context: 0 rtnl_mutex &wg->device_update_lock &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &____s->seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem pcpu_lock stock_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rhashtable_bucket irq_context: 0 misc_mtx (console_sem).lock irq_context: 0 misc_mtx &wq->mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx pin_fs_lock irq_context: 0 misc_mtx &wq->mutex &pool->lock irq_context: 0 misc_mtx &wq->mutex &x->wait#10 irq_context: 0 misc_mtx wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock (worker)->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &base->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tb->tb6_lock &base->lock &obj_hash[i].lock irq_context: 0 misc_mtx wq_mayday_lock irq_context: 0 &tty->ldisc_sem &rq->__lock irq_context: 0 &tty->ldisc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 remove_cache_srcu pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 fs_reclaim &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 batched_entropy_u32.lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 pcpu_alloc_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 shrinker_mutex irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock &obj_hash[i].lock pool_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#7 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 key#14 irq_context: 0 sb_writers#3 jbd2_handle bit_wait_table + i irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle bit_wait_table + i irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &c->lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock &dir->lock#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock &ul->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 &c->lock irq_context: 0 sk_lock-AF_INET6 &rq->__lock irq_context: 0 sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &folio_wait_table[i] &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#30 &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 key#25 irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &c->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &cfs_rq->removed.lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &base->lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) &ht->mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock &r->producer_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 &wb->work_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 &wb->work_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 &wb->work_lock &base->lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 bit_wait_table + i irq_context: hardirq &ret->b_uptodate_lock bit_wait_table + i &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 key_types_sem &type->lock_class &____s->seqcount#2 irq_context: 0 key_types_sem &type->lock_class &pcp->lock &zone->lock irq_context: 0 key_types_sem &type->lock_class &pcp->lock &zone->lock &____s->seqcount irq_context: 0 key_types_sem &type->lock_class &____s->seqcount irq_context: 0 &wb->work_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &rq->__lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &rq->__lock &base->lock irq_context: 0 &type->s_umount_key#30 &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &n->list_lock irq_context: 0 &type->s_umount_key#30 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock quarantine_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu pool_lock#2 irq_context: 0 loop_validate_mutex &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) crngs.lock irq_context: 0 &sig->cred_guard_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)bat_events &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#57 &rq->__lock irq_context: 0 &type->s_umount_key#57 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fs_reclaim key irq_context: 0 fs_reclaim pcpu_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fs_reclaim percpu_counters_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &ssp->srcu_sup->srcu_cb_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex &lo->lo_mutex &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock irq_context: 0 rtnl_mutex devnet_rename_sem batched_entropy_u8.lock irq_context: 0 key_types_sem crngs.lock irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &p->alloc_lock &p->pi_lock &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem fs_reclaim &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &p->alloc_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 rtnl_mutex net_rwsem percpu_counters_lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 list_lrus_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex udc_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex udc_lock fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex udc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex udc_lock pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &xa->xa_lock#15 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &xa->xa_lock#15 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &sb->s_type->i_lock_key#33 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &s->s_inode_list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &sb->s_type->i_lock_key#33 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#68 irq_context: 0 &type->s_umount_key#68 &x->wait#23 irq_context: 0 &type->s_umount_key#68 shrinker_mutex irq_context: 0 &type->s_umount_key#68 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#68 pool_lock#2 irq_context: 0 &type->s_umount_key#68 sb_mutex irq_context: 0 &type->s_umount_key#68 sb_mutex rename_lock.seqcount irq_context: 0 &type->s_umount_key#68 sb_mutex &dentry->d_lock irq_context: 0 &type->s_umount_key#68 sb_mutex &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#68 sb_mutex &sb->s_type->i_lock_key#33 irq_context: 0 &type->s_umount_key#68 sb_mutex &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#68 sb_mutex &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#68 sb_mutex &obj_hash[i].lock irq_context: 0 &type->s_umount_key#68 sb_mutex pool_lock#2 irq_context: 0 &type->s_umount_key#68 sb_mutex &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#68 sb_mutex &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#68 sb_mutex &dentry->d_lock pool_lock#2 irq_context: 0 &type->s_umount_key#68 sb_mutex rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#68 sb_mutex &dentry->d_lock/1 irq_context: 0 sb_mutex irq_context: 0 sb_mutex sb_lock irq_context: 0 sb_mutex unnamed_dev_ida.xa_lock irq_context: 0 sb_mutex &obj_hash[i].lock irq_context: 0 sb_mutex pool_lock#2 irq_context: 0 &xa->xa_lock#15 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex udc_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &c->lock irq_context: 0 &type->s_umount_key#68 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem key_serial_lock irq_context: 0 &type->s_umount_key#68 sb_mutex &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#68 sb_mutex &rq->__lock irq_context: 0 &type->s_umount_key#68 sb_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &xa->xa_lock#15 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 shrinker_mutex &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 shrinker_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex fs_reclaim &rq->__lock irq_context: 0 &type->s_umount_key#68 sb_mutex &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &obj_hash[i].lock pool_lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 &n->list_lock &c->lock irq_context: 0 key_types_sem key_construction_mutex irq_context: 0 rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem init_user_ns.keyring_sem irq_context: 0 key_types_sem root_key_user.lock irq_context: 0 key_types_sem key_construction_mutex keyring_name_lock irq_context: 0 key_types_sem rcu_read_lock &pool->lock irq_context: 0 key_types_sem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 key_types_sem rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh &r->producer_lock#3 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem remove_cache_srcu pool_lock#2 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem remove_cache_srcu &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->rq_qos_mutex &rq->__lock irq_context: 0 &q->rq_qos_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &dentry->d_lock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &dentry->d_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex free_vmap_area_lock &meta->lock irq_context: 0 crypto_alg_sem &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#18 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex free_vmap_area_lock kfence_freelist_lock irq_context: 0 crypto_alg_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 sk_lock-AF_CAN fs_reclaim irq_context: 0 sk_lock-AF_CAN fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_CAN fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_CAN fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN &c->lock irq_context: 0 sk_lock-AF_CAN pool_lock#2 irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock fs_reclaim irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock &____s->seqcount irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock &c->lock irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_node_0 irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock &net->can.rcvlists_lock irq_context: 0 sk_lock-AF_CAN &priv->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.gp_wq irq_context: 0 rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_CAN &priv->lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_node_0 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock &rq->__lock irq_context: hardirq|softirq &x->wait#14 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#699 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex bpf_devs_lock &rq->__lock irq_context: 0 key_types_sem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 key_types_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 key_types_sem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem &x->wait#17 irq_context: 0 key_types_sem running_helpers_waitq.lock irq_context: 0 key_types_sem key_construction_mutex &obj_hash[i].lock irq_context: 0 key_types_sem key_construction_mutex &base->lock irq_context: 0 key_types_sem key_construction_mutex &base->lock &obj_hash[i].lock irq_context: 0 key_types_sem key_construction_mutex &rq->__lock irq_context: 0 key_types_sem key_construction_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex bpf_devs_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem key_construction_mutex &type->lock_class#3/1 irq_context: 0 key_types_sem key_construction_mutex &type->lock_class#3/1 &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 key_types_sem key_construction_mutex &type->lock_class#3/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem key_construction_mutex &type->lock_class#3/1 rcu_read_lock &pool->lock irq_context: 0 key_types_sem key_construction_mutex &type->lock_class#3/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 key_types_sem key_construction_mutex &type->lock_class#3/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &c->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work quarantine_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uts_sem &rq->__lock irq_context: 0 key_types_sem key_construction_mutex &type->lock_class#3/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 key_types_sem key_construction_mutex &type->lock_class#3/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events key_gc_work &type->lock_class irq_context: 0 (wq_completion)events key_gc_work &type->lock_class fs_reclaim irq_context: 0 (wq_completion)events key_gc_work &type->lock_class fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events key_gc_work &type->lock_class pool_lock#2 irq_context: 0 (wq_completion)events key_gc_work &type->lock_class &c->lock irq_context: 0 uts_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu batched_entropy_u8.lock irq_context: 0 &fsnotify_mark_srcu kfence_freelist_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock batched_entropy_u8.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock kfence_freelist_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &meta->lock irq_context: 0 (wq_completion)events key_gc_work &type->lock_class &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PPPOX &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 key_types_sem rcu_read_lock &rq->__lock irq_context: 0 key_types_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem &type->lock_class &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 key_types_sem bit_wait_table + i irq_context: 0 key_types_sem bit_wait_table + i &p->pi_lock irq_context: 0 key_types_sem bit_wait_table + i &p->pi_lock &rq->__lock irq_context: softirq &keypair->receiving_counter.lock irq_context: softirq &peer->keypairs.keypair_update_lock irq_context: softirq &list->lock#14 irq_context: softirq rcu_read_lock_bh &base->lock irq_context: 0 rtnl_mutex &wg->device_update_lock tk_core.seq.seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 &mm->mmap_lock key irq_context: 0 &mm->mmap_lock pcpu_lock irq_context: 0 &mm->mmap_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 &xt[i].mutex remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) rcu_node_0 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rcu_state.expedited_wq irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &child->perf_event_mutex &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &child->perf_event_mutex &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#228 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN &priv->lock &obj_hash[i].lock irq_context: 0 &child->perf_event_mutex pool_lock#2 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 misc_mtx rcu_read_lock rcu_node_0 irq_context: 0 &xt[i].mutex remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &rq->__lock irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &root->kernfs_rwsem &meta->lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_AX25 rlock-AF_AX25 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &rq->__lock irq_context: 0 cb_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_CAN remove_cache_srcu irq_context: 0 sb_writers#4 &mm->mmap_lock &rq->__lock irq_context: 0 sk_lock-AF_CAN remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#228 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &dev->mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 &type->lock_class fs_reclaim irq_context: 0 &mm->mmap_lock remove_cache_srcu stock_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex quarantine_lock irq_context: softirq rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->page_table_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->lock_class fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx rfkill_global_mutex irq_context: 0 misc_mtx rfkill_global_mutex fs_reclaim irq_context: 0 misc_mtx rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx rfkill_global_mutex pool_lock#2 irq_context: 0 misc_mtx rfkill_global_mutex &rfkill->lock irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx irq_context: 0 cb_lock genl_mutex rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock rcu_read_lock pool_lock#2 irq_context: 0 cb_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 cb_lock genl_mutex hwsim_radio_lock irq_context: 0 cb_lock genl_mutex &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex &x->wait#9 irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex gdp_mutex irq_context: 0 cb_lock genl_mutex gdp_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex lock irq_context: 0 cb_lock genl_mutex lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex bus_type_sem irq_context: 0 cb_lock genl_mutex sysfs_symlink_target_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex &dev->power.lock irq_context: 0 cb_lock genl_mutex dpm_list_mtx irq_context: 0 cb_lock genl_mutex uevent_sock_mutex irq_context: 0 cb_lock genl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &c->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#14 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pool_lock#2 irq_context: 0 tty_mutex (work_completion)(&buf->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &wg->device_update_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN &priv->j1939_socks_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock smack_known_lock.wait_lock irq_context: 0 sb_writers#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_wait_done_commit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#7 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#49 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#49 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#51 &rq->__lock irq_context: 0 &type->s_umount_key#51 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &wq->mutex &rq->__lock irq_context: 0 &type->s_umount_key#30 &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 kfence_freelist_lock irq_context: 0 sb_writers#7 kn->active#4 batched_entropy_u8.lock irq_context: 0 sb_writers#7 kn->active#4 kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 bit_wait_table + i irq_context: hardirq &ret->b_uptodate_lock bit_wait_table + i &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#7 kn->active#4 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &mapping->i_private_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &lock->wait_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &lock->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 sb_internal jbd2_handle irq_context: 0 sb_writers#3 sb_internal &journal->j_state_lock irq_context: 0 cb_lock genl_mutex subsys mutex#53 irq_context: 0 cb_lock genl_mutex subsys mutex#53 &k->k_lock irq_context: 0 cb_lock genl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#3 sb_internal &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex device_links_lock irq_context: 0 cb_lock genl_mutex &k->k_lock irq_context: 0 cb_lock genl_mutex deferred_probe_mutex irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal &journal->j_state_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal &journal->j_state_lock &base->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#3 sb_internal &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &____s->seqcount irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 &type->s_umount_key#30 batched_entropy_u8.lock irq_context: 0 &type->s_umount_key#30 kfence_freelist_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#30 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss pcpu_lock stock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex stock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex key irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex percpu_counters_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex pcpu_lock stock_lock irq_context: 0 sb_writers#3 sb_internal &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem bit_wait_table + i irq_context: 0 (wq_completion)bat_events &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex kfence_freelist_lock irq_context: 0 cb_lock genl_mutex genl_mutex.wait_lock irq_context: 0 cb_lock genl_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 tasklist_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex pcpu_alloc_mutex irq_context: 0 cb_lock genl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex pcpu_alloc_mutex &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex wq_pool_mutex irq_context: 0 cb_lock genl_mutex wq_pool_mutex &wq->mutex irq_context: 0 cb_lock genl_mutex crngs.lock irq_context: 0 cb_lock genl_mutex triggers_list_lock irq_context: 0 cb_lock genl_mutex leds_list_lock irq_context: 0 cb_lock genl_mutex &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex param_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex param_lock rate_ctrl_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex (console_sem).lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &k->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &k->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex kobj_ns_type_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx bus_type_sem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx sysfs_symlink_target_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &dev->power.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dpm_list_mtx irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx reg_requests_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &base->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &k->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex bus_type_sem irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 cb_lock genl_mutex rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &dev->power.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex dpm_list_mtx irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &rfkill->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &k->k_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex subsys mutex#40 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex subsys mutex#40 &k->k_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex triggers_list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex leds_list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex.wait_lock irq_context: 0 cb_lock genl_mutex pin_fs_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (work_completion)(&(&wb->dwork)->work) irq_context: 0 rcu_read_lock (wq_completion)writeback irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (work_completion)(&(&wb->bw_dwork)->work) irq_context: 0 &bdi->cgwb_release_mutex irq_context: 0 &bdi->cgwb_release_mutex cgwb_lock irq_context: 0 subsys mutex#37 &k->k_lock klist_remove_lock irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx stack_depot_init_mutex irq_context: 0 &q->sysfs_dir_lock sysfs_symlink_target_lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex pcpu_lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem pool_lock#2 irq_context: softirq rcu_callback rcu_read_lock &q->mq_freeze_wq irq_context: 0 &q->debugfs_mutex pin_fs_lock irq_context: 0 &q->debugfs_mutex &dentry->d_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 batched_entropy_u8.lock crngs.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#4 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &list->lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &tn->lock irq_context: 0 misc_mtx hrtimer_bases.lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &x->wait#9 irq_context: softirq rcu_read_lock rcu_read_lock quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &k->k_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &dir->lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &dev->power.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx dev_base_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx input_pool.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &tbl->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx sysctl_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx failover_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx proc_subdir_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx proc_inum_ida.xa_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx proc_subdir_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &pnettable->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx smc_ib_devices.mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &ndev->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &local->iflist_mtx irq_context: 0 cb_lock genl_mutex (inetaddr_chain).rwsem irq_context: 0 cb_lock genl_mutex inet6addr_chain.lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rhashtable_bucket irq_context: 0 cb_lock genl_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex nl_table_wait.lock irq_context: 0 cb_lock genl_mutex.wait_lock irq_context: 0 cb_lock &p->pi_lock irq_context: 0 cb_lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 cb_lock rtnl_mutex &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 misc_mtx hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &xa->xa_lock#8 irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &base->lock irq_context: 0 major_names_lock &n->list_lock irq_context: 0 major_names_lock &n->list_lock &c->lock irq_context: 0 major_names_lock &rq->__lock irq_context: 0 major_names_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->debugfs_mutex rcu_read_lock &dentry->d_lock irq_context: 0 &q->debugfs_mutex &fsnotify_mark_srcu irq_context: 0 &q->debugfs_mutex &sb->s_type->i_lock_key#7 irq_context: 0 &q->debugfs_mutex &s->s_inode_list_lock irq_context: 0 &q->debugfs_mutex &xa->xa_lock#8 irq_context: 0 &q->debugfs_mutex &obj_hash[i].lock irq_context: 0 &q->debugfs_mutex pool_lock#2 irq_context: 0 &q->debugfs_mutex &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &meta->lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->debugfs_mutex &dentry->d_lock &obj_hash[i].lock irq_context: 0 &q->debugfs_mutex &dentry->d_lock pool_lock#2 irq_context: 0 &q->debugfs_mutex rcu_read_lock mount_lock irq_context: 0 &q->debugfs_mutex rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock quarantine_lock irq_context: 0 &q->debugfs_mutex mount_lock irq_context: 0 &q->debugfs_mutex mount_lock mount_lock.seqcount irq_context: 0 major_names_lock &obj_hash[i].lock irq_context: 0 &data->open_mutex &base->lock irq_context: 0 &data->open_mutex &base->lock &obj_hash[i].lock irq_context: 0 major_names_lock loop_ctl_mutex irq_context: 0 kn->active#52 &____s->seqcount#2 irq_context: 0 subsys mutex#36 &k->k_lock klist_remove_lock irq_context: 0 kn->active#52 &____s->seqcount irq_context: 0 kn->active#52 &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_sockopt_mutex &rcu_state.expedited_wq irq_context: 0 nf_sockopt_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 nf_sockopt_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 nf_sockopt_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock batched_entropy_u8.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#228 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (work_completion)(&q->timeout_work) irq_context: 0 (wq_completion)kintegrityd irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &fsnotify_mark_srcu irq_context: 0 sb_writers &s->s_inode_list_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 kn->active#52 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &table->lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#74 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem rcu_node_0 irq_context: 0 nf_nat_proto_mutex &rq->__lock irq_context: 0 nf_nat_proto_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#6 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 lock#4 &obj_hash[i].lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu irq_context: 0 nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu quarantine_lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu &rq->__lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu &c->lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu &n->list_lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 ebt_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 ebt_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 ebt_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#246 irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#116 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex _xmit_ETHER/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_AX25 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &n->list_lock irq_context: 0 &fc->uapi_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock &n->list_lock &c->lock irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rdev->mgmt_registrations_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (&dwork->timer) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &base->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#8 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &fsnotify_mark_srcu irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_lock_key#7 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &s->s_inode_list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#8 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &dentry->d_lock/1 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx mount_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx mount_lock mount_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (&dwork->timer)#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&link->color_collision_detect_work)->work) irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 cb_lock rtnl_mutex.wait_lock irq_context: 0 cb_lock &rdev->wiphy.mtx rtnl_mutex.wait_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &list->lock#15 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &ifibss->incomplete_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx (console_sem).lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock irq_context: 0 rtnl_mutex _xmit_ETHER console_owner_lock irq_context: 0 rtnl_mutex _xmit_ETHER console_owner irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &data->mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)cfg80211 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &rdev->bss_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx fs_reclaim irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx nl_table_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &c->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &list->lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)events wireless_nlevent_work irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &c->lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &____s->seqcount irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem pool_lock#2 irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem nl_table_lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem nl_table_wait.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_node_0 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 cb_lock genl_mutex &rcu_state.expedited_wq irq_context: 0 cb_lock genl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key#15 irq_context: 0 rtnl_mutex net_rwsem rcu_node_0 irq_context: 0 rtnl_mutex net_rwsem &rcu_state.expedited_wq irq_context: 0 rtnl_mutex net_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex net_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex net_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#46/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &fsnotify_mark_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->lock_class &type->lock_class/1 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &rq->__lock irq_context: 0 &type->lock_class &type->lock_class/1 &n->list_lock irq_context: 0 misc_mtx rfkill_global_mutex &c->lock irq_context: 0 &type->lock_class &type->lock_class/1 &n->list_lock &c->lock irq_context: 0 misc_mtx rfkill_global_mutex &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex.wait_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &____s->seqcount irq_context: 0 tty_mutex (work_completion)(&buf->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#8 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN &priv->lock &n->list_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &hsr->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &base->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &priv->j1939_socks_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem nl_table_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem nl_table_wait.lock irq_context: 0 cb_lock rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &priv->lock irq_context: 0 sk_lock-AF_PPPOX &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 key_types_sem &type->lock_class#2 &obj_hash[i].lock irq_context: 0 key_types_sem &type->lock_class#2 pool_lock#2 irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#50 remove_cache_srcu irq_context: 0 kn->active#50 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#50 remove_cache_srcu &c->lock irq_context: 0 kn->active#50 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers &obj_hash[i].lock irq_context: 0 sb_writers &sbinfo->stat_lock irq_context: 0 sb_writers &xa->xa_lock#8 irq_context: 0 sb_writers &fsnotify_mark_srcu irq_context: 0 &q->rq_qos_mutex irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &q->rq_qos_mutex &stats->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock key#16 irq_context: 0 cb_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.changes_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tvlv.container_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &list->lock#15 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &ifibss->incomplete_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx (console_sem).lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner &port_lock_key irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &q->rq_qos_mutex (&cb->timer) irq_context: 0 &q->rq_qos_mutex &obj_hash[i].lock irq_context: 0 &q->rq_qos_mutex &base->lock irq_context: 0 &q->rq_qos_mutex pool_lock#2 irq_context: 0 &tags->lock irq_context: 0 &xa->xa_lock#10 irq_context: 0 &q->unused_hctx_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &data->mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &base->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &n->list_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &xt[i].mutex &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex &mm->mmap_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &ipvlan->addrs_lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#16 irq_context: softirq &list->lock#16 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->queue_lock &blkcg->lock percpu_ref_switch_lock irq_context: 0 &q->queue_lock &blkcg->lock percpu_ref_switch_lock &obj_hash[i].lock irq_context: 0 &q->queue_lock &blkcg->lock percpu_ref_switch_lock pool_lock#2 irq_context: 0 (&sq->pending_timer) irq_context: 0 (work_completion)(&td->dispatch_work) irq_context: 0 &lo->lo_work_lock irq_context: 0 &x->wait#8 &p->pi_lock irq_context: 0 &x->wait#8 &p->pi_lock &rq->__lock irq_context: 0 &x->wait#8 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 kn->active#50 remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex fs_reclaim &rq->__lock irq_context: 0 &sig->cred_guard_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_iattr_rwsem rcu_read_lock &rq->__lock irq_context: 0 &root->kernfs_iattr_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &p->lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->w) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xa->xa_lock#8 &c->lock irq_context: 0 &ret->b_state_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock &c->lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock &n->list_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock bit_wait_table + i irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 &pl->lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 &pl->lock key#12 irq_context: 0 &type->s_umount_key#30 lock#4 &lruvec->lru_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 key#12 irq_context: 0 &type->s_umount_key#30 rcu_read_lock &n->list_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &n->list_lock &c->lock irq_context: hardirq &ret->b_uptodate_lock bit_wait_table + i irq_context: hardirq &ret->b_uptodate_lock bit_wait_table + i &p->pi_lock irq_context: hardirq &ret->b_uptodate_lock bit_wait_table + i &p->pi_lock &rq->__lock irq_context: hardirq &ret->b_uptodate_lock bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &ret->b_state_lock &journal->j_list_lock irq_context: 0 krc.lock &obj_hash[i].lock irq_context: 0 krc.lock &base->lock irq_context: 0 (&lo->timer) irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &x->wait#20 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu key irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock &base->lock irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &base->lock irq_context: 0 &type->s_umount_key#30 &____s->seqcount#2 irq_context: 0 &type->s_umount_key#30 &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 pcpu_lock irq_context: 0 sk_lock-AF_INET6 percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rq->__lock cpu_asid_lock irq_context: softirq rcu_callback rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM nr_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NETROM rlock-AF_NETROM irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex (&sq->pending_timer) irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex &base->lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex percpu_counters_lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex pcpu_lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex &q->queue_lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &xa->xa_lock#10 irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) pcpu_lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) blk_queue_ida.xa_lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) percpu_ref_switch_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &xa->xa_lock#15 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 stock_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &dentry->d_lock &wq#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_NETROM irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_ROSE irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE irq_context: softirq rcu_callback percpu_ref_switch_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: softirq rcu_callback rcu_read_lock &q->mq_freeze_wq &p->pi_lock irq_context: softirq rcu_callback rcu_read_lock &q->mq_freeze_wq &p->pi_lock &rq->__lock irq_context: softirq rcu_callback rcu_read_lock &q->mq_freeze_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 cb_lock &rq->__lock irq_context: softirq rcu_read_lock hwsim_radio_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex pool_lock irq_context: 0 key_types_sem rcu_node_0 irq_context: 0 &q->debugfs_mutex &obj_hash[i].lock pool_lock irq_context: 0 &q->debugfs_mutex &rq->__lock irq_context: 0 &q->debugfs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem &n->list_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock stock_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem nl_table_wait.lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock key irq_context: 0 rtnl_mutex rcu_read_lock pcpu_lock irq_context: 0 rtnl_mutex rcu_read_lock percpu_counters_lock irq_context: 0 rtnl_mutex rcu_read_lock pcpu_lock stock_lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock.wait_lock irq_context: 0 key_types_sem &n->list_lock &c->lock irq_context: 0 uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 kfence_freelist_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &n->list_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &obj_hash[i].lock pool_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) kfence_freelist_lock irq_context: 0 &p->lock &rq->__lock cpu_asid_lock irq_context: 0 ebt_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &lock->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#7 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#7 tomoyo_ss &meta->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_owner_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock console_owner irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#16 irq_context: softirq rcu_read_lock lock#6 irq_context: softirq rcu_read_lock lock#6 kcov_remote_lock irq_context: softirq rcu_read_lock &ifibss->incomplete_lock irq_context: softirq rcu_read_lock &rdev->wiphy_work_lock irq_context: softirq rcu_read_lock &local->rx_path_lock irq_context: softirq rcu_read_lock &local->rx_path_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sta->lock irq_context: softirq rcu_read_lock &local->rx_path_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx pin_fs_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rlock-AF_NETLINK irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx nl_table_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 &type->s_umount_key#46/1 irq_context: 0 &type->s_umount_key#46/1 fs_reclaim irq_context: 0 &type->s_umount_key#46/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#46/1 pool_lock#2 irq_context: 0 &type->s_umount_key#46/1 pcpu_alloc_mutex irq_context: 0 &type->s_umount_key#46/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#46/1 shrinker_mutex irq_context: 0 &type->s_umount_key#46/1 list_lrus_mutex irq_context: 0 &type->s_umount_key#46/1 sb_lock irq_context: 0 &type->s_umount_key#46/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#46/1 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_lock_key#32 irq_context: 0 &type->s_umount_key#46/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#46/1 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#46/1 &c->lock irq_context: 0 &type->s_umount_key#46/1 &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#46/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#46/1 &____s->seqcount irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_lock_key#32 &dentry->d_lock irq_context: 0 &type->s_umount_key#46/1 binderfs_minors_mutex irq_context: 0 &type->s_umount_key#46/1 binderfs_minors_mutex binderfs_minors.xa_lock irq_context: 0 &type->s_umount_key#46/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &sb->s_type->i_lock_key#32 irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 rename_lock.seqcount irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 fs_reclaim irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 pool_lock#2 irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &dentry->d_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &dentry->d_lock &wq irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &sb->s_type->i_lock_key#32 &dentry->d_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 rcu_read_lock iunique_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 tk_core.seq.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex fs_reclaim irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex pool_lock#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &____s->seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &c->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex css_set_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex cgroup_file_kn_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq rcu_callback rlock-AF_NETLINK irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex lock cgroup_idr_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex cgroup_idr_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex task_group_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 &type->i_mutex_dir_key#6 irq_context: 0 &type->i_mutex_dir_key#6 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#6 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#6 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 irq_context: 0 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#6 &sb->s_type->i_lock_key#30 &dentry->d_lock &wq irq_context: 0 kn->active#48 fs_reclaim irq_context: 0 kn->active#48 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#48 &c->lock irq_context: 0 kn->active#49 fs_reclaim irq_context: 0 kn->active#49 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &x->wait#2 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock hwsim_radio_lock pool_lock#2 irq_context: softirq rcu_read_lock hwsim_radio_lock &c->lock irq_context: softirq rcu_read_lock hwsim_radio_lock &list->lock#16 irq_context: softirq rcu_read_lock &local->rx_path_lock &list->lock#15 irq_context: softirq rcu_read_lock &local->rx_path_lock &rdev->wiphy_work_lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock hwsim_radio_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lock#6 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lock#6 kcov_remote_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock krc.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock krc.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &obj_hash[i].lock irq_context: softirq (&hsr->prune_timer) irq_context: softirq (&hsr->prune_timer) &hsr->list_lock irq_context: softirq (&hsr->prune_timer) &obj_hash[i].lock irq_context: softirq (&hsr->prune_timer) &base->lock irq_context: softirq (&hsr->prune_timer) &base->lock &obj_hash[i].lock irq_context: softirq &(&conn->info_timer)->timer irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&conn->info_timer)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&conn->info_timer)->work) &conn->chan_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#16 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#16 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem inode_hash_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem inode_hash_lock &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &sb->s_type->i_lock_key#30 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &s->s_inode_list_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &xa->xa_lock#8 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &fsnotify_mark_srcu irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &c->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &base->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->pi_lock &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events kernfs_notify_work &root->kernfs_supers_rwsem inode_hash_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex (wq_completion)cpuset_migrate_mm irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &wq->mutex irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &wq->mutex &pool->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &wq->mutex &x->wait#10 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 rename_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &dentry->d_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &root->kernfs_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem tk_core.seq.seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex css_set_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &____s->seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex lock cgroup_idr_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cgroup_idr_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex jump_label_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex jump_label_mutex patch_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock cpuset_mutex callback_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &pcp->lock &zone->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex percpu_counters_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock jump_label_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock jump_label_mutex patch_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_mutex fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex shrinker_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 rcu_read_lock &dentry->d_lock &lru->node[i].lock irq_context: 0 &type->i_mutex_dir_key#7 irq_context: 0 &type->i_mutex_dir_key#7 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#7 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#7 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#7 rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#31 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &sb->s_type->i_lock_key#31 irq_context: 0 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 irq_context: 0 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key#7 &sb->s_type->i_lock_key#31 &dentry->d_lock &wq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &____s->seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 pool_lock#2 irq_context: 0 kn->active#50 fs_reclaim irq_context: 0 kn->active#50 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#50 &c->lock irq_context: 0 kn->active#50 &____s->seqcount#2 irq_context: 0 kn->active#50 &____s->seqcount irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 cb_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &____s->seqcount#2 irq_context: 0 tomoyo_ss &____s->seqcount#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pool_lock#2 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->pi_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->alloc_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &p->alloc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex cpuset_attach_wq.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &obj_hash[i].lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock pool_lock#2 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock krc.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex (wq_completion)cpuset_migrate_mm irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex &wq->mutex irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex &wq->mutex &pool->lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex &wq->mutex &x->wait#10 irq_context: 0 stock_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &xa->xa_lock#15 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &xa->xa_lock#15 pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 stock_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#15 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#15 &____s->seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#15 pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 cb_lock &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &lock->wait_lock irq_context: 0 &type->s_umount_key#46/1 &____s->seqcount#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &____s->seqcount#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &c->lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem stock_lock irq_context: 0 kn->active#51 fs_reclaim irq_context: 0 kn->active#51 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#51 stock_lock irq_context: 0 kn->active#51 pool_lock#2 irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#52 fs_reclaim irq_context: 0 kn->active#52 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#52 stock_lock irq_context: 0 kn->active#52 &c->lock irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &of->mutex kn->active#52 memcg_max_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &xa->xa_lock#15 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &obj_hash[i].lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 stock_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cgroup_mutex.wait_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &____s->seqcount#2 irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex.wait_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &p->pi_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex fs_reclaim irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex devcgroup_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex cpu_hotplug_lock freezer_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#4 &____s->seqcount#2 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cgroup_mutex.wait_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 rcu_read_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &wg->device_update_lock &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#4 &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim rcu_node_0 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim &rcu_state.expedited_wq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 proto_tab_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 &type->s_umount_key#30 &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &n->list_lock &c->lock irq_context: 0 &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &c->lock irq_context: 0 &root->kernfs_rwsem quarantine_lock irq_context: 0 &base->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &base->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu pcpu_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 &p->lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &sem->wait_lock irq_context: 0 &dev->mutex remove_cache_srcu &rq->__lock irq_context: 0 &dev->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 &____s->seqcount#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex &rq->__lock irq_context: 0 rtnl_mutex &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &p->pi_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#7 stock_lock irq_context: 0 &type->i_mutex_dir_key#7 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &c->lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#15 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#15 &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &xa->xa_lock#15 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#7 &root->kernfs_rwsem stock_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &root->kernfs_rwsem &____s->seqcount irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 kn->active#50 stock_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem freezer_mutex freezer_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss &c->lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->alloc_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &p->alloc_lock &newf->file_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 &xa->xa_lock#15 pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &xa->xa_lock#15 &c->lock irq_context: 0 nf_hook_mutex irq_context: 0 nf_hook_mutex fs_reclaim irq_context: 0 nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nf_hook_mutex stock_lock irq_context: 0 nf_hook_mutex pool_lock#2 irq_context: 0 ebt_mutex &mm->mmap_lock irq_context: 0 ebt_mutex &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 ebt_mutex &mm->mmap_lock fs_reclaim irq_context: 0 ebt_mutex &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 ebt_mutex &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 ebt_mutex &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 ebt_mutex &mm->mmap_lock &____s->seqcount irq_context: 0 ebt_mutex &mm->mmap_lock stock_lock irq_context: 0 ebt_mutex &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &c->lock irq_context: 0 ebt_mutex &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &sb->s_type->i_mutex_key#10 stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 sb_writers#4 key irq_context: 0 sb_writers#4 pcpu_lock irq_context: 0 &xt[i].mutex &c->lock irq_context: 0 sb_writers#4 percpu_counters_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock irq_context: 0 tty_mutex &n->list_lock irq_context: 0 &xt[i].mutex free_vmap_area_lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &xt[i].mutex free_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &xt[i].mutex free_vmap_area_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex stock_lock irq_context: 0 &xt[i].mutex vmap_area_lock irq_context: 0 &xt[i].mutex &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &c->lock irq_context: 0 &xt[i].mutex &per_cpu(xt_recseq, i) irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &____s->seqcount irq_context: 0 &xt[i].mutex purge_vmap_area_lock irq_context: 0 &vma->vm_lock->lock stock_lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 nf_hook_mutex &c->lock irq_context: 0 nf_hook_mutex &____s->seqcount#2 irq_context: 0 nf_hook_mutex &____s->seqcount irq_context: 0 nf_nat_proto_mutex irq_context: 0 nf_nat_proto_mutex fs_reclaim irq_context: 0 nf_nat_proto_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nf_nat_proto_mutex pool_lock#2 irq_context: 0 nf_nat_proto_mutex nf_hook_mutex irq_context: 0 nf_nat_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 nf_nat_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nf_nat_proto_mutex nf_hook_mutex stock_lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 nf_nat_proto_mutex cpu_hotplug_lock irq_context: 0 nf_nat_proto_mutex &obj_hash[i].lock irq_context: 0 nf_nat_proto_mutex stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &c->lock irq_context: 0 &xt[i].mutex &n->list_lock irq_context: 0 &xt[i].mutex &n->list_lock &c->lock irq_context: 0 pcpu_alloc_mutex fs_reclaim irq_context: 0 pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 pcpu_alloc_mutex &____s->seqcount irq_context: 0 cb_lock rcu_read_lock &____s->seqcount irq_context: 0 nf_nat_proto_mutex &c->lock irq_context: 0 nf_nat_proto_mutex &____s->seqcount irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 nf_nat_proto_mutex &____s->seqcount#2 irq_context: 0 &pipe->mutex/1 stock_lock irq_context: softirq rcu_callback cgroup_threadgroup_rwsem.rss.gp_wait.lock irq_context: softirq rcu_callback cgroup_threadgroup_rwsem.rss.gp_wait.lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock &xa->xa_lock#8 irq_context: 0 &vma->vm_lock->lock &sb->s_type->i_lock_key irq_context: 0 &vma->vm_lock->lock &info->lock irq_context: 0 &vma->vm_lock->lock lock#4 irq_context: 0 &vma->vm_lock->lock tk_core.seq.seqcount irq_context: 0 &vma->vm_lock->lock mount_lock irq_context: 0 &vma->vm_lock->lock &xa->xa_lock#8 pool_lock#2 irq_context: 0 cb_lock &____s->seqcount#2 irq_context: 0 &vma->vm_lock->lock lock#4 &lruvec->lru_lock irq_context: 0 &xt[i].mutex &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &xa->xa_lock#15 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &____s->seqcount#2 irq_context: 0 rtnl_mutex &dev->tx_global_lock irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_ETHER#2 irq_context: 0 rtnl_mutex &dev->tx_global_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &sch->q.lock irq_context: 0 loop_validate_mutex irq_context: 0 loop_validate_mutex &lo->lo_mutex irq_context: 0 &fsnotify_mark_srcu fs_reclaim irq_context: 0 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 &fsnotify_mark_srcu &group->notification_lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &____s->seqcount#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq (&tun->flow_gc_timer) irq_context: softirq (&tun->flow_gc_timer) &tun->lock irq_context: 0 uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex &base->lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock kernfs_idr_lock irq_context: 0 &type->i_mutex_dir_key/1 &fsnotify_mark_srcu &conn->lock irq_context: 0 &type->i_mutex_dir_key/1 &conn->lock irq_context: 0 &type->i_mutex_dir_key/1 &group->mark_mutex irq_context: 0 &type->i_mutex_dir_key/1 &group->mark_mutex &mark->lock irq_context: 0 &type->i_mutex_dir_key/1 &mark->lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock pool_lock#2 irq_context: 0 &xt[i].mutex init_mm.page_table_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock stock_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &pcp->lock &zone->lock irq_context: softirq rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem stock_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock mmu_notifier_invalidate_range_start irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex irq_context: 0 &mm->mmap_lock stock_lock irq_context: 0 &vma->vm_lock->lock &xa->xa_lock#8 &c->lock irq_context: 0 rtnl_mutex &ndev->lock &tb->tb6_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock hwsim_radio_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &sch->q.lock irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &sch->q.lock irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &meta->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) &zone->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &____s->seqcount#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 fs_reclaim irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 &group->notification_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 &group->notification_waitq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 &group->notification_waitq &ep->lock irq_context: 0 &type->i_mutex_dir_key/1 &group->inotify_data.idr_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &fq->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &local->active_txq_lock[i] irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx (work_completion)(&sta->drv_deliver_wk) irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->s_umount_key#21/1 &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rcu_state.gp_wq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE slock-AF_ROSE irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &c->lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &____s->seqcount#2 irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &____s->seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 &xa->xa_lock#15 irq_context: 0 &sb->s_type->i_mutex_key#9 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 stock_lock irq_context: 0 &r->consumer_lock irq_context: 0 &mm->mmap_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_ROSE irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex &obj_hash[i].lock pool_lock irq_context: 0 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex pool_lock#2 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &rq->__lock &cfs_rq->removed.lock irq_context: 0 tasklist_lock &sighand->siglock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem key irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &br->hash_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock &br->hash_lock &____s->seqcount irq_context: 0 &type->i_mutex_dir_key/1 destroy_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock pool_lock irq_context: 0 fill_pool_map-wait-type-override &rq->__lock cpu_asid_lock irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#49 &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex.wait_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &p->pi_lock irq_context: 0 btf_idr_lock irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem kfence_freelist_lock irq_context: 0 rtnl_mutex dev_addr_sem &meta->lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &base->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &____s->seqcount#2 irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex &pnettable->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u8.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock &meta->lock irq_context: 0 &group->mark_mutex &fsnotify_mark_srcu &rq->__lock irq_context: 0 &group->mark_mutex &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &sem->wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &____s->seqcount#2 irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock kfence_freelist_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_read_lock rcu_node_0 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock quarantine_lock irq_context: 0 cb_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex key irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 _xmit_ETHER pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock stock_lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: 0 cb_lock genl_mutex &x->wait#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &base->lock irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: hardirq &fq->mq_flush_lock quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock batched_entropy_u8.lock crngs.lock irq_context: 0 &fc->uapi_mutex remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 _xmit_ETHER &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem &sn->pipefs_sb_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &sn->pipefs_sb_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uts_sem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &____s->seqcount irq_context: 0 key_types_sem &type->lock_class fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &obj_hash[i].lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock stock_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock key irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 pcpu_lock stock_lock irq_context: 0 br_ioctl_mutex rtnl_mutex krc.lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PPPOX rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_PPPOX rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &priv->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &priv->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock irq_context: 0 uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#18 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#18 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#18 &xa->xa_lock#15 irq_context: 0 cb_lock genl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 fs_reclaim &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &group->notification_waitq &ep->lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 stock_lock irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 &lru->node[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock &c->lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex _xmit_ETHER rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex cgroup_mutex.wait_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 &sbinfo->stat_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 rcu_read_lock &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock cpu_asid_lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_wq[3] irq_context: 0 (wq_completion)nfc2_nci_rx_wq#694 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#692 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &____s->seqcount#2 irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &base->lock &obj_hash[i].lock irq_context: 0 tomoyo_ss rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock &____s->seqcount irq_context: 0 cb_lock remove_cache_srcu irq_context: 0 cb_lock remove_cache_srcu quarantine_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: softirq (&sdp->delay_work) irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock remove_cache_srcu &rq->__lock irq_context: 0 cb_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock &xa->xa_lock#8 stock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &meta->lock irq_context: 0 &fc->uapi_mutex irq_context: 0 cb_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock remove_cache_srcu &c->lock irq_context: 0 cb_lock remove_cache_srcu &n->list_lock irq_context: 0 cb_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 cb_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 cb_lock remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_CAN irq_context: 0 sk_lock-AF_CAN slock-AF_CAN irq_context: 0 sk_lock-AF_CAN &mm->mmap_lock irq_context: 0 slock-AF_CAN irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock &table->lock#3 irq_context: softirq rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &n->list_lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 cgroup_mutex &n->list_lock &c->lock irq_context: 0 sk_lock-AF_X25 irq_context: 0 sk_lock-AF_X25 slock-AF_X25 irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock irq_context: 0 slock-AF_X25 irq_context: 0 kn->active#49 &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &n->list_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &____s->seqcount#2 irq_context: 0 kn->active#52 &n->list_lock irq_context: 0 kn->active#52 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &dir->lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex &c->lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex &____s->seqcount#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex blkcg_pol_mutex &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#7 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &tb->tb6_lock &____s->seqcount#2 irq_context: 0 ebt_mutex &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 &type->i_mutex_dir_key#6 &root->kernfs_rwsem pool_lock#2 irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7/1 cgroup_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &rq->__lock irq_context: 0 sb_writers#9 &type->i_mutex_dir_key#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &c->lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &p->pi_lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 tasklist_lock &p->alloc_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex.wait_lock irq_context: 0 sb_writers#9 &of->mutex &p->pi_lock irq_context: 0 sb_writers#9 &of->mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#9 &of->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &____s->seqcount#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#9 &c->lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &____s->seqcount#2 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &____s->seqcount irq_context: 0 kn->active#51 &c->lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem stock_lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pcpu_alloc_mutex pool_lock#2 irq_context: 0 pcpu_alloc_mutex free_vmap_area_lock irq_context: 0 pcpu_alloc_mutex vmap_area_lock irq_context: 0 pcpu_alloc_mutex &c->lock irq_context: 0 pcpu_alloc_mutex init_mm.page_table_lock irq_context: 0 pcpu_alloc_mutex &pcp->lock &zone->lock irq_context: 0 pcpu_alloc_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 nf_nat_proto_mutex &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem stock_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &____s->seqcount#2 irq_context: 0 nf_hook_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &xt[i].mutex &pcp->lock &zone->lock irq_context: 0 &xt[i].mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &rq->__lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &rq->__lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u32.lock crngs.lock irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 nf_nat_proto_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &xt[i].mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pcpu_lock stock_lock irq_context: 0 &xt[i].mutex fs_reclaim &rq->__lock irq_context: 0 &xt[i].mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh key#20 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 tasklist_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu irq_context: 0 sb_writers#3 &s->s_inode_list_lock irq_context: 0 sb_writers#3 sb_internal irq_context: 0 sb_writers#3 sb_internal mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 sb_internal pool_lock#2 irq_context: 0 sb_writers#3 sb_internal &journal->j_state_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle irq_context: 0 sb_writers#3 sb_internal jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &xa->xa_lock#8 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 sb_internal &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal &rq->__lock irq_context: 0 sb_writers#3 sb_internal &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 inode_hash_lock irq_context: 0 sb_writers#3 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 &type->i_mutex_dir_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &fsnotify_mark_srcu irq_context: 0 sb_writers#7 tomoyo_ss &____s->seqcount#2 irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &wb->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback put_task_map-wait-type-override stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &xa->xa_lock#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &obj_hash[i].lock irq_context: 0 &sb->s_type->i_lock_key#32 irq_context: 0 misc_mtx nfc_index_ida.xa_lock irq_context: 0 misc_mtx &n->list_lock &c->lock irq_context: 0 misc_mtx pcpu_alloc_mutex irq_context: 0 misc_mtx pcpu_alloc_mutex pcpu_lock irq_context: 0 misc_mtx cpu_hotplug_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex pool_lock#2 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &obj_hash[i].lock irq_context: 0 misc_mtx kthread_create_lock irq_context: 0 misc_mtx &x->wait irq_context: 0 misc_mtx wq_pool_mutex irq_context: 0 misc_mtx wq_pool_mutex &wq->mutex irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 tomoyo_ss &n->list_lock irq_context: 0 sb_writers#7 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#7 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#7 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock once_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock once_lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#7 &xattrs->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq rcu_read_lock hwsim_radio_lock &n->list_lock irq_context: softirq rcu_read_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &c->lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &x->wait#9 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 proto_tab_lock irq_context: 0 proto_tab_lock pool_lock#2 irq_context: 0 proto_tab_lock &dir->lock irq_context: 0 proto_tab_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex pool_lock#2 irq_context: 0 misc_mtx nfc_devlist_mutex &k->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex &k->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN slock-AF_CAN irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 misc_mtx nfc_devlist_mutex bus_type_sem irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN clock-AF_CAN irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_CAN irq_context: 0 misc_mtx nfc_devlist_mutex sysfs_symlink_target_lock irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem irq_context: 0 misc_mtx nfc_devlist_mutex &dev->power.lock irq_context: 0 misc_mtx nfc_devlist_mutex dpm_list_mtx irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex fs_reclaim irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 slock-AF_X25 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 x25_list_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 rlock-AF_X25 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_X25 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex subsys mutex#39 irq_context: 0 misc_mtx nfc_devlist_mutex subsys mutex#39 &k->k_lock irq_context: 0 misc_mtx &k->k_lock irq_context: 0 misc_mtx llcp_devices_lock irq_context: 0 misc_mtx &dev->mutex fs_reclaim irq_context: 0 misc_mtx &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx &dev->mutex pool_lock#2 irq_context: 0 misc_mtx &dev->mutex &x->wait#9 irq_context: 0 misc_mtx &dev->mutex &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &k->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex bus_type_sem irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex sysfs_symlink_target_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &dev->power.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex dpm_list_mtx irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rfkill->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &k->k_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex subsys mutex#40 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex subsys mutex#40 &k->k_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex leds_list_lock irq_context: 0 sk_lock-AF_NFC irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_NFC slock-AF_NFC irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 slock-AF_NFC irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex.wait_lock irq_context: 0 misc_mtx &dev->mutex &p->pi_lock irq_context: 0 misc_mtx &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex &rq->__lock irq_context: 0 misc_mtx &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nl_table_lock irq_context: 0 misc_mtx nl_table_wait.lock irq_context: 0 &p->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock irq_context: 0 &mm->mmap_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &sem->wait_lock irq_context: 0 sb_writers#3 sb_internal &c->lock irq_context: 0 &ndev->req_lock &wq->mutex irq_context: 0 &ndev->req_lock &wq->mutex &rq->__lock irq_context: 0 &ndev->req_lock &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#47 irq_context: 0 &type->s_umount_key#48 irq_context: 0 &type->s_umount_key#40 irq_context: 0 &type->s_umount_key#49 irq_context: 0 &type->s_umount_key#6 irq_context: 0 &type->s_umount_key#33 irq_context: 0 &type->s_umount_key#32 irq_context: 0 &type->s_umount_key#50 irq_context: 0 &type->s_umount_key#51 irq_context: 0 &type->s_umount_key#52 irq_context: 0 &type->s_umount_key#53 irq_context: 0 &type->s_umount_key#14 irq_context: 0 &type->s_umount_key#54 irq_context: 0 &type->s_umount_key#55 irq_context: 0 &type->s_umount_key#56 irq_context: 0 &type->s_umount_key#57 irq_context: 0 &type->s_umount_key#58 irq_context: 0 &type->s_umount_key#34 irq_context: 0 &type->s_umount_key#59 irq_context: 0 &type->s_umount_key#36 irq_context: 0 &type->s_umount_key#60 irq_context: 0 &type->s_umount_key#60 &rq->__lock irq_context: 0 &type->s_umount_key#60 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &meta->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) kfence_freelist_lock irq_context: 0 &ndev->req_lock &wq->mutex &pool->lock irq_context: 0 &ndev->req_lock &wq->mutex &x->wait#10 irq_context: 0 &ndev->req_lock (&ndev->cmd_timer) irq_context: 0 &ndev->req_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock &base->lock irq_context: 0 &ndev->req_lock (&ndev->data_timer) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &mapping->i_private_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) pool_lock#2 irq_context: 0 rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &group->mark_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock &rq->__lock irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex (work_completion)(&rfkill->uevent_work) irq_context: 0 &dev->mutex (work_completion)(&rfkill->sync_work) irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &dev->mutex kernfs_idr_lock irq_context: 0 &dev->mutex &k->k_lock klist_remove_lock irq_context: 0 &dev->mutex subsys mutex#40 irq_context: 0 &dev->mutex subsys mutex#40 &k->k_lock irq_context: 0 &dev->mutex subsys mutex#40 &k->k_lock klist_remove_lock irq_context: 0 &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &rfkill->lock irq_context: 0 &dev->mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &dev->mutex triggers_list_lock irq_context: 0 &dev->mutex leds_list_lock irq_context: 0 &local->sockets.lock irq_context: 0 &local->raw_sockets.lock irq_context: 0 (&local->link_timer) irq_context: 0 (work_completion)(&local->tx_work) irq_context: 0 (work_completion)(&local->rx_work) irq_context: 0 (work_completion)(&local->timeout_work) irq_context: 0 (&local->sdreq_timer) irq_context: 0 (work_completion)(&local->sdreq_timeout_work) irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 nfc_devlist_mutex kernfs_idr_lock irq_context: 0 nfc_devlist_mutex subsys mutex#39 &k->k_lock klist_remove_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 nfc_devlist_mutex &x->wait#9 irq_context: 0 nfc_devlist_mutex deferred_probe_mutex irq_context: 0 nfc_devlist_mutex device_links_lock irq_context: 0 nfc_devlist_mutex mmu_notifier_invalidate_range_start irq_context: 0 nfc_devlist_mutex &n->list_lock irq_context: 0 nfc_devlist_mutex &n->list_lock &c->lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 nfc_devlist_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &c->lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 nfc_devlist_mutex uevent_sock_mutex nl_table_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &type->s_umount_key#30 irq_context: 0 &type->s_umount_key#30 (wq_completion)ext4-rsv-conversion irq_context: 0 &type->s_umount_key#30 &wq->mutex irq_context: 0 &type->s_umount_key#30 &wq->mutex &pool->lock irq_context: 0 &type->s_umount_key#30 &wq->mutex &x->wait#10 irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq rcu_callback per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &type->s_umount_key#30 jbd2_handle irq_context: 0 &type->s_umount_key#30 &journal->j_wait_commit irq_context: 0 &type->s_umount_key#30 &journal->j_wait_done_commit irq_context: 0 &type->s_umount_key#30 &rq->__lock irq_context: 0 &type->s_umount_key#30 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &____s->seqcount irq_context: 0 inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &xa->xa_lock#8 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &journal->j_wait_done_commit &p->pi_lock &rq->__lock irq_context: 0 &journal->j_wait_done_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &sb->s_type->i_lock_key#3 irq_context: 0 &type->s_umount_key#30 &wb->list_lock irq_context: 0 &type->s_umount_key#30 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#30 rcu_read_lock &memcg->move_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 &s->s_inode_wblist_lock irq_context: 0 &type->s_umount_key#30 mmu_notifier_invalidate_range_start irq_context: 0 &type->s_umount_key#30 pool_lock#2 irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock pcpu_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock percpu_counters_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock &blkcg->lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &q->queue_lock &blkcg->lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 &c->lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 key#11 irq_context: 0 &type->s_umount_key#30 lock#4 irq_context: 0 &type->s_umount_key#30 lock#5 irq_context: 0 &type->s_umount_key#30 rcu_read_lock key#11 irq_context: 0 &type->s_umount_key#30 &rq_wait->wait irq_context: 0 &type->s_umount_key#30 &__ctx->lock irq_context: 0 &type->s_umount_key#30 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 rcu_node_0 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_node_0 irq_context: hardirq rcu_read_lock &xa->xa_lock#8 key#11 irq_context: hardirq rcu_read_lock &xa->xa_lock#8 key#12 irq_context: hardirq rcu_read_lock &xa->xa_lock#8 key#14 irq_context: hardirq &rq_wait->wait irq_context: hardirq &rq_wait->wait &p->pi_lock irq_context: hardirq &rq_wait->wait &p->pi_lock &rq->__lock irq_context: hardirq &rq_wait->wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &type->s_umount_key#30 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 rcu_read_lock &retval->lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 rcu_read_lock &c->lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &____s->seqcount#2 irq_context: 0 &type->s_umount_key#30 rcu_read_lock &____s->seqcount irq_context: 0 &type->s_umount_key#30 rcu_read_lock &nvmeq->sq_lock irq_context: 0 &type->s_umount_key#30 &folio_wait_table[i] irq_context: 0 &type->s_umount_key#61 irq_context: 0 &type->s_umount_key#29 irq_context: 0 &type->s_umount_key#62 irq_context: 0 &type->s_umount_key#44 irq_context: 0 &type->s_umount_key#45 irq_context: 0 &type->s_umount_key#63 irq_context: 0 &type->s_umount_key#64 irq_context: 0 &type->s_umount_key#65 irq_context: 0 &type->s_umount_key#43 irq_context: 0 &type->s_umount_key#66 irq_context: 0 &type->s_umount_key#30 &fq->mq_flush_lock irq_context: 0 &type->s_umount_key#30 &fq->mq_flush_lock tk_core.seq.seqcount irq_context: 0 &type->s_umount_key#30 &fq->mq_flush_lock &q->requeue_lock irq_context: 0 &type->s_umount_key#30 &fq->mq_flush_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &fq->mq_flush_lock rcu_read_lock &pool->lock irq_context: 0 &type->s_umount_key#30 &fq->mq_flush_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 &fq->mq_flush_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &type->s_umount_key#30 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#30 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &x->wait#26 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock &base->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_NFC irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu &c->lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &ei->i_es_lock key#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 kn->active#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &c->lock irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex kfence_freelist_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu irq_context: 0 cb_lock genl_mutex remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &rq->__lock irq_context: 0 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu pool_lock#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) irq_context: softirq (&ndev->rs_timer) &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) &pcp->lock &zone->lock irq_context: softirq (&ndev->rs_timer) &____s->seqcount irq_context: softirq (&ndev->rs_timer) pool_lock#2 irq_context: softirq (&ndev->rs_timer) &c->lock irq_context: softirq (&ndev->rs_timer) &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) &dir->lock#2 irq_context: softirq (&ndev->rs_timer) &ul->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &____s->seqcount#8 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#11 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock batched_entropy_u32.lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &base->lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override &n->list_lock irq_context: 0 remove_cache_srcu fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rq->__lock irq_context: 0 &xt[i].mutex &lock->wait_lock irq_context: 0 sb_writers#7 kn->active#4 &n->list_lock irq_context: 0 sb_writers#7 kn->active#4 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu &n->list_lock irq_context: 0 &fsnotify_mark_srcu &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &____s->seqcount irq_context: softirq (&peer->timer_persistent_keepalive) irq_context: softirq (&peer->timer_persistent_keepalive) pool_lock#2 irq_context: softirq (&peer->timer_persistent_keepalive) &list->lock#14 irq_context: softirq (&peer->timer_persistent_keepalive) tk_core.seq.seqcount irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex quarantine_lock irq_context: 0 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 tomoyo_ss quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#8 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#8 &xa->xa_lock#15 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#8 &xa->xa_lock#15 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#8 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#8 stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh &r->producer_lock#3 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#3 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &____s->seqcount irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq &(&nsim_dev->trap_data->trap_report_dw)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &ei->i_es_lock key#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) quarantine_lock irq_context: 0 sched_map-wait-type-override &rq->__lock irq_context: 0 sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#16 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#16 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &base->lock irq_context: 0 sk_lock-AF_INET slock-AF_INET#2 &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_node_0 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 cgroup_threadgroup_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 kn->active#4 &____s->seqcount irq_context: 0 sb_writers#7 kn->active#4 pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 quarantine_lock irq_context: 0 &f->f_pos_lock &rq->__lock irq_context: 0 &f->f_pos_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->tt.work)->timer irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) key#16 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) key#21 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &bat_priv->tt.req_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &bat_priv->tt.roam_list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_sockopt_mutex &rq->__lock irq_context: 0 nf_sockopt_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex rcu_read_lock rcu_node_0 irq_context: 0 &xt[i].mutex rcu_read_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock hwsim_radio_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sighand->siglock stock_lock irq_context: 0 &sighand->siglock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock rcu_read_lock &n->list_lock irq_context: 0 cb_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 cb_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &meta->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &n->list_lock irq_context: 0 sb_writers#7 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock &____s->seqcount irq_context: 0 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex remove_cache_srcu irq_context: 0 &xt[i].mutex remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#16 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#16 irq_context: 0 sb_writers#3 sb_internal &n->list_lock irq_context: 0 sb_writers#3 sb_internal &n->list_lock &c->lock irq_context: 0 cb_lock fs_reclaim &rq->__lock irq_context: 0 cb_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &c->lock batched_entropy_u8.lock irq_context: softirq &c->lock kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &bgl->locks[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#7 kn->active#4 &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle key#4 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 &____s->seqcount#2 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_node_0 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock pool_lock#2 irq_context: softirq &tx->clean_lock quarantine_lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock cpu_asid_lock irq_context: 0 &vma->vm_lock->lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss rcu_read_lock &obj_hash[i].lock irq_context: 0 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &n->list_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 ebt_mutex &rq->__lock irq_context: 0 ebt_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ebt_mutex ebt_mutex.wait_lock irq_context: 0 ebt_mutex.wait_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &____s->seqcount irq_context: 0 sk_lock-AF_NETLINK &rq->__lock irq_context: 0 sk_lock-AF_NETLINK &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex remove_cache_srcu &c->lock irq_context: 0 &xt[i].mutex remove_cache_srcu &n->list_lock irq_context: 0 &xt[i].mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &xt[i].mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &xt[i].mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &xt[i].mutex remove_cache_srcu &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &pcp->lock &zone->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu quarantine_lock irq_context: 0 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal &____s->seqcount#2 irq_context: 0 sb_writers#3 sb_internal &pcp->lock &zone->lock irq_context: 0 sb_writers#3 sb_internal &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 sb_internal &____s->seqcount irq_context: softirq fs/notify/mark.c:89 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) &n->list_lock irq_context: softirq (&ndev->rs_timer) &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &____s->seqcount#2 irq_context: 0 &mm->mmap_lock &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rq->__lock &cfs_rq->removed.lock irq_context: 0 &tsk->futex_exit_mutex &rq->__lock irq_context: 0 &tsk->futex_exit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tsk->futex_exit_mutex &cfs_rq->removed.lock irq_context: 0 &tsk->futex_exit_mutex &obj_hash[i].lock irq_context: 0 &tsk->futex_exit_mutex pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock pool_lock#2 irq_context: 0 &fsnotify_mark_srcu &____s->seqcount#2 irq_context: 0 &fsnotify_mark_srcu &pcp->lock &zone->lock irq_context: 0 &fsnotify_mark_srcu &____s->seqcount irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &mm->mmap_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &child->perf_event_mutex &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: softirq (&peer->timer_persistent_keepalive) &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &xa->xa_lock#8 &c->lock irq_context: 0 &fsnotify_mark_srcu &group->notification_waitq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 sb_writers#7 tomoyo_ss &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &cfs_rq->removed.lock irq_context: 0 rtnl_mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 key irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 pool_lock#2 irq_context: 0 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock &____s->seqcount#2 irq_context: 0 &vma->vm_lock->lock &n->list_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 batched_entropy_u8.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 kfence_freelist_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#14 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 oom_adj_mutex rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &pcp->lock &zone->lock irq_context: 0 rtnl_mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &meta->lock irq_context: softirq rcu_read_lock &br->multicast_lock &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock &br->multicast_lock &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&dom->period_timer) &p->sequence key#14 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 key#22 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock rcu_read_lock &ndev->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 misc_mtx rfkill_global_mutex &rq->__lock irq_context: 0 sk_lock-AF_INET &rq->__lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_node_0 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &ndev->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &pcp->lock &zone->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &bgl->locks[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock cpu_asid_lock irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 &base->lock irq_context: 0 &type->i_mutex_dir_key/1 &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tvlv.container_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tvlv.container_list_lock pool_lock#2 irq_context: 0 lock pidmap_lock &____s->seqcount#2 irq_context: 0 lock pidmap_lock &pcp->lock &zone->lock irq_context: 0 lock pidmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 lock map_idr_lock irq_context: 0 lock map_idr_lock &c->lock irq_context: 0 lock map_idr_lock pool_lock#2 irq_context: 0 map_idr_lock irq_context: 0 map_idr_lock &obj_hash[i].lock irq_context: 0 map_idr_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) stock_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &br->multicast_lock &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rename_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle key#4 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#18 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#18 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#18 stock_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 sb_writers#3 sb_internal jbd2_handle key irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &type->s_umount_key#30 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu pool_lock#2 irq_context: 0 kn->active#4 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &base->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &base->lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &new_node->seq_out_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock batched_entropy_u32.lock crngs.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock smack_known_lock.wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock.wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock.wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock.wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu &meta->lock irq_context: 0 remove_cache_srcu kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss percpu_counters_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss pcpu_lock stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &base->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) &base->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh pool_lock#2 irq_context: 0 &group->mark_mutex &rq->__lock irq_context: 0 &group->mark_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &fq->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->active_txq_lock[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->active_txq_lock[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &local->queue_stop_reason_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &fq->lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &list->lock#16 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock &list->lock#16 irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu pool_lock#2 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &list->lock#12 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rq->__lock cpu_asid_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock rcu_node_0 irq_context: 0 remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &bgl->locks[i].lock irq_context: 0 cb_lock genl_mutex fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &k->list_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock gdp_mutex irq_context: 0 tty_mutex &tty->legacy_mutex console_lock gdp_mutex &k->list_lock irq_context: softirq (&peer->timer_retransmit_handshake) irq_context: softirq (&peer->timer_retransmit_handshake) &peer->endpoint_lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 &tsk->futex_exit_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 &xt[i].mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ret->b_state_lock &journal->j_list_lock &obj_hash[i].lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock pool_lock#2 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &mapping->i_private_lock irq_context: 0 key#23 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &n->list_lock &c->lock irq_context: 0 &xt[i].mutex remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &r->producer_lock#3 irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock &c->lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 &xt[i].mutex rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_sockopt_mutex &cfs_rq->removed.lock irq_context: 0 nf_sockopt_mutex &obj_hash[i].lock irq_context: 0 nf_sockopt_mutex pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &n->list_lock &c->lock irq_context: 0 nf_sockopt_mutex nf_sockopt_mutex.wait_lock irq_context: 0 nf_sockopt_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 nf_sockopt_mutex.wait_lock irq_context: 0 &xt[i].mutex batched_entropy_u8.lock irq_context: 0 &xt[i].mutex kfence_freelist_lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock &meta->lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock kfence_freelist_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rq->__lock cpu_asid_lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &p->pi_lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &kcov->lock kcov_remote_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &cfs_rq->removed.lock irq_context: 0 tasklist_lock rcu_read_lock &sighand->siglock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &____s->seqcount#2 irq_context: 0 tomoyo_ss remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#4 &cfs_rq->removed.lock irq_context: 0 &dentry->d_lock &sb->s_type->i_lock_key#22 &lru->node[i].lock irq_context: 0 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_internal irq_context: 0 sb_internal jbd2_handle &ei->i_raw_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &ei->i_raw_lock irq_context: 0 sb_internal jbd2_handle &xa->xa_lock#8 irq_context: 0 sb_internal jbd2_handle &ei->i_es_lock irq_context: 0 inode_hash_lock &sb->s_type->i_lock_key#22 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#66 irq_context: 0 &type->s_umount_key#66 &x->wait#23 irq_context: 0 &type->s_umount_key#66 shrinker_mutex irq_context: 0 &type->s_umount_key#66 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#66 pool_lock#2 irq_context: 0 &type->s_umount_key#66 rename_lock.seqcount irq_context: 0 &type->s_umount_key#66 &dentry->d_lock irq_context: 0 &type->s_umount_key#66 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#66 &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#66 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#66 &sb->s_type->i_lock_key#32 irq_context: 0 &type->s_umount_key#66 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#66 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#66 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#66 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#66 &dentry->d_lock pool_lock#2 irq_context: 0 &type->s_umount_key#66 binderfs_minors_mutex irq_context: 0 &type->s_umount_key#66 binderfs_minors_mutex binderfs_minors.xa_lock irq_context: 0 &mq_lock irq_context: 0 (wq_completion)events free_ipc_work irq_context: 0 (wq_completion)events free_ipc_work &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &x->wait#2 irq_context: 0 (wq_completion)events free_ipc_work sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events free_ipc_work &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kcov_remote_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock mount_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events free_ipc_work mount_lock irq_context: 0 (wq_completion)events free_ipc_work mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)events free_ipc_work &fsnotify_mark_srcu irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 &x->wait#23 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 shrinker_mutex irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 rename_lock.seqcount irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 &sb->s_type->i_lock_key#19 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 &s->s_inode_list_lock irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 &xa->xa_lock#8 irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 &fsnotify_mark_srcu irq_context: 0 (wq_completion)events free_ipc_work &type->s_umount_key#59 &dentry->d_lock/1 irq_context: 0 (wq_completion)events free_ipc_work sb_lock irq_context: 0 (wq_completion)events free_ipc_work unnamed_dev_ida.xa_lock irq_context: 0 (wq_completion)events free_ipc_work list_lrus_mutex irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#15 irq_context: 0 (wq_completion)events free_ipc_work pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work sb_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work sb_lock pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work mnt_id_ida.xa_lock irq_context: 0 (wq_completion)events free_ipc_work &ids->rwsem irq_context: 0 (wq_completion)events free_ipc_work (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)events free_ipc_work &ht->mutex irq_context: 0 (wq_completion)events free_ipc_work &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work percpu_counters_lock irq_context: 0 (wq_completion)events free_ipc_work pcpu_lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock sysctl_lock irq_context: 0 (wq_completion)events free_ipc_work rcu_read_lock &sb->s_type->i_lock_key#23 irq_context: 0 (wq_completion)events free_ipc_work &sb->s_type->i_lock_key#23 irq_context: 0 (wq_completion)events free_ipc_work &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock irq_context: 0 (wq_completion)events free_ipc_work rename_lock.seqcount irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock &sb->s_type->i_lock_key#23 &lru->node[i].lock irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)events free_ipc_work &s->s_inode_list_lock irq_context: 0 (wq_completion)events free_ipc_work &xa->xa_lock#8 irq_context: 0 (wq_completion)events free_ipc_work proc_inum_ida.xa_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (work_completion)(&data->suspend_work) irq_context: 0 &hdev->unregister_lock irq_context: 0 hci_dev_list_lock irq_context: 0 (work_completion)(&hdev->power_on) irq_context: 0 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (work_completion)(&hdev->reenable_adv_work) irq_context: 0 &hdev->cmd_sync_work_lock irq_context: 0 &hdev->req_lock (work_completion)(&(&hdev->interleave_scan)->work) irq_context: 0 &hdev->req_lock hci_dev_list_lock irq_context: 0 &hdev->req_lock (work_completion)(&hdev->tx_work) irq_context: 0 &hdev->req_lock (work_completion)(&hdev->rx_work) irq_context: 0 &hdev->req_lock &wq->mutex irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 &hdev->req_lock &wq->mutex &pool->lock irq_context: 0 &hdev->req_lock &wq->mutex &x->wait#10 irq_context: 0 &hdev->req_lock &hdev->lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &x->wait#2 irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock &obj_hash[i].lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &pcp->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock &base->lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 slock-AF_INET#2 &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&sdp->delay_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock &base->lock irq_context: 0 &hdev->req_lock &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 &hdev->req_lock &hdev->lock &x->wait#2 irq_context: 0 &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 oom_adj_mutex &rq->__lock irq_context: 0 sb_writers#4 oom_adj_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 oom_adj_mutex oom_adj_mutex.wait_lock irq_context: 0 sb_writers#4 oom_adj_mutex.wait_lock irq_context: 0 sb_writers#4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 &hdev->req_lock &hdev->lock pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 &hdev->req_lock &hdev->lock subsys mutex#74 irq_context: 0 &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock klist_remove_lock irq_context: 0 &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 &hdev->req_lock &hdev->lock &c->lock irq_context: 0 &hdev->req_lock &hdev->lock &____s->seqcount#2 irq_context: 0 &hdev->req_lock &hdev->lock &____s->seqcount irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 &hdev->req_lock tk_core.seq.seqcount irq_context: 0 &hdev->req_lock hci_sk_list.lock irq_context: 0 &hdev->req_lock &list->lock#5 irq_context: 0 &hdev->req_lock (work_completion)(&hdev->cmd_work) irq_context: 0 &hdev->req_lock (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 &hdev->lock irq_context: 0 &hdev->lock fs_reclaim irq_context: 0 &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &hdev->lock pool_lock#2 irq_context: 0 &hdev->lock tk_core.seq.seqcount irq_context: 0 &hdev->lock hci_sk_list.lock irq_context: 0 &hdev->lock &obj_hash[i].lock irq_context: 0 hci_sk_list.lock irq_context: 0 (work_completion)(&rfkill->uevent_work) irq_context: 0 (work_completion)(&rfkill->sync_work) irq_context: 0 subsys mutex#40 irq_context: 0 subsys mutex#40 &k->k_lock irq_context: 0 subsys mutex#40 &k->k_lock klist_remove_lock irq_context: 0 &rfkill->lock irq_context: 0 uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 uevent_sock_mutex nl_table_lock irq_context: 0 uevent_sock_mutex &c->lock irq_context: 0 uevent_sock_mutex &n->list_lock irq_context: 0 uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 uevent_sock_mutex nl_table_wait.lock irq_context: 0 subsys mutex#74 irq_context: 0 subsys mutex#74 &k->k_lock irq_context: 0 subsys mutex#74 &k->k_lock klist_remove_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 &sb->s_type->i_mutex_key#3 &xa->xa_lock#8 irq_context: 0 &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 &xa->xa_lock#17 irq_context: 0 hci_index_ida.xa_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex &tun->lock irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_NETROM irq_context: 0 &data->open_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock irq_context: 0 rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &sch->q.lock irq_context: 0 rtnl_mutex __ip_vs_mutex irq_context: 0 rtnl_mutex __ip_vs_mutex &ipvs->dest_trash_lock irq_context: 0 rtnl_mutex _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock irq_context: 0 rtnl_mutex flowtable_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &data->fib_event_queue_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#18 &dentry->d_lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->legacy_mutex &port->mutex &tpk_port.spinlock irq_context: 0 &xt[i].mutex remove_cache_srcu stock_lock irq_context: 0 &xt[i].mutex remove_cache_srcu key irq_context: 0 &xt[i].mutex remove_cache_srcu pcpu_lock irq_context: 0 &xt[i].mutex remove_cache_srcu percpu_counters_lock irq_context: 0 sb_writers#7 tomoyo_ss pool_lock#2 irq_context: 0 root_key_user.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock &pcp->lock &zone->lock irq_context: 0 key_construction_mutex keyring_name_lock irq_context: 0 sk_lock-AF_INET6 tcpv6_prot_mutex irq_context: 0 sk_lock-AF_INET6 tcpv6_prot_mutex &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock cpu_asid_lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock &n->list_lock irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 namespace_sem &____s->seqcount#2 irq_context: 0 rtnl_mutex &idev->mc_query_lock irq_context: 0 rtnl_mutex &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) irq_context: softirq rcu_callback &c->lock irq_context: 0 rtnl_mutex dev_base_lock &xa->xa_lock#4 irq_context: 0 rtnl_mutex cpu_hotplug_lock &list->lock#12 irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) pool_lock#2 irq_context: 0 rtnl_mutex bpf_devs_lock irq_context: 0 rtnl_mutex &hwstats->hwsdev_list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &tbl->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem class irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem (&tbl->proxy_timer) irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &base->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem krc.lock irq_context: 0 rtnl_mutex &ul->lock irq_context: 0 rtnl_mutex &net->xdp.lock irq_context: 0 rtnl_mutex mirred_list_lock irq_context: 0 rtnl_mutex &ul->lock#2 irq_context: 0 rtnl_mutex &ul->lock#2 pool_lock#2 irq_context: 0 rtnl_mutex &ul->lock#2 &dir->lock#2 irq_context: 0 rtnl_mutex &ul->lock#2 &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 rtnl_mutex &idev->mc_report_lock irq_context: 0 rtnl_mutex rcu_read_lock sysctl_lock irq_context: 0 rtnl_mutex rcu_read_lock &sb->s_type->i_lock_key#23 irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#23 irq_context: 0 rtnl_mutex &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 rtnl_mutex &dentry->d_lock irq_context: 0 rtnl_mutex rename_lock.seqcount irq_context: 0 rtnl_mutex rcu_read_lock &dentry->d_lock irq_context: 0 rtnl_mutex &dentry->d_lock &sb->s_type->i_lock_key#23 &lru->node[i].lock irq_context: 0 rtnl_mutex &dentry->d_lock &dentry->d_lock/1 irq_context: 0 rtnl_mutex &dentry->d_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &dentry->d_lock pool_lock#2 irq_context: 0 rtnl_mutex &s->s_inode_list_lock irq_context: 0 rtnl_mutex &xa->xa_lock#8 irq_context: 0 rtnl_mutex &fsnotify_mark_srcu irq_context: 0 rtnl_mutex &dentry->d_lock &lru->node[i].lock irq_context: 0 rtnl_mutex &pnn->pndevs.lock irq_context: 0 rtnl_mutex &pnn->routes.lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &sem->wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &sem->wait_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &p->pi_lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex subsys mutex#20 &k->k_lock klist_remove_lock irq_context: 0 rtnl_mutex deferred_probe_mutex irq_context: 0 rtnl_mutex device_links_lock irq_context: 0 rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_node_0 irq_context: 0 rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 tomoyo_ss stock_lock irq_context: 0 tomoyo_ss pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem &____s->seqcount#2 irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &rq->__lock irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dev_base_lock irq_context: 0 lweventlist_lock irq_context: 0 pernet_ops_rwsem nf_hook_mutex &c->lock irq_context: 0 pernet_ops_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem remove_cache_srcu irq_context: 0 pernet_ops_rwsem remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &xa->xa_lock#4 &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex kfence_freelist_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &____s->seqcount#2 irq_context: 0 sb_writers#4 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tun->lock irq_context: 0 &dir->lock#2 irq_context: 0 &dir->lock#2 &obj_hash[i].lock irq_context: 0 &dir->lock#2 pool_lock#2 irq_context: 0 netdev_unregistering_wq.lock irq_context: 0 &dir->lock &obj_hash[i].lock irq_context: 0 &dir->lock pool_lock#2 irq_context: 0 css_set_lock cgroup_file_kn_lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex &rq->__lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock tasklist_lock irq_context: 0 (wq_completion)netns irq_context: 0 (wq_completion)netns net_cleanup_work irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem net_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->nsid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 &group->mark_mutex remove_cache_srcu &c->lock irq_context: 0 &group->mark_mutex remove_cache_srcu &n->list_lock irq_context: 0 &group->mark_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &group->mark_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &group->mark_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &group->mark_mutex remove_cache_srcu &rq->__lock irq_context: 0 &group->mark_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->mark_mutex &lock->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &wg->socket_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &x->wait#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->dat.work)->timer irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->bla.work)->timer irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq rcu_callback proc_inum_ida.xa_lock irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock &obj_hash[i].lock irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock pool_lock#2 irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock krc.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &hash->list_locks[i] irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) key#20 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &entry->crc_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rq->__lock irq_context: 0 &group->mark_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-slock-AF_INET#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &table->hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &table->hash[i].lock &table->hash2[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-clock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &xa->xa_lock#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &fsnotify_mark_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock smack_known_lock.wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->endpoint_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &mm->mmap_lock remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &tn->node_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ebt_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &nft_net->commit_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_ct_ecache_mutex irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock irq_context: softirq (&sk->sk_timer) irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 tk_core.seq.seqcount irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 &obj_hash[i].lock irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 &base->lock irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 &base->lock &obj_hash[i].lock irq_context: 0 purge_vmap_area_lock quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 sb_internal jbd2_handle bit_wait_table + i irq_context: 0 ebt_mutex &cfs_rq->removed.lock irq_context: 0 ebt_mutex &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 rcu_state.exp_mutex pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem pcpu_lock stock_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx rfkill_global_mutex &n->list_lock irq_context: 0 misc_mtx rfkill_global_mutex &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 &p->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &n->list_lock &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rq->__lock irq_context: 0 rtnl_mutex &tbl->lock &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex triggers_list_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &c->lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock &c->lock irq_context: 0 &type->s_umount_key#46/1 &n->list_lock irq_context: 0 &type->s_umount_key#46/1 &n->list_lock &c->lock irq_context: 0 kn->active#49 &n->list_lock irq_context: 0 kn->active#49 &n->list_lock &c->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &local->filter_lock irq_context: 0 kn->active#50 &n->list_lock irq_context: 0 kn->active#50 &n->list_lock &c->lock irq_context: 0 nf_nat_proto_mutex &n->list_lock irq_context: 0 nf_nat_proto_mutex &n->list_lock &c->lock irq_context: softirq (&p->forward_delay_timer) irq_context: softirq (&p->forward_delay_timer) &br->lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &c->lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock pool_lock#2 irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock nl_table_lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &obj_hash[i].lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock nl_table_wait.lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &n->list_lock irq_context: softirq (&p->forward_delay_timer) &br->lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex kfence_freelist_lock irq_context: 0 file_rwsem &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET k-clock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET#2 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET#2 pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock key irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex &tbl->lock &n->list_lock &c->lock irq_context: 0 bt_proto_lock &n->list_lock irq_context: 0 bt_proto_lock &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &meta->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem batched_entropy_u8.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &mm->mmap_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#15 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#11 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#8 key#11 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &retval->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &nvmeq->sq_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &rq_wait->wait irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &__ctx->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &mapping->i_mmap_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem &____s->seqcount#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&hctx->run_work)->work) rcu_read_lock &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 &pl->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 &pl->lock key#12 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 key#14 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 key#11 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &xa->xa_lock#8 key#12 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_es_lock key#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_node_0 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: hardirq &ei->i_completed_io_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &retval->lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_es_lock key#2 irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sn->gssp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &cd->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem cache_list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem cache_list_lock &cd->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->can.stattimer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem vmap_area_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem purge_vmap_area_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem purge_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem xfrm_state_gc_work irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->xfrm.xfrm_state_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &hashinfo->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock sysctl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &sb->s_type->i_lock_key#23 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_lock_key#23 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rename_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock &sb->s_type->i_lock_key#23 &lru->node[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock &lru->node[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&net->ipv6.addr_chk_work)->work) irq_context: 0 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ip6_fl_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#18 rcu_read_lock rename_lock.seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex rcu_read_lock kfence_freelist_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->rules_mod_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem percpu_counters_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->ipv6.ip6_fib_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&mrt->ipmr_expire_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_app_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock &meta->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem __ip_vs_mutex irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&ipvs->dest_trash_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ipvs->expire_nodest_conn_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ipvs->defense_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ipvs->est_reload_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nfnl_subsys_ipset irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem recent_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hashlimit_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem trans_gc_work irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_log_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &nf_conntrack_locks[i] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &____s->seqcount#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex rcu_read_lock &nf_nat_locks[i] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&cnet->ecache.dwork)->work) irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sn->pipefs_sb_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem tcp_metrics_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_INET irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx &____s->seqcount#2 irq_context: 0 rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&net->xfrm.policy_hash_work) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->xfrm.xfrm_policy_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&net->xfrm.state_hash_work) irq_context: 0 &p->lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xa->xa_lock#4 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem genl_sk_destructing_waitq.lock irq_context: softirq rcu_read_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rq->__lock irq_context: 0 &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx misc_mtx.wait_lock irq_context: 0 misc_mtx nfc_devlist_mutex &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 misc_mtx.wait_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)events fqdir_free_work &p->pi_lock irq_context: 0 rtnl_mutex lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &____s->seqcount#2 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2 &____s->seqcount irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events fqdir_free_work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex nfc_devlist_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex sched_map-wait-type-override &pool->lock irq_context: 0 misc_mtx nfc_devlist_mutex dpm_list_mtx &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex.wait_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &zone->lock &____s->seqcount irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &p->pi_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&w->w) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &n->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override pool_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &n->list_lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &c->lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &n->list_lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &obj_hash[i].lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &rq->__lock irq_context: softirq rcu_callback &x->wait#2 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work &dir->lock irq_context: 0 (wq_completion)netns net_cleanup_work &dir->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work &dir->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work &obj_hash[i].lock pool_lock irq_context: 0 tomoyo_ss remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock stock_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock key irq_context: 0 &vma->vm_lock->lock rcu_read_lock pcpu_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock percpu_counters_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &dev->mutex &n->list_lock irq_context: 0 &dev->mutex &n->list_lock &c->lock irq_context: 0 &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 &dev->mutex rfkill_global_mutex &rq->__lock irq_context: 0 &dev->mutex rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rfkill_global_mutex.wait_lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work pool_lock irq_context: 0 nfc_devlist_mutex &sem->wait_lock irq_context: 0 nfc_devlist_mutex &p->pi_lock irq_context: 0 nfc_devlist_mutex &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &u->iolock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex lock kernfs_idr_lock &c->lock irq_context: 0 nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 stock_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 &dev->mutex remove_cache_srcu irq_context: 0 &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 &dev->mutex remove_cache_srcu &c->lock irq_context: 0 &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &dev->mutex remove_cache_srcu pool_lock#2 irq_context: 0 nfc_devlist_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 nfc_devlist_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#18 &dentry->d_lock &wq irq_context: 0 &type->i_mutex_dir_key/1 &group->notification_waitq &ep->lock &ep->wq irq_context: 0 &type->i_mutex_dir_key/1 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key/1 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#18 &sb->s_type->i_lock_key#32 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 rtnl_mutex gdp_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE rose_list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &type->i_mutex_dir_key/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock pool_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 misc_mtx hrtimer_bases.lock irq_context: 0 misc_mtx hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 misc_mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex &c->lock irq_context: 0 misc_mtx &dev->mutex &n->list_lock irq_context: 0 misc_mtx &dev->mutex &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex &sem->wait_lock irq_context: 0 tasklist_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 percpu_counters_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 pcpu_lock stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &n->list_lock &c->lock irq_context: 0 tomoyo_ss remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &sem->wait_lock irq_context: 0 sb_writers#3 &sem->wait_lock irq_context: 0 rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &p->pi_lock irq_context: 0 sb_writers#3 &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock &zone->lock irq_context: softirq rcu_read_lock rcu_read_lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock _xmit_NONE#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#3 rcu_read_lock rcu_node_0 irq_context: 0 &dev->mutex &sem->wait_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &group->mark_mutex &n->list_lock irq_context: 0 &group->mark_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &meta->lock irq_context: 0 &mm->mmap_lock sb_writers#5 &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pcpu_alloc_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &n->list_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 pcpu_alloc_mutex.wait_lock irq_context: 0 sb_writers#3 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#3 rcu_node_0 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem netns_bpf_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &x->wait#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->fs_probe_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->cells_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->cells_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem bit_wait_table + i irq_context: 0 (wq_completion)afs irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) &net->cells_lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i &p->pi_lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&net->fs_timer) irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) &(&net->fs_lock)->lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i &p->pi_lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC k-slock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rx->incoming_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->conn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &call->waitq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC rcu_read_lock &call->notify_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC (rxrpc_call_limiter).lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rx->recvmsg_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rx->call_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->call_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC (&call->timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &list->lock#17 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)kafsd irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &local->services_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)krxrpcd irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rlock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &x->wait irq_context: 0 &rxnet->local_mutex irq_context: 0 (&local->client_conn_reap_timer) irq_context: 0 &rxnet->conn_lock irq_context: 0 &table->hash[i].lock irq_context: 0 &table->hash[i].lock &table->hash2[i].lock irq_context: 0 k-clock-AF_INET6 irq_context: 0 &list->lock#18 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_lock_key#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xa->xa_lock#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_subdir_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ent->pde_unload_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem proc_inum_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex (work_completion)(&data->gc_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex net_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&ovs_net->masks_rebalance)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&ovs_net->dp_notify_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &srv->idr_lock irq_context: 0 &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex &wq->mutex &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock (worker)->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock (worker)->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->service_conn_reaper) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC k-slock-AF_TIPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &nt->cluster_scope_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC k-clock-AF_TIPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_TIPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ptype_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&tn->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_callback rlock-AF_RXRPC irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &tn->nametbl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&c->work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)krdsd irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rtn->rds_tcp_accept_w) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &h->lhash2[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &icsk->icsk_accept_queue.rskq_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 k-slock-AF_INET6 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rds_tcp_conn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem loop_conns_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)l2tp irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &x->wait#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&rxnet->peer_keepalive_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rxnet->peer_keepalive_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&rxnet->service_conn_reap_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &x->wait#10 irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->service_conn_reaper) &rxnet->conn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rxnet->conn_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex &br->hash_lock &n->list_lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &c->lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &n->list_lock irq_context: 0 rtnl_mutex &br->lock lweventlist_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex netpoll_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pn->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tn->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_NONE#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_ETHER#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sch->q.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &sch->q.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex &ipvs->dest_trash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &im->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex class irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&tbl->proxy_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex flowtable_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock rt6_exception_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ifa->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tb->tb6_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#5 rcu_node_0 irq_context: 0 sb_writers#5 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.fib6_gc_lock rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.fib6_gc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fib_info_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fib_info_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fib_info_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &tbl->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem class irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem (&tbl->proxy_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->xdp.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex mirred_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &nft_net->commit_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex proc_subdir_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ent->pde_unload_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_report_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->routes.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnettable->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex target_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_NONE irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &x->wait#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dpm_list_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex deferred_probe_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex device_links_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_base_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem lweventlist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem napi_hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem netdev_unregistering_wq.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 sb_internal mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_TUNNEL6#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_TUNNEL6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 rtnl_mutex _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex _xmit_ETHER &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &table->lock#3 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 &sb->s_type->i_mutex_key#18 &sb->s_type->i_lock_key#32 &dentry->d_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: softirq rcu_callback &x->wait#24 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal batched_entropy_u8.lock irq_context: 0 sb_writers#3 sb_internal kfence_freelist_lock irq_context: 0 sb_writers#3 sb_internal &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_SIT#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_SIT irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_node_0 irq_context: 0 key_types_sem batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex __ip_vs_mutex pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex key#24 irq_context: 0 rtnl_mutex &wg->device_update_lock &n->list_lock irq_context: 0 rtnl_mutex &wg->device_update_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#74 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock &net->can.rcvlists_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock &obj_hash[i].lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &____s->seqcount#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex lweventlist_lock &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock percpu_counters_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock &cfs_rq->removed.lock irq_context: 0 wq_pool_attach_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &root->kernfs_rwsem stock_lock irq_context: 0 &root->kernfs_rwsem key irq_context: 0 &root->kernfs_rwsem pcpu_lock irq_context: 0 &root->kernfs_rwsem percpu_counters_lock irq_context: 0 &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock key irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock stock_lock irq_context: 0 &root->kernfs_iattr_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock &priv->lock irq_context: 0 sk_lock-AF_CAN &n->list_lock irq_context: 0 sk_lock-AF_CAN &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#5 &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_node_0 irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &p->lock fs_reclaim &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &cfs_rq->removed.lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 &n->list_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 &sighand->siglock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &cfs_rq->removed.lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock rcu_node_0 irq_context: 0 &type->s_umount_key#30 &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock &base->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events free_ipc_work sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock &base->lock irq_context: 0 rtnl_mutex &wg->device_update_lock &rnp->exp_lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_TUNNEL#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_TUNNEL irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_node_0 irq_context: 0 &xt[i].mutex remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 &xt[i].mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock nl_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_node_0 irq_context: 0 rtnl_mutex pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &xt[i].mutex quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#16 &c->lock irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#18 &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle key#4 irq_context: softirq rcu_read_lock hwsim_radio_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_INET &rq->__lock cpu_asid_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &rq->__lock cpu_asid_lock irq_context: 0 &dentry->d_lock &sb->s_type->i_lock_key#24 &lru->node[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &c->lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &rq->__lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex rtnl_mutex.wait_lock irq_context: 0 vlan_ioctl_mutex &p->pi_lock irq_context: 0 vlan_ioctl_mutex &p->pi_lock &rq->__lock irq_context: 0 vlan_ioctl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_IPGRE#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_IPGRE krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: hardirq hrtimer_bases.lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &fn->fou_lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &meta->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 batched_entropy_u8.lock irq_context: 0 &type->i_mutex_dir_key#5 kfence_freelist_lock irq_context: softirq (&peer->timer_persistent_keepalive) &n->list_lock irq_context: softirq (&peer->timer_persistent_keepalive) &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->sync_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &xa->xa_lock#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (inetaddr_chain).rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem inet6addr_chain.lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &wdev->pmsr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->wiphy_work_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &fq->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &xa->xa_lock#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &xa->xa_lock#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &local->active_txq_lock[i] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (work_completion)(&sta->drv_deliver_wk) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ifibss->incomplete_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx hrtimer_bases.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx hrtimer_bases.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &data->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &local->queue_stop_reason_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &fq->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &local->queue_stop_reason_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &x->wait#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &sighand->siglock &n->list_lock irq_context: 0 rcu_read_lock &sighand->siglock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&ifibss->timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&wdev->disconnect_wk) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&wdev->pmsr_free_wk) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &____s->seqcount#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx _xmit_ETHER &local->filter_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&local->dynamic_ps_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&dwork->timer)#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&dwork->timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &list->lock#16 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 &sig->cred_guard_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &wq->mutex &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->dev_wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &local->iflist_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem pool_lock#2 irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_base_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_base_lock &xa->xa_lock#4 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx bpf_devs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx net_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx net_rwsem &list->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &in_dev->mc_tomb_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &tbl->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx class irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&tbl->proxy_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ul->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &net->xdp.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx mirred_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &nft_net->commit_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &tn->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &tb->tb6_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx proc_subdir_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ent->pde_unload_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx proc_inum_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &net->ipv6.addrconf_hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ndev->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &ndev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_query_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_report_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &pnn->pndevs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &pnn->routes.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &pnettable->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx smc_ib_devices.mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx target_list_lock irq_context: 0 tomoyo_ss mount_lock irq_context: 0 tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (work_completion)(&(&link->color_collision_detect_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_hotplug_mutex &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &x->wait#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dpm_list_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx deferred_probe_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx device_links_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock xps_map_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->mgmt_registrations_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &wdev->event_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fq->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (&dwork->timer)#4 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock nsim_bus_dev_ids.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock subsys mutex#76 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &k->k_lock klist_remove_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex device_links_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &k->list_lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &rq->__lock irq_context: 0 rcu_read_lock stock_lock irq_context: 0 rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)events free_ipc_work sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events free_ipc_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events free_ipc_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss &pcp->lock &zone->lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &kcov->lock kcov_remote_lock &c->lock irq_context: 0 &kcov->lock kcov_remote_lock &n->list_lock irq_context: 0 &kcov->lock kcov_remote_lock &n->list_lock &c->lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 &data->open_mutex lock kernfs_idr_lock &c->lock irq_context: 0 &data->open_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 &data->open_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->work_lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)bat_events pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 pcpu_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG fs_reclaim irq_context: 0 sk_lock-AF_ALG fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_ALG pool_lock#2 irq_context: 0 sk_lock-AF_ALG &mm->mmap_lock irq_context: 0 misc_mtx pcpu_alloc_mutex.wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_node_0 irq_context: 0 sk_lock-AF_ALG &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ALG &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &obj_hash[i].lock pool_lock irq_context: softirq (&lapb->t1timer) &lapb->lock &list->lock#20 irq_context: softirq &list->lock#20 irq_context: softirq rcu_read_lock x25_neigh_list_lock irq_context: softirq rcu_read_lock x25_list_lock irq_context: softirq rcu_read_lock x25_forward_list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &sem->wait_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex rcu_node_0 irq_context: 0 &fsnotify_mark_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &journal->j_wait_done_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#18 &xa->xa_lock#15 &c->lock irq_context: 0 &sb->s_type->i_mutex_key#18 &xa->xa_lock#15 &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#18 &xa->xa_lock#15 &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 percpu_counters_lock irq_context: 0 &data->open_mutex subsys mutex#74 &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex triggers_list_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 pcpu_lock stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &____s->seqcount irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#18 &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock stock_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock &obj_hash[i].lock irq_context: 0 wq_pool_attach_mutex &rq->__lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock key irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock pcpu_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 &n->list_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock percpu_counters_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal quarantine_lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 &obj_hash[i].lock pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->work_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock &n->list_lock &c->lock irq_context: 0 wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_node_0 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &c->lock irq_context: 0 rcu_read_lock rcu_read_lock &____s->seqcount#4 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex console_owner_lock irq_context: 0 rtnl_mutex console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu pcpu_lock stock_lock irq_context: 0 sk_lock-AF_ALG &rq->__lock irq_context: 0 sk_lock-AF_ALG &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NFC &rq->__lock irq_context: 0 &tsk->futex_exit_mutex &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim &obj_hash[i].lock irq_context: 0 sb_writers#3 &base->lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex &base->lock irq_context: 0 cb_lock genl_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex pool_lock#2 irq_context: 0 sb_writers#7 &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &base->lock irq_context: 0 subsys mutex#37 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &n->list_lock irq_context: 0 &type->i_mutex_dir_key#5 stock_lock irq_context: 0 &type->i_mutex_dir_key#5 key irq_context: 0 (wq_completion)events key_gc_work sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &____s->seqcount irq_context: 0 rtnl_mutex &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq init_task.mems_allowed_seq.seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#4 &n->list_lock irq_context: 0 sb_writers#4 &n->list_lock &c->lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &n->list_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu stock_lock irq_context: 0 remove_cache_srcu key irq_context: 0 remove_cache_srcu pcpu_lock irq_context: 0 remove_cache_srcu percpu_counters_lock irq_context: 0 remove_cache_srcu pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&local->restart_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &list->lock#16 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rdev->wiphy.mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rfkill->uevent_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rfkill->sync_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#40 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#40 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#40 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &x->wait#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dpm_list_mtx irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem deferred_probe_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem device_links_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rfkill->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &c->lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem triggers_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem leds_list_lock irq_context: 0 cb_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx subsys mutex#54 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->wiphy_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->conn_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->event_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&rdev->dfs_update_channels_wk)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&(&rdev->background_cac_done_wk)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->destroy_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->propagate_radar_detect_wk) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->propagate_cac_done_wk) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->mgmt_registrations_update_wk) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (work_completion)(&rdev->background_cac_abort_wk) irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex device_links_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#53 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#53 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem subsys mutex#53 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem gdp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (&local->sta_cleanup) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rq->__lock cpu_asid_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_QIPCRTR irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &root->kernfs_rwsem rcu_node_0 irq_context: 0 &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &obj_hash[i].lock pool_lock irq_context: 0 &p->lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 rtnl_mutex &tn->lock &rq->__lock irq_context: 0 rtnl_mutex _xmit_ETHER &____s->seqcount#2 irq_context: 0 rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock pool_lock#2 irq_context: 0 misc_mtx quarantine_lock irq_context: 0 &vma->vm_lock->lock &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &base->lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 &child->perf_event_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 &vma->vm_lock->lock remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem pool_lock#2 irq_context: 0 sb_writers#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &____s->seqcount irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_netgroup_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets_rwsem rdma_nets.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem devices_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-clock-AF_NETLINK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &nlk->wait irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rdma_nets.xa_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &dev->mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &obj_hash[i].lock irq_context: 0 &dev->mutex uevent_sock_mutex.wait_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex.wait_lock irq_context: 0 rtnl_mutex uevent_sock_mutex.wait_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 remove_cache_srcu rcu_node_0 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &n->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex nfc_devlist_mutex.wait_lock irq_context: 0 nfc_devlist_mutex &rq->__lock irq_context: softirq (&peer->timer_send_keepalive) irq_context: softirq (&peer->timer_send_keepalive) pool_lock#2 irq_context: softirq (&peer->timer_send_keepalive) &list->lock#14 irq_context: softirq (&peer->timer_send_keepalive) tk_core.seq.seqcount irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &r->producer_lock#2 irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &obj_hash[i].lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &base->lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: softirq (&peer->timer_send_keepalive) &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_QIPCRTR &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_QIPCRTR slock-AF_QIPCRTR irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_QIPCRTR clock-AF_QIPCRTR irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_QIPCRTR irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx batched_entropy_u8.lock crngs.lock irq_context: 0 nfc_devlist_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &journal->j_state_lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &ret->b_state_lock bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 slock-AF_ALG irq_context: 0 misc_mtx nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 cb_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->endpoint_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex gdp_mutex &____s->seqcount#2 irq_context: 0 rtnl_mutex gdp_mutex &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#30 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &lg->lg_mutex rcu_read_lock &pa->pa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &xa->xa_lock#18 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &app->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&app->join_timer)#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&app->periodic_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &list->lock#11 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&app->join_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &app->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &list->lock#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&priv->scan_result)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&priv->connect)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&hsr->prune_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&hsr->announce_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &xa->xa_lock#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &fsnotify_mark_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &xa->xa_lock#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dentry->d_lock &dentry->d_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dentry->d_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock mount_lock mount_lock.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex mount_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex mount_lock mount_lock.seqcount irq_context: 0 rtnl_mutex gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &dev->mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (console_sem).lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_lock console_srcu console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &x->wait#2 irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->softif_vlan_list_lock kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &meta->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &hsr->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bat_priv->tt.commit_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex ptype_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &entry->crc_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex key#19 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bat_priv->forw_bcast_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bat_priv->forw_bat_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&forw_packet_aggr->delayed_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 ebt_mutex &mm->mmap_lock &rq->__lock irq_context: 0 ebt_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bat_priv->tt.changes_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock deferred_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock (console_sem).lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock console_lock console_srcu console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock deferred_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock &c->lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &br->multicast_lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) (&p->rexmit_timer) irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) (&p->timer) irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) pool_lock#2 irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock irq_context: softirq (&mp->timer) irq_context: softirq (&mp->timer) &br->multicast_lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rhashtable_bucket irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) (&mp->timer) irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 percpu_counters_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (switchdev_blocking_notif_chain).rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&pmctx->ip6_mc_router_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&pmctx->ip4_mc_router_timer) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&ht->run_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex deferred_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->multicast_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex pcpu_lock irq_context: 0 sb_writers#4 &rq->__lock irq_context: 0 sb_writers#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock lweventlist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock lweventlist_lock &dir->lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback put_task_map-wait-type-override per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 cb_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&br->gc_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex j1939_netdev_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &vlan_netdev_xmit_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &batadv_netdev_xmit_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key#3 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock &qdisc_xmit_lock_key#4 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock _xmit_LOOPBACK#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 &sch->q.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-sk_lock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-slock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &table->hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &table->hash[i].lock &table->hash2[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-clock-AF_INET6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &sb->s_type->i_lock_key#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-sk_lock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-slock-AF_INET#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex k-clock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER (console_sem).lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &list->lock#14 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (&peer->timer_retransmit_handshake) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (&peer->timer_send_keepalive) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (&peer->timer_new_handshake) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (&peer->timer_zero_key_material) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (&peer->timer_persistent_keepalive) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock (work_completion)(&peer->clear_peer_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &handshake->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &r->consumer_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->socket_update_lock irq_context: 0 tomoyo_ss batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &nn->netlink_tap_lock irq_context: 0 &lo->lo_mutex &rq->__lock irq_context: 0 &lo->lo_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->mii_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->arp_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->alb_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->ad_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->mcast_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&bond->slave_arr_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&br->hello_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&br->topology_change_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&br->tcn_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip4_mc_router_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip4_other_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip4_other_query.delay_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip4_own_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip6_mc_router_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip6_other_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip6_other_query.delay_timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (&brmctx->ip6_own_query.timer) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &im->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 krc.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock pool_lock#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem key irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem percpu_counters_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 sb_writers#6 rcu_node_0 irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &lock->wait_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback &n->list_lock irq_context: softirq rcu_callback &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &k->k_lock klist_remove_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock dpm_list_mtx irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock deferred_probe_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock device_links_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 &wb->list_lock &sb->s_type->i_lock_key#24 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 rcu_node_0 irq_context: 0 sb_writers#3 &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &rq->__lock cpu_asid_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex raw_notifier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bcm_notifier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex isotp_notifier_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &batadv_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &batadv_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &batadv_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &bond->stats_lock/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &meta->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex napi_hash_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex req_lock irq_context: 0 &type->i_mutex_dir_key/1 irq_context: 0 &type->i_mutex_dir_key/1 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 pool_lock#2 irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss pool_lock#2 irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss rcu_read_lock &dentry->d_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock mmu_notifier_invalidate_range_start irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss tomoyo_policy_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &x->wait#11 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#75 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#75 &k->k_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#75 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex kernfs_idr_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tap_major->minor_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&port->bc_work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &ipvlan->addrs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#3/1 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nf_hook_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nf_hook_mutex cpu_hotplug_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&port->wq) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &k->k_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) pool_lock irq_context: 0 &xt[i].mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex fs_reclaim &obj_hash[i].lock irq_context: 0 &xt[i].mutex fs_reclaim pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 krc.lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex remove_cache_srcu &____s->seqcount irq_context: 0 cb_lock genl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#7 kn->active#4 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fib_info_lock &obj_hash[i].lock pool_lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_node_0 irq_context: 0 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 &ep->mtx &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock irq_context: 0 kn->active#4 rcu_node_0 irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex pool_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_hotplug_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex input_pool.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&slave->notify_work)->work) irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &bond->ipsec_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex crngs.lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq (&lapb->t1timer) &lapb->lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_NETROM#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &sem->wait_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem pcpu_lock stock_lock irq_context: 0 &hdev->lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &ndev->lock &base->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &____s->seqcount irq_context: 0 rtnl_mutex &ndev->lock &base->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &table->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &x->wait#17 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq irq_context: 0 &type->i_mutex_dir_key#5 &cfs_rq->removed.lock irq_context: softirq rcu_read_lock &br->hash_lock &n->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (&peer->timer_retransmit_handshake) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (&peer->timer_send_keepalive) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (&peer->timer_new_handshake) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (&peer->timer_zero_key_material) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (&peer->timer_persistent_keepalive) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (work_completion)(&peer->clear_peer_work) irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wq->mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wq->mutex &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wq->mutex &x->wait#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock napi_hash_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex devnet_rename_sem &sem->wait_lock irq_context: 0 rtnl_mutex devnet_rename_sem &p->pi_lock irq_context: 0 rtnl_mutex devnet_rename_sem &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &wq->mutex &rq->__lock irq_context: 0 &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 &c->lock irq_context: 0 nfc_devlist_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 kfence_freelist_lock irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim &rq->__lock irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex quarantine_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_node_0 irq_context: 0 rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock xps_map_mutex &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 quarantine_lock irq_context: 0 rtnl_mutex _xmit_IPGRE &c->lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &cfs_rq->removed.lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->keypairs.keypair_update_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex stock_lock irq_context: 0 rtnl_mutex pcpu_lock stock_lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &cfs_rq->removed.lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work pool_lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &c->lock irq_context: softirq (&timer) irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &x->wait#2 irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &br->hash_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex netpoll_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_node_0 irq_context: 0 &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex nfc_devlist_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock quarantine_lock irq_context: 0 &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex.wait_lock irq_context: 0 misc_mtx nfc_devlist_mutex stock_lock irq_context: 0 misc_mtx nfc_devlist_mutex key irq_context: 0 misc_mtx nfc_devlist_mutex pcpu_lock irq_context: 0 misc_mtx nfc_devlist_mutex percpu_counters_lock irq_context: 0 misc_mtx nfc_devlist_mutex pcpu_lock stock_lock irq_context: 0 misc_mtx nfc_devlist_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &obj_hash[i].lock irq_context: softirq rcu_read_lock hwsim_radio_lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override pool_lock irq_context: softirq rcu_read_lock hwsim_radio_lock &zone->lock &____s->seqcount irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &meta->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex &____s->seqcount#2 irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &rq->__lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &____s->seqcount#2 irq_context: softirq (&peer->timer_send_keepalive) &n->list_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex netpoll_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex.wait_lock irq_context: 0 rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 krc.lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 rtnl_mutex &batadv_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex kn->active#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (debug_obj_work).work quarantine_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock batched_entropy_u8.lock irq_context: 0 &p->lock kfence_freelist_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET fs_reclaim &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rnp->exp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rnp->exp_wq[2] irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex rcu_node_0 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_ROSE rlock-AF_ROSE irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 misc_mtx &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &dev->mutex rfkill_global_mutex rfkill_global_mutex.wait_lock irq_context: 0 &dev->mutex rfkill_global_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &q->sysfs_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pcpu_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->wiphy_work_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &____s->seqcount irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &base->lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 fill_pool_map-wait-type-override &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#7 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#86 irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 rtnl_mutex &macsec_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock key irq_context: 0 &data->open_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex pin_fs_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 kfence_freelist_lock irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#46 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &base->lock &obj_hash[i].lock irq_context: 0 put_task_map-wait-type-override#2 &obj_hash[i].lock irq_context: 0 put_task_map-wait-type-override#2 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macvlan_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem wq_pool_mutex &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 misc_mtx remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 misc_mtx remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 misc_mtx remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &c->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 &data->open_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 &tsk->futex_exit_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock irq_context: 0 &tsk->futex_exit_mutex &rcu_state.expedited_wq irq_context: 0 &tsk->futex_exit_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 misc_mtx remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock &rq->__lock irq_context: 0 &ndev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &c->lock irq_context: 0 misc_mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 sb_internal &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss kfence_freelist_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 sb_internal &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu pool_lock#2 irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 pool_lock#2 irq_context: 0 &ep->mtx stock_lock irq_context: 0 &ep->mtx key irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mount_lock irq_context: 0 &ep->mtx pcpu_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock lweventlist_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&n->timer) irq_context: softirq (&n->timer) &n->lock irq_context: softirq (&n->timer) &n->lock &obj_hash[i].lock irq_context: softirq (&n->timer) &n->lock &base->lock irq_context: softirq (&n->timer) &n->lock &base->lock &obj_hash[i].lock irq_context: 0 &ep->mtx percpu_counters_lock irq_context: 0 &ep->mtx pcpu_lock stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 kfence_freelist_lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#88 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fsnotify_mark_srcu key irq_context: softirq (&peer->timer_retransmit_handshake) &peer->endpoint_lock &obj_hash[i].lock irq_context: softirq (&peer->timer_retransmit_handshake) &peer->endpoint_lock pool_lock#2 irq_context: 0 &fsnotify_mark_srcu pcpu_lock irq_context: 0 &fsnotify_mark_srcu percpu_counters_lock irq_context: 0 &fsnotify_mark_srcu pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal rcu_node_0 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#88 &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#101 irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#101 &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.gp_wq irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.gp_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.expedited_wq irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx remove_cache_srcu pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &meta->lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem (&timer.timer) irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem fw_lock &x->wait#22 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu pool_lock#2 irq_context: softirq (&lapb->t1timer) &lapb->lock &n->list_lock irq_context: softirq (&lapb->t1timer) &lapb->lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 lweventlist_lock irq_context: 0 (wq_completion)events key_gc_work &type->lock_class &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh key#20 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &n->list_lock &c->lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#3 jbd2_handle &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex lweventlist_lock kfence_freelist_lock irq_context: 0 &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 &root->kernfs_iattr_rwsem pool_lock#2 irq_context: 0 misc_mtx nfc_devlist_mutex &n->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex &n->list_lock &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &peer->endpoint_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &lock->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock batched_entropy_u8.lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &n->list_lock irq_context: 0 sb_writers#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 fill_pool_map-wait-type-override pool_lock irq_context: 0 &vma->vm_lock->lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock fs_reclaim &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events &rq->__lock irq_context: 0 (wq_completion)bat_events &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &n->list_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &p->pi_lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &q->sysfs_dir_lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem kfence_freelist_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[3] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &____s->seqcount irq_context: 0 nfc_devlist_mutex remove_cache_srcu irq_context: 0 nfc_devlist_mutex remove_cache_srcu quarantine_lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &c->lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &n->list_lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &rq->__lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &group->mark_mutex &cfs_rq->removed.lock irq_context: 0 &group->mark_mutex &obj_hash[i].lock irq_context: 0 &p->lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu rcu_node_0 irq_context: 0 nfc_devlist_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 nfc_devlist_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &fsnotify_mark_srcu remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 &dev->mutex kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &dev->mutex kernfs_idr_lock pool_lock#2 irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &type->s_umount_key#66 &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#66 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &type->s_umount_key#66 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle pool_lock#2 irq_context: 0 &xt[i].mutex &base->lock irq_context: 0 &xt[i].mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &q->sysfs_dir_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &bgl->locks[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &c->lock irq_context: 0 &q->debugfs_mutex &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &lock->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &c->lock irq_context: 0 kn->active#50 &rq->__lock irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events key_gc_work &type->lock_class &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#88 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#8 irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#975 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#975 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#228 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#224 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#223 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#968 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#962 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tn->lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx nfc_devlist_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#539 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &x->wait#10 irq_context: 0 sk_lock-AF_NFC &c->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &table->lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &____s->seqcount irq_context: 0 &p->lock &of->mutex kn->active#4 batched_entropy_u8.lock irq_context: 0 &p->lock &of->mutex kn->active#4 kfence_freelist_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &meta->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex pool_lock#2 irq_context: 0 &disk->open_mutex &lo->lo_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &obj_hash[i].lock pool_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inet6addr_validator_chain).rwsem &rq->__lock irq_context: 0 rtnl_mutex (inet6addr_validator_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_CAN &mm->mmap_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex batched_entropy_u8.lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq irq_context: 0 misc_mtx nfc_devlist_mutex kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &c->lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#539 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: softirq (&mp->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &rcu_state.expedited_wq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &ht->mutex &rq->__lock irq_context: 0 &sb->s_type->i_lock_key#20 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tn->lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_log_mutex &rq->__lock irq_context: 0 &hdev->lock &n->list_lock irq_context: 0 &hdev->lock &n->list_lock &c->lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex kfence_freelist_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &meta->lock irq_context: 0 &group->mark_mutex fs_reclaim &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_log_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &fsnotify_mark_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_log_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 &obj_hash[i].lock pool_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &dev->mutex stock_lock irq_context: 0 &dev->mutex key irq_context: 0 &dev->mutex pcpu_lock irq_context: 0 &dev->mutex percpu_counters_lock irq_context: 0 &dev->mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_log_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_log_mutex pool_lock#2 irq_context: 0 &type->s_umount_key#66 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &type->s_umount_key#66 &dentry->d_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#66 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#66 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events free_ipc_work sched_map-wait-type-override &rq->__lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) quarantine_lock irq_context: 0 loop_validate_mutex &rq->__lock irq_context: 0 loop_validate_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock &mm->page_table_lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &kernfs_locks->open_file_mutex[count] fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_gp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 &dev->mutex uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override pool_lock irq_context: softirq (&timer) &obj_hash[i].lock irq_context: softirq (&timer) &base->lock irq_context: softirq (&timer) &base->lock &obj_hash[i].lock irq_context: softirq (&timer) rcu_read_lock pool_lock#2 irq_context: softirq (&timer) rcu_read_lock &c->lock irq_context: softirq (&timer) rcu_read_lock &____s->seqcount#2 irq_context: softirq (&timer) rcu_read_lock &____s->seqcount irq_context: softirq (&timer) rcu_read_lock &n->list_lock irq_context: softirq (&timer) rcu_read_lock &n->list_lock &c->lock irq_context: softirq (&timer) &txlock irq_context: softirq (&timer) &txlock &list->lock#3 irq_context: softirq (&timer) &txwq irq_context: softirq (&timer) &txwq &p->pi_lock irq_context: softirq (&timer) &txwq &p->pi_lock &rq->__lock irq_context: softirq (&timer) &txwq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_persistent_keepalive) &____s->seqcount#2 irq_context: softirq (&peer->timer_persistent_keepalive) &____s->seqcount irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&peer->timer_send_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 rcu_read_lock_bh &list->lock#12 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 &eql->queue.lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 &eql->queue.lock &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_SLIP#2 &eql->queue.lock pool_lock#2 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM (console_sem).lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner &port_lock_key irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_lock console_srcu console_owner console_owner_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM pool_lock#2 irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key nr_node_list_lock irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh &nr_netdev_xmit_lock_key pool_lock#2 irq_context: 0 rcu_read_lock_bh _xmit_X25#2 irq_context: 0 rcu_read_lock_bh _xmit_X25#2 &lapbeth->up_lock irq_context: 0 rcu_read_lock_bh _xmit_X25#2 &lapbeth->up_lock &obj_hash[i].lock irq_context: 0 rcu_read_lock_bh _xmit_X25#2 &lapbeth->up_lock pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 &vma->vm_lock->lock mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock key irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock pcpu_lock stock_lock irq_context: 0 &xt[i].mutex fs_reclaim stock_lock irq_context: 0 &xt[i].mutex fs_reclaim key irq_context: 0 cb_lock genl_mutex &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex fs_reclaim pcpu_lock irq_context: 0 &xt[i].mutex fs_reclaim percpu_counters_lock irq_context: 0 sb_writers#3 jbd2_handle rcu_node_0 irq_context: 0 sb_writers#3 jbd2_handle &rcu_state.expedited_wq irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 misc_mtx stock_lock irq_context: 0 misc_mtx key irq_context: 0 misc_mtx pcpu_lock irq_context: 0 misc_mtx percpu_counters_lock irq_context: 0 misc_mtx pcpu_lock stock_lock irq_context: 0 &xt[i].mutex fs_reclaim pcpu_lock stock_lock irq_context: 0 br_ioctl_mutex rtnl_mutex fs_reclaim irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock key irq_context: 0 &vma->vm_lock->lock pcpu_lock irq_context: 0 &vma->vm_lock->lock percpu_counters_lock irq_context: 0 &vma->vm_lock->lock pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock pcpu_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 br_ioctl_mutex rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock (console_sem).lock irq_context: 0 cb_lock console_lock console_srcu console_owner_lock irq_context: 0 &disk->open_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 cb_lock console_lock console_srcu console_owner irq_context: 0 rtnl_mutex lweventlist_lock &n->list_lock irq_context: 0 rtnl_mutex lweventlist_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex gdp_mutex &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER/1 &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &____s->seqcount#2 irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_send_keepalive) &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[2] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback put_task_map-wait-type-override stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &base->lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 &obj_hash[i].lock irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) rcu_read_lock rcu_node_0 irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#319 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#536 irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET#2 rcu_read_lock &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3/1 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pcp->lock &zone->lock irq_context: 0 &fc->uapi_mutex &pcp->lock &zone->lock irq_context: 0 rcu_state.exp_mutex.wait_lock irq_context: 0 &fc->uapi_mutex uuid_mutex irq_context: 0 uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock lock kernfs_idr_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 rcu_read_lock _xmit_ETHER &____s->seqcount#2 irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx rcu_node_0 irq_context: 0 misc_mtx &rcu_state.expedited_wq irq_context: 0 misc_mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rq->__lock cpu_asid_lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) rcu_node_0 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex nbd_index_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &rq->__lock irq_context: softirq rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sk_lock-AF_NFC slock-AF_NFC &sk->sk_lock.wq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) pool_lock#2 irq_context: 0 lock map_idr_lock &n->list_lock irq_context: 0 lock map_idr_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem &____s->seqcount#2 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_NFC &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 slock-AF_NFC &sk->sk_lock.wq irq_context: 0 slock-AF_NFC &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_NFC &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex _xmit_ETHER/1 _xmit_ETHER &____s->seqcount#2 irq_context: 0 rtnl_mutex _xmit_ETHER/1 _xmit_ETHER &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &n->list_lock &c->lock irq_context: 0 slock-AF_NFC &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NFC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &meta->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 rtnl_mutex &ul->lock#2 &n->list_lock irq_context: 0 rtnl_mutex &ul->lock#2 &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers &obj_hash[i].lock pool_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem kfence_freelist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &base->lock &obj_hash[i].lock irq_context: 0 &dev->mutex (work_completion)(&rfkill->uevent_work) &rq->__lock irq_context: 0 &dev->mutex (work_completion)(&rfkill->uevent_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &____s->seqcount#2 irq_context: 0 &q->sysfs_dir_lock &n->list_lock irq_context: 0 &q->sysfs_dir_lock &n->list_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &xa->xa_lock#15 irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex &____s->seqcount#2 irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock &p->pi_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock nl_table_wait.lock &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &base->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) init_task.mems_allowed_seq.seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle pcpu_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &____s->seqcount#2 irq_context: 0 &x->wait#20 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex pool_lock#2 irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu pool_lock#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock quarantine_lock irq_context: 0 misc_mtx batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &obj_hash[i].lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock &cfs_rq->removed.lock irq_context: 0 fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle pcpu_lock stock_lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock &base->lock irq_context: 0 rtnl_mutex _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: softirq &tx->clean_lock &base->lock irq_context: softirq &tx->clean_lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &data->open_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) kfence_freelist_lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 devnet_rename_sem irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 stock_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 lweventlist_lock &dir->lock#2 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem quarantine_lock irq_context: 0 rtnl_mutex &dev->tx_global_lock _xmit_NONE#2 irq_context: 0 rtnl_mutex fib_info_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex fib_info_lock pool_lock#2 irq_context: 0 rtnl_mutex raw_notifier_lock irq_context: 0 rtnl_mutex bcm_notifier_lock irq_context: 0 rtnl_mutex isotp_notifier_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#492 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 free_vmap_area_lock &pcp->lock &zone->lock irq_context: 0 free_vmap_area_lock &____s->seqcount irq_context: 0 kn->active#18 &____s->seqcount#2 irq_context: 0 kn->active#18 &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ebt_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 uts_sem &obj_hash[i].lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &meta->lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#193 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 &journal->j_list_lock &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock#2 &c->lock irq_context: 0 &journal->j_list_lock kfence_freelist_lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx nl_table_lock irq_context: 0 raw_notifier_lock irq_context: 0 rtnl_mutex sk_lock-AF_CAN irq_context: 0 rtnl_mutex sk_lock-AF_CAN slock-AF_CAN irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &n->list_lock &c->lock irq_context: 0 uts_sem pool_lock#2 irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_callback key#24 irq_context: 0 &dev->mutex rfkill_global_mutex &data->mtx irq_context: 0 &dev->mutex rfkill_global_mutex &data->read_wait#2 irq_context: 0 purge_vmap_area_lock &meta->lock irq_context: 0 rtnl_mutex sk_lock-AF_CAN pcpu_lock irq_context: 0 rtnl_mutex sk_lock-AF_CAN clock-AF_CAN irq_context: 0 cb_lock genl_mutex &sem->wait_lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx kfence_freelist_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx stock_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx key irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx percpu_counters_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_lock stock_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#48 &n->list_lock irq_context: 0 kn->active#48 &n->list_lock &c->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &root->kernfs_iattr_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &rq->__lock irq_context: 0 purge_vmap_area_lock kfence_freelist_lock irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 ebt_mutex &c->lock irq_context: 0 ebt_mutex &n->list_lock irq_context: 0 ebt_mutex &n->list_lock &c->lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &n->list_lock irq_context: 0 nf_nat_proto_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 nf_hook_mutex &n->list_lock irq_context: 0 nf_hook_mutex &n->list_lock &c->lock irq_context: 0 kn->active#4 stock_lock irq_context: 0 kn->active#4 &obj_hash[i].lock irq_context: 0 kn->active#4 key irq_context: 0 kn->active#4 pcpu_lock irq_context: 0 kn->active#4 percpu_counters_lock irq_context: 0 kn->active#4 pcpu_lock stock_lock irq_context: 0 kn->active#4 pool_lock#2 irq_context: 0 kn->active#4 &cfs_rq->removed.lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 kn->active#51 &____s->seqcount#2 irq_context: 0 kn->active#51 &____s->seqcount irq_context: 0 &mm->mmap_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#534 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ndev->lock &ifa->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex batched_entropy_u8.lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: 0 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &____s->seqcount#2 irq_context: 0 misc_mtx remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex remove_cache_srcu irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &group->inotify_data.idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &group->inotify_data.idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &group->inotify_data.idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &cfs_rq->removed.lock irq_context: 0 sb_writers#7 kn->active#4 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle pcpu_lock irq_context: 0 &u->iolock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&bat_priv->bla.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem kernfs_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem subsys mutex#73 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem deferred_probe_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem device_links_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex sysfs_symlink_target_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) umhelper_sem gdp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) fw_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex reg_indoor_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex krc.lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex reg_requests_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex reg_pending_beacons_lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&fw_work->work) rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 slock-AF_INET#2 &sk->sk_lock.wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &x->wait#24 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1015 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &meta->lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &type->s_umount_key#66 fill_pool_map-wait-type-override pool_lock irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &n->list_lock &c->lock irq_context: 0 &hdev->lock &c->lock irq_context: 0 &pcp->lock rcu_read_lock &p->pi_lock irq_context: 0 &pcp->lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &pcp->lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex slock-AF_CAN irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1015 &rq->__lock irq_context: 0 &data->open_mutex &____s->seqcount#2 irq_context: 0 &data->open_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle pcpu_lock stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#8 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &lock->wait_lock irq_context: 0 cb_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 bit_wait_table + i irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rq->__lock cpu_asid_lock irq_context: 0 &type->s_umount_key#32 &rq->__lock irq_context: 0 &type->s_umount_key#32 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1015 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1015 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex pool_lock#2 irq_context: 0 &dev->mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1007 irq_context: 0 (wq_completion)events_unbound &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &base->lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#229 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_wait_updates &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#2/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock kfence_freelist_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#229 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#362 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER batched_entropy_u8.lock irq_context: 0 file_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback quarantine_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu quarantine_lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 file_rwsem rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#776 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#74 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#74 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#72 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#72 &rq->__lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex net_rwsem nl_table_lock irq_context: 0 rtnl_mutex net_rwsem &____s->seqcount irq_context: 0 rtnl_mutex net_rwsem nl_table_wait.lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#4 &p->pi_lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#4 &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#34 &rq->__lock irq_context: 0 &type->s_umount_key#34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &base->lock &obj_hash[i].lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 kn->active#4 &lock->wait_lock irq_context: 0 kn->active#4 &p->pi_lock irq_context: 0 kn->active#4 &p->pi_lock &rq->__lock irq_context: 0 kn->active#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rcu_state.expedited_wq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &base->lock irq_context: 0 &mm->mmap_lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_node_0 irq_context: 0 namespace_sem &n->list_lock irq_context: 0 namespace_sem &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &pipe->mutex/1 fs_reclaim &rq->__lock irq_context: 0 &pipe->mutex/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 iattr_mutex &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh &base->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_persistent_keepalive) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 rtnl_mutex pcpu_alloc_mutex.wait_lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 rcu_node_0 irq_context: 0 sk_lock-AF_INET6 rcu_read_lock rcu_node_0 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &nf_nat_locks[i] irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sb_writers#3 sb_internal jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rq->__lock cpu_asid_lock irq_context: 0 rcu_read_lock &sighand->siglock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq &(&wb->bw_dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &base->lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#72 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#71 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#71 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#71 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#31 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 misc_mtx pcpu_alloc_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 lock#4 &lruvec->lru_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &group->mark_mutex batched_entropy_u8.lock irq_context: 0 &group->mark_mutex kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &group->mark_mutex rcu_node_0 irq_context: 0 &type->s_umount_key#66 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 key#23 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &n->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem hwsim_radio_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &hdev->req_lock rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex key irq_context: 0 misc_mtx &dev->mutex &____s->seqcount#2 irq_context: 0 &type->s_umount_key#66 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &type->s_umount_key#49 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#7 kn->active#4 rcu_node_0 irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->lock &____s->seqcount#8 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock nl_table_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock nl_table_wait.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rcu_read_lock lock#8 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock rcu_read_lock id_table_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &dir->lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock &n->list_lock &c->lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock &base->lock irq_context: 0 &xt[i].mutex purge_vmap_area_lock &base->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx &rq->__lock cpu_asid_lock irq_context: 0 tomoyo_ss rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &n->list_lock &c->lock irq_context: 0 &hdev->req_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock stock_lock irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_NETLINK rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &n->list_lock irq_context: 0 rtnl_mutex &bat_priv->tvlv.handler_list_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pcpu_lock stock_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 rcu_state.barrier_mutex.wait_lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#225 irq_context: 0 rtnl_mutex nf_hook_mutex &n->list_lock irq_context: 0 rtnl_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &____s->seqcount#2 irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &pcp->lock &zone->lock irq_context: 0 nf_sockopt_mutex stock_lock irq_context: 0 nf_sockopt_mutex key irq_context: 0 nf_sockopt_mutex pcpu_lock irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq irq_context: 0 nf_sockopt_mutex percpu_counters_lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 nf_sockopt_mutex pcpu_lock stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#31 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#31 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#31 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#493 irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &dentry->d_lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex pool_lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &hwstats->hwsdev_list_lock &rq->__lock irq_context: 0 rtnl_mutex &ndev->lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#169 irq_context: 0 misc_mtx pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &n->list_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &type->s_umount_key#30 lock#4 &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#30 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#493 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 rtnl_mutex &br->lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex &br->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 pcpu_lock stock_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &____s->seqcount irq_context: 0 &xt[i].mutex fs_reclaim &rcu_state.expedited_wq irq_context: 0 &xt[i].mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex &____s->seqcount irq_context: 0 &xt[i].mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 &fc->uapi_mutex fs_reclaim irq_context: 0 &fc->uapi_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &c->lock irq_context: 0 &fc->uapi_mutex pool_lock#2 irq_context: 0 &fc->uapi_mutex sb_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override pool_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 jbd2_handle irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem key irq_context: 0 rtnl_mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ipvs->est_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 &type->s_umount_key#66 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#224 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#492 irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &n->list_lock irq_context: 0 &data->open_mutex uevent_sock_mutex &n->list_lock irq_context: 0 &data->open_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)afs (work_completion)(&net->cells_manager) &rq->__lock irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#492 &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#492 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#489 irq_context: 0 rtnl_mutex &wg->device_update_lock batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#540 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 dev_addr_sem rcu_read_lock rcu_node_0 irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 dev_addr_sem rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#540 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#537 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#535 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#230 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh noop_qdisc.q.lock irq_context: softirq net/wireless/reg.c:533 irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq net/wireless/reg.c:533 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim &rq->__lock irq_context: 0 nf_sockopt_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex reg_indoor_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex krc.lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex reg_requests_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex reg_pending_beacons_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex fs_reclaim irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (crda_timeout).work &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events reg_work rtnl_mutex &c->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events reg_work rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events reg_work rtnl_mutex.wait_lock irq_context: 0 (wq_completion)events reg_work &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#489 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 (work_completion)(&tty->hangup_work) &rq->__lock irq_context: 0 (work_completion)(&tty->hangup_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 fs_reclaim pool_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#230 irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock &wq#2 irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_rx_wq#354 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock quarantine_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#354 &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#354 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) rcu_read_lock &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#226 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#225 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem purge_vmap_area_lock quarantine_lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &base->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 rcu_node_0 irq_context: 0 lock pidmap_lock batched_entropy_u8.lock irq_context: 0 lock pidmap_lock kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#962 &rq->__lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 cgroup_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 cgroup_mutex css_set_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#54 &rq->__lock irq_context: 0 &type->s_umount_key#54 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 &data->open_mutex wq_pool_mutex &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock cpu_asid_lock irq_context: 0 &data->open_mutex wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NFC &k->list_lock irq_context: 0 sk_lock-AF_NFC &k->k_lock irq_context: 0 sk_lock-AF_NFC llcp_devices_lock irq_context: 0 sk_lock-AF_NFC llcp_devices_lock &k->list_lock irq_context: 0 sk_lock-AF_NFC llcp_devices_lock &k->k_lock irq_context: 0 sk_lock-AF_NFC fs_reclaim irq_context: 0 sk_lock-AF_NFC fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_NFC pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NFC &local->sdp_lock irq_context: 0 sk_lock-AF_NFC &local->sdp_lock &local->sockets.lock irq_context: 0 sk_lock-AF_NFC &local->sockets.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NFC irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NFC slock-AF_NFC irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NFC &local->sockets.lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_NFC irq_context: 0 &xt[i].mutex remove_cache_srcu &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)krxrpcd (work_completion)(&rxnet->peer_keepalive_work) fill_pool_map-wait-type-override pool_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &pcp->lock &zone->lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#962 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#699 &rq->__lock irq_context: 0 rtnl_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq irq_context: 0 pcpu_alloc_mutex stock_lock irq_context: 0 pcpu_alloc_mutex key irq_context: softirq rcu_read_lock rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: softirq rcu_read_lock hwsim_radio_lock init_task.mems_allowed_seq.seqcount irq_context: 0 &hdev->req_lock &n->list_lock irq_context: 0 &hdev->req_lock &n->list_lock &c->lock irq_context: 0 pcpu_alloc_mutex percpu_counters_lock irq_context: 0 pcpu_alloc_mutex pcpu_lock stock_lock irq_context: 0 br_ioctl_mutex rtnl_mutex irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 lweventlist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#689 irq_context: 0 br_ioctl_mutex rtnl_mutex &mm->mmap_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock pool_lock#2 irq_context: 0 &xt[i].mutex fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pcpu_alloc_mutex rcu_read_lock &rq->__lock irq_context: 0 pcpu_alloc_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock bit_wait_table + i irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock sb_writers#3 batched_entropy_u8.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_writers#3 kfence_freelist_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &meta->lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6 k-slock-AF_INET6 irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock quarantine_lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex &wg->device_update_lock kfence_freelist_lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock remove_cache_srcu irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex &wg->device_update_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#185 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 &wq->mutex rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start key irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq rcu_callback put_task_map-wait-type-override fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_callback put_task_map-wait-type-override fill_pool_map-wait-type-override &n->list_lock irq_context: softirq rcu_callback put_task_map-wait-type-override fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 batched_entropy_u8.lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 &group->inotify_data.idr_lock &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 kfence_freelist_lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 loop_validate_mutex &lo->lo_mutex &rq->__lock irq_context: 0 loop_validate_mutex &lo->lo_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound connector_reaper_work &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 keyring_serialise_link_lock &rq->__lock irq_context: 0 keyring_serialise_link_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &rq->__lock cpu_asid_lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rename_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex gdp_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex gdp_mutex &p->pi_lock irq_context: 0 rtnl_mutex gdp_mutex &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex gdp_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macsec_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: softirq rcu_callback batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex stock_lock irq_context: 0 &xt[i].mutex key irq_context: 0 &xt[i].mutex pcpu_lock irq_context: 0 &xt[i].mutex percpu_counters_lock irq_context: 0 &xt[i].mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 sk_lock-AF_CAN &jsk->sk_session_queue_lock irq_context: 0 sk_lock-AF_CAN &priv->lock irq_context: 0 &type->lock_class keyring_serialise_link_lock &____s->seqcount irq_context: 0 sk_lock-AF_CAN (console_sem).lock irq_context: 0 sb_writers#3 &sb->s_type->i_lock_key#22 bit_wait_table + i irq_context: 0 sk_lock-AF_CAN console_lock console_srcu console_owner_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 ebt_mutex &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 ebt_mutex &mm->mmap_lock &obj_hash[i].lock irq_context: 0 ebt_mutex &mm->mmap_lock pool_lock#2 irq_context: 0 sb_writers#3 jbd2_handle &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex remove_cache_srcu irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN console_lock console_srcu console_owner irq_context: 0 sk_lock-AF_CAN console_lock console_srcu console_owner &port_lock_key irq_context: 0 sk_lock-AF_CAN console_lock console_srcu console_owner console_owner_lock irq_context: 0 sk_lock-AF_CAN &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex nbd_index_mutex &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex nbd_index_mutex &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 lweventlist_lock &dir->lock#2 irq_context: 0 &wq->mutex &cfs_rq->removed.lock irq_context: 0 &wq->mutex &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#446 irq_context: 0 &wq->mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &dentry->d_lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &meta->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work kfence_freelist_lock irq_context: 0 &type->s_umount_key#56 &rq->__lock irq_context: 0 &type->s_umount_key#56 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &dentry->d_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 (wq_completion)ext4-rsv-conversion &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &base->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex rcu_read_lock rcu_node_0 irq_context: 0 ebt_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex batched_entropy_u8.lock irq_context: 0 nfc_devlist_mutex kfence_freelist_lock irq_context: 0 nfc_devlist_mutex &meta->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_NETLINK &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex.wait_lock irq_context: 0 &rnp->exp_lock irq_context: 0 rcu_state.exp_mutex irq_context: 0 rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 rtnl_mutex &rnp->exp_lock irq_context: 0 rtnl_mutex &rnp->exp_wq[3] irq_context: 0 rtnl_mutex &idev->mc_lock krc.lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock krc.lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &idev->mc_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &obj_hash[i].lock irq_context: 0 &ep->mtx fs_reclaim &rq->__lock irq_context: 0 &ep->mtx &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rcu_state.expedited_wq irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &____s->seqcount irq_context: 0 &hdev->req_lock &hdev->lock &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock &p->pi_lock irq_context: 0 &hdev->req_lock &hdev->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &p->pi_lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 &dir->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &dir->lock pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &sem->wait_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &p->pi_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &type->lock_class keyring_serialise_link_lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#446 &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex pool_lock#2 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock key#24 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &base->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &base->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 kn->active#49 remove_cache_srcu irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#51 remove_cache_srcu irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 &u->iolock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &u->iolock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &u->iolock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &u->iolock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#49 remove_cache_srcu quarantine_lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock cpu_asid_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex.wait_lock irq_context: 0 sb_writers#3 sb_internal rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &p->pi_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &p->pi_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 kn->active#49 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#49 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->lock_class &base->lock irq_context: 0 kn->active#49 remove_cache_srcu &obj_hash[i].lock irq_context: 0 tomoyo_ss &rcu_state.expedited_wq irq_context: 0 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &local->filter_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#8 &p->pi_lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#8 &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#8 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock stock_lock irq_context: 0 &mm->mmap_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 &hdev->req_lock &meta->lock irq_context: 0 rcu_state.barrier_mutex &rq->__lock cpu_asid_lock irq_context: 0 kn->active#51 remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock irq_context: 0 remove_cache_srcu &base->lock irq_context: 0 remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &macvlan_netdev_addr_lock_key/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex net_rwsem pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock stock_lock irq_context: 0 rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &obj_hash[i].lock pool_lock irq_context: 0 &q->queue_lock &blkcg->lock percpu_ref_switch_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &n->list_lock &c->lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &n->list_lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &n->list_lock &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock#2 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ul->lock#2 &dir->lock#2 irq_context: 0 lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#170 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &____s->seqcount#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock quarantine_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex freezer_mutex.wait_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex leds_list_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.gp_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 init_user_ns.keyring_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock pcpu_lock irq_context: 0 prog_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) vmap_area_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) purge_vmap_area_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem &meta->lock irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock &c->lock irq_context: 0 sb_writers key irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock k-slock-AF_INET6 rcu_read_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex batched_entropy_u8.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex kfence_freelist_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &meta->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#4 rcu_node_0 irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex.wait_lock irq_context: 0 rtnl_mutex &wg->device_update_lock &p->pi_lock irq_context: 0 rtnl_mutex &wg->device_update_lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#50 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &fsnotify_mark_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 vmap_purge_lock &rq->__lock cpu_asid_lock irq_context: 0 lock prog_idr_lock irq_context: 0 lock prog_idr_lock pool_lock#2 irq_context: 0 sk_lock-AF_PACKET &rnp->exp_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[1] irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_lock irq_context: 0 &sb->s_type->i_mutex_key#10 &rnp->exp_wq[1] irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_node_0 irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 bpf_lock irq_context: 0 prog_idr_lock &obj_hash[i].lock irq_context: 0 prog_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock quarantine_lock irq_context: 0 vmap_purge_lock &cfs_rq->removed.lock irq_context: 0 vmap_purge_lock &obj_hash[i].lock irq_context: 0 lock prog_idr_lock &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex kfence_freelist_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx subsys mutex#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&ifa->dad_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex leds_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) purge_vmap_area_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers pcpu_lock irq_context: 0 sb_writers percpu_counters_lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &c->lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &n->list_lock irq_context: 0 sk_lock-AF_PACKET &po->bind_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 sb_writers pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 &mm->mmap_lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 rtnl_mutex dev_hotplug_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock pcpu_lock stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh quarantine_lock irq_context: 0 &group->mark_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_base_lock &xa->xa_lock#4 &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#446 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->w) rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->mutex &rq->__lock irq_context: 0 misc_mtx console_lock console_srcu console_owner_lock irq_context: 0 misc_mtx console_lock console_srcu console_owner irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 rcu_node_0 irq_context: 0 sb_writers#6 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#446 irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex pool_lock#2 irq_context: 0 misc_mtx console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock fs_reclaim irq_context: 0 (wq_completion)nfc2_nci_rx_wq#695 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tn->lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tn->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#438 irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &obj_hash[i].lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work rcu_node_0 irq_context: 0 rcu_read_lock &dentry->d_lock &p->pi_lock irq_context: 0 misc_mtx console_lock console_srcu console_owner console_owner_lock irq_context: 0 &vma->vm_lock->lock fs_reclaim stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock &____s->seqcount irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock stock_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#435 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tn->lock pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 batched_entropy_u8.lock crngs.lock irq_context: 0 pernet_ops_rwsem batched_entropy_u8.lock irq_context: 0 pernet_ops_rwsem kfence_freelist_lock irq_context: 0 sk_lock-AF_INET6 slock-AF_INET6 &sk->sk_lock.wq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 slock-AF_INET6 &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mount_lock irq_context: 0 sched_map-wait-type-override rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &vma->vm_lock->lock fs_reclaim key irq_context: 0 (wq_completion)nfc2_nci_tx_wq#693 irq_context: 0 &vma->vm_lock->lock fs_reclaim pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_node_0 irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 (console_sem).lock &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#699 irq_context: 0 key_types_sem &cfs_rq->removed.lock irq_context: 0 key_types_sem key_construction_mutex &type->lock_class#3/1 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 lock prog_idr_lock &n->list_lock irq_context: 0 lock prog_idr_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &____s->seqcount#2 irq_context: 0 vmap_purge_lock rcu_read_lock rcu_node_0 irq_context: 0 vmap_purge_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 vmap_purge_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 vmap_purge_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events key_gc_work &type->lock_class &n->list_lock irq_context: 0 key_types_sem &type->lock_class fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#231 irq_context: 0 &xt[i].mutex free_vmap_area_lock &base->lock irq_context: 0 &vma->vm_lock->lock fs_reclaim percpu_counters_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex flowtable_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (crypto_chain).rwsem rcu_read_lock rcu_node_0 irq_context: 0 (crypto_chain).rwsem rcu_read_lock &rq->__lock irq_context: 0 (crypto_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#231 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) batched_entropy_u8.lock crngs.lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex flowtable_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &vma->vm_lock->lock fs_reclaim pcpu_lock stock_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#231 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) pool_lock#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock sb_writers#4 remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem &n->list_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#17 namespace_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex triggers_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq security/keys/gc.c:28 irq_context: softirq security/keys/gc.c:28 rcu_read_lock &pool->lock irq_context: softirq security/keys/gc.c:28 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq security/keys/gc.c:28 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock key#22 irq_context: 0 sk_lock-AF_CAN remove_cache_srcu rcu_node_0 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq security/keys/gc.c:28 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq security/keys/gc.c:28 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &sem->wait_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &group->notification_waitq &p->pi_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_node_0 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: softirq (&in_dev->mr_ifc_timer) fill_pool_map-wait-type-override &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) fill_pool_map-wait-type-override &n->list_lock irq_context: softirq (&in_dev->mr_ifc_timer) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex key irq_context: 0 cb_lock genl_mutex rtnl_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex percpu_counters_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER/1 _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex &meta->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &rq->__lock irq_context: 0 pcpu_alloc_mutex rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 &dentry->d_lock fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 &data->open_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 &data->open_mutex uevent_sock_mutex &rcu_state.expedited_wq irq_context: 0 &data->open_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &data->open_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &xt[i].mutex free_vmap_area_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 br_ioctl_mutex rtnl_mutex krc.lock &base->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 &c->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->routes.lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &base->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &base->lock &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#6 &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 &root->kernfs_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 device_links_lock &rq->__lock irq_context: 0 &ep->mtx &rcu_state.expedited_wq irq_context: 0 &ep->mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ep->mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ep->mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 device_links_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 &p->lock remove_cache_srcu stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex kfence_freelist_lock irq_context: 0 &p->lock remove_cache_srcu key irq_context: 0 proto_tab_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock irq_context: 0 pcpu_alloc_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock kfence_freelist_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&priv->connect)->work) &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 sb_internal jbd2_handle pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_NETLINK key irq_context: 0 sk_lock-AF_NETLINK pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_NETLINK percpu_counters_lock irq_context: 0 sk_lock-AF_NETLINK pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &____s->seqcount#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &____s->seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 lock#4 &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dentry->d_lock rcu_read_lock &p->pi_lock irq_context: 0 &dentry->d_lock rcu_read_lock &sb->s_type->i_lock_key#23 &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &nft_net->commit_mutex &rq->__lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &nft_net->commit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_NETLINK &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#313 irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex bpf_devs_lock irq_context: 0 &p->lock remove_cache_srcu pcpu_lock irq_context: 0 &p->lock remove_cache_srcu percpu_counters_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex.wait_lock irq_context: 0 rtnl_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tsk->futex_exit_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &tsk->futex_exit_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &obj_hash[i].lock pool_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 kn->active#4 fs_reclaim &cfs_rq->removed.lock irq_context: 0 kn->active#4 fs_reclaim &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &____s->seqcount#2 irq_context: 0 &u->iolock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock batched_entropy_u8.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &p->lock remove_cache_srcu pcpu_lock stock_lock irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &meta->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 &meta->lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 &type->s_umount_key#50 &rq->__lock irq_context: 0 &ep->mtx &____s->seqcount#2 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &base->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 lock#4 &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_CAN &jsk->filters_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 krc.lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 krc.lock &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 k-slock-AF_INET6 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 rcu_state.barrier_mutex pool_lock#2 irq_context: 0 lweventlist_lock pool_lock#2 irq_context: 0 lweventlist_lock &dir->lock#2 irq_context: 0 rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &n->list_lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock stock_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock key irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock pcpu_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock percpu_counters_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &cfs_rq->removed.lock irq_context: 0 rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#9 &rq->__lock cpu_asid_lock irq_context: 0 &data->open_mutex &sem->wait_lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &pool->lock &p->pi_lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 &rq->__lock cpu_asid_lock irq_context: 0 &hdev->req_lock kfence_freelist_lock irq_context: 0 rtnl_mutex &tbl->lock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 rtnl_mutex &wg->device_update_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock (console_sem).lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock (console_sem).lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock (console_sem).lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock (console_sem).lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 tomoyo_ss &obj_hash[i].lock pool_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock console_srcu console_owner_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu rcu_node_0 irq_context: 0 nfc_devlist_mutex kn->active#4 irq_context: 0 nfc_devlist_mutex kn->active#4 &root->deactivate_waitq irq_context: 0 nfc_devlist_mutex kn->active#4 &rq->__lock irq_context: 0 nfc_devlist_mutex kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq &p->pi_lock irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq &p->pi_lock &rq->__lock irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &p->lock &of->mutex &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock console_srcu console_owner irq_context: 0 tty_mutex &tty->legacy_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem &tbl->lock &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#4 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 init_user_ns.keyring_sem &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#313 &rq->__lock irq_context: 0 init_user_ns.keyring_sem &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#4 remove_cache_srcu &c->lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 tty_mutex &____s->seqcount#2 irq_context: 0 tty_mutex &____s->seqcount irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NFC &local->raw_sockets.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NFC &local->raw_sockets.lock irq_context: 0 &root->kernfs_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->lock_class &type->lock_class/1 irq_context: 0 &type->lock_class &type->lock_class/1 fs_reclaim irq_context: 0 &type->lock_class &type->lock_class/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->lock_class &type->lock_class/1 &c->lock irq_context: 0 &type->lock_class &type->lock_class/1 pool_lock#2 irq_context: 0 &type->lock_class &type->lock_class/1 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock &____s->seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 remove_cache_srcu &n->list_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 rcu_read_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#692 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#693 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &n->list_lock &c->lock irq_context: softirq &(&bat_priv->orig_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &pcp->lock &zone->lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events key_gc_work rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events key_gc_work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &obj_hash[i].lock irq_context: 0 key_types_sem key_construction_mutex key_construction_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem key_construction_mutex.wait_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock pcpu_lock stock_lock irq_context: 0 fs_reclaim &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 &hdev->lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle irq_context: 0 misc_mtx nfc_devlist_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#117 irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#117 &devlink_port->type_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#999 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 key irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 pcpu_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 percpu_counters_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &base->lock irq_context: 0 (wq_completion)events (debug_obj_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#999 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#999 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->lock_class root_key_user.lock irq_context: 0 &sig->cred_guard_mutex &meta->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock irq_context: 0 &type->lock_class &obj_hash[i].lock irq_context: 0 &type->lock_class &rq->__lock irq_context: 0 &type->lock_class &c->lock irq_context: 0 &type->lock_class &n->list_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &type->lock_class &n->list_lock &c->lock irq_context: 0 &type->lock_class &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->page_table_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->page_table_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_read_lock &pool->lock irq_context: 0 &type->lock_class pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx &rq->__lock irq_context: 0 misc_mtx rfkill_global_mutex &data->mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_NFC &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock stock_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock pcpu_lock stock_lock irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: hardirq &fq->mq_flush_lock fill_pool_map-wait-type-override &c->lock irq_context: hardirq &fq->mq_flush_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: hardirq &fq->mq_flush_lock fill_pool_map-wait-type-override &n->list_lock irq_context: hardirq &fq->mq_flush_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: hardirq &fq->mq_flush_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex &br->hash_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#45 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#45 sb_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#45 &dentry->d_lock irq_context: 0 &sighand->siglock batched_entropy_u8.lock irq_context: 0 tty_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock batched_entropy_u32.lock irq_context: 0 rcu_read_lock &sighand->siglock &____s->seqcount#2 irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 &base->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#999 irq_context: 0 rtnl_mutex dev_addr_sem nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nl_table_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#13 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &fsnotify_mark_srcu fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 rcu_read_lock &xa->xa_lock#8 &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#2 &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#2 &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock cpu_asid_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock batched_entropy_u8.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock kfence_freelist_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &meta->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock kfence_freelist_lock irq_context: 0 &u->iolock rcu_read_lock &rcu_state.gp_wq irq_context: 0 kn->active#51 remove_cache_srcu &c->lock irq_context: 0 kn->active#51 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#51 remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 tty_mutex rcu_node_0 irq_context: 0 tty_mutex &rcu_state.expedited_wq irq_context: 0 tty_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex &obj_hash[i].lock pool_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&in_dev->mr_ifc_timer) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 &base->lock irq_context: 0 &type->s_umount_key#30 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#13 mount_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock console_owner_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock console_owner irq_context: 0 sb_writers#13 &sb->s_type->i_mutex_key#18 irq_context: 0 sb_writers#13 &sb->s_type->i_mutex_key#18 rename_lock.seqcount irq_context: 0 sb_writers#13 &sb->s_type->i_mutex_key#18 rcu_read_lock &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 irq_context: 0 &u->iolock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &u->iolock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 slock-AF_INET#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &bgl->locks[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 misc_mtx fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &u->iolock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &n->list_lock irq_context: 0 &ep->mtx rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &x->wait#3 &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &u->iolock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &u->iolock rcu_read_lock &obj_hash[i].lock irq_context: 0 &u->iolock rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#5 quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 rlock-AF_INET irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET/1 &list->lock#19 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock irq_context: 0 &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 &hdev->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &hdev->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &hdev->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &hdev->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &journal->j_state_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &hdev->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &pnettable->lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &sctp_ep_hashtable[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 &net->sctp.addr_wq_lock slock-AF_INET/1 clock-AF_INET irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &c->lock irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events key_gc_work krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &rq->__lock irq_context: softirq (&in_dev->mr_ifc_timer) fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &data->open_mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 &group->inotify_data.idr_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu quarantine_lock irq_context: 0 loop_ctl_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &p->lock &of->mutex kn->active#4 rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&w->w) &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock &br->multicast_lock &____s->seqcount#2 irq_context: 0 &p->lock &of->mutex kn->active#4 rcu_read_lock &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.barrier_mutex pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 rcu_read_lock rcu_node_0 irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 rcu_node_0 irq_context: 0 tomoyo_ss remove_cache_srcu key irq_context: 0 tomoyo_ss remove_cache_srcu pcpu_lock irq_context: 0 tomoyo_ss remove_cache_srcu percpu_counters_lock irq_context: 0 tty_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &rq->__lock irq_context: 0 tty_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#46/1 &sb->s_type->i_mutex_key#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &c->lock irq_context: 0 sb_writers#8 &rq->__lock irq_context: 0 sb_writers#8 &n->list_lock irq_context: 0 sb_writers#8 &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock krc.lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex irq_context: 0 rtnl_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[0] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_wait_updates &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_owner_lock irq_context: 0 misc_mtx batched_entropy_u8.lock irq_context: 0 misc_mtx kfence_freelist_lock irq_context: 0 sb_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 misc_mtx &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx &dev->mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 fs_reclaim rcu_node_0 irq_context: 0 kn->active#4 fs_reclaim &rcu_state.expedited_wq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#489 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex console_owner irq_context: 0 (wq_completion)nfc2_nci_rx_wq#992 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#489 &cfs_rq->removed.lock irq_context: 0 sb_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#986 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 rtnl_mutex _xmit_ETHER batched_entropy_u8.lock irq_context: 0 rtnl_mutex _xmit_ETHER kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock kfence_freelist_lock irq_context: 0 tty_mutex &tty->legacy_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)nfc2_nci_tx_wq#986 &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#986 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key#3 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#489 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#487 irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &rq->__lock irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#30 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex.wait_lock irq_context: 0 nf_sockopt_mutex rcu_read_lock &rq->__lock irq_context: 0 nf_sockopt_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq rcu_callback stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events_power_efficient (gc_work).work rcu_node_0 irq_context: 0 kn->active#16 &n->list_lock irq_context: 0 kn->active#16 &n->list_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#30 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#29 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#490 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#488 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#201 irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex (inetaddr_validator_chain).rwsem &rq->__lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem iattr_mutex &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_validator_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &obj_hash[i].lock pool_lock irq_context: 0 loop_ctl_mutex &rq->__lock irq_context: 0 loop_ctl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#171 irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_CAN &priv->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fib_info_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &base->lock irq_context: 0 sb_writers#3 &base->lock &obj_hash[i].lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock kfence_freelist_lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex pcpu_lock irq_context: 0 rtnl_mutex uevent_sock_mutex percpu_counters_lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sk_lock-AF_INET rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#171 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_node_0 irq_context: 0 kn->active#20 &____s->seqcount#2 irq_context: 0 sb_writers#5 lock#4 &obj_hash[i].lock irq_context: 0 &mm->mmap_lock sb_writers#3 &cfs_rq->removed.lock irq_context: softirq rcu_callback pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 loop_ctl_mutex &c->lock irq_context: 0 rtnl_mutex &rq->__lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &rq->__lock &base->lock irq_context: 0 rtnl_mutex &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 rcu_state.exp_mutex stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: 0 &q->mq_freeze_lock &rq->__lock irq_context: 0 &q->mq_freeze_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &q->queue_lock &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rcu_state.expedited_wq irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nl_table_wait.lock &p->pi_lock irq_context: 0 misc_mtx nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_node_0 irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &obj_hash[i].lock irq_context: 0 &p->alloc_lock &x->wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rcu_state.exp_mutex key irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &rdev->wiphy.mtx &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rq->__lock irq_context: 0 sb_writers#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_lock_key &xa->xa_lock#8 irq_context: 0 &info->lock irq_context: 0 &sbinfo->stat_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_node_0 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 (console_sem).lock &p->pi_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &____s->seqcount irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 (console_sem).lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.exp_mutex pcpu_lock irq_context: 0 rcu_state.exp_mutex percpu_counters_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &list->lock#14 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &list->lock#14 pool_lock#2 irq_context: 0 rcu_state.exp_mutex pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#171 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 sb_writers#4 oom_adj_mutex &rq->__lock cpu_asid_lock irq_context: 0 key_types_sem asymmetric_key_parsers_sem &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 lock#4 &obj_hash[i].lock irq_context: 0 &lruvec->lru_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->xdp.lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->xdp.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &____s->seqcount irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 pool_lock#2 irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex pool_lock#2 irq_context: 0 rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem stock_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#166 irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock bit_wait_table + i &p->pi_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#172 irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex deferred_probe_mutex &rq->__lock irq_context: 0 nfc_devlist_mutex deferred_probe_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &ct->lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &rq->__lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 &type->i_mutex_dir_key#4 &rq->__lock cpu_asid_lock irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&sk->sk_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock &xa->xa_lock#17 irq_context: 0 &hdev->req_lock &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock &xa->xa_lock#17 &pcp->lock &zone->lock irq_context: 0 &hdev->req_lock &hdev->lock &xa->xa_lock#17 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &hdev->req_lock &hdev->lock &xa->xa_lock#17 &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock kfence_freelist_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &xa->xa_lock#17 pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#449 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#449 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#446 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &lock->wait_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &p->pi_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#46/1 binderfs_minors_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 misc_mtx wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 cb_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex &meta->lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &____s->seqcount#2 irq_context: 0 &group->mark_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &meta->lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex fs_reclaim irq_context: 0 rtnl_mutex pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex pcpu_alloc_mutex &____s->seqcount irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock irq_context: 0 rtnl_mutex stack_depot_init_mutex &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 &lru->node[i].lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &base->lock &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex stock_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#201 irq_context: softirq rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx kfence_freelist_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock &n->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex kfence_freelist_lock irq_context: 0 br_ioctl_mutex rtnl_mutex pcpu_alloc_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 br_ioctl_mutex rtnl_mutex batched_entropy_u32.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &meta->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &n->list_lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &____s->seqcount#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &n->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &n->list_lock &c->lock irq_context: 0 &u->iolock &pcp->lock &zone->lock irq_context: 0 &u->iolock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock key irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock pcpu_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock percpu_counters_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &xa->xa_lock#4 irq_context: 0 br_ioctl_mutex rtnl_mutex net_rwsem irq_context: 0 br_ioctl_mutex rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex &tn->lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock rcu_node_0 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 stock_lock irq_context: 0 sb_writers#3 pcpu_lock stock_lock irq_context: softirq (&app->join_timer) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rcu_state.expedited_wq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &x->wait#9 irq_context: 0 br_ioctl_mutex rtnl_mutex &k->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 rcu_read_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex gdp_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock xps_map_mutex &rq->__lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#783 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#49 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 &ep->mtx fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#185 irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &dentry->d_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->routes.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)gid-cache-wq &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &n->list_lock irq_context: 0 rtnl_mutex &dev_addr_list_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss stock_lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &buf->lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)bond0#118 irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 rtnl_mutex uevent_sock_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: softirq (&app->join_timer) fill_pool_map-wait-type-override &c->lock irq_context: softirq (&app->join_timer) fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &macvlan_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 batched_entropy_u8.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock batched_entropy_u64.lock crngs.lock base_crng.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex quarantine_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#764 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#764 &rq->__lock irq_context: 0 (wq_completion)bond0#118 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 stock_lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle mmu_notifier_invalidate_range_start pool_lock#2 irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &base->lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#444 irq_context: 0 misc_mtx &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &cfs_rq->removed.lock irq_context: 0 misc_mtx rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events key_gc_work &type->lock_class &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 pcpu_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#694 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 pcpu_lock stock_lock irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sk_lock-AF_PACKET &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET6 clock-AF_INET6 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#231 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 sb_internal batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 key_construction_mutex &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock stock_lock irq_context: 0 sb_writers#6 &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock &c->lock irq_context: 0 pernet_ops_rwsem lock kernfs_idr_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock lock kernfs_idr_lock &c->lock irq_context: 0 uevent_sock_mutex remove_cache_srcu irq_context: 0 uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &mm->mmap_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 &obj_hash[i].lock pool_lock irq_context: 0 &q->sysfs_lock &n->list_lock irq_context: 0 &u->iolock rcu_node_0 irq_context: 0 &u->iolock &rcu_state.expedited_wq irq_context: 0 &u->iolock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &u->iolock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &u->iolock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_TUNNEL6#2 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem bit_wait_table + i irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock hwsim_radio_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 &f->f_pos_lock sb_writers#4 remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 &obj_hash[i].lock pool_lock irq_context: softirq rcu_read_lock hwsim_radio_lock kfence_freelist_lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu &c->lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &kernfs_locks->open_file_mutex[count] &cfs_rq->removed.lock irq_context: 0 sb_writers#7 tomoyo_ss mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &q->sysfs_lock &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key/1 &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &n->list_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx cpu_hotplug_lock &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &n->list_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex nbd_index_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &dev->mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET rcu_node_0 irq_context: 0 cb_lock genl_mutex nbd_index_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex nbd_index_mutex rlock-AF_NETLINK irq_context: 0 cb_lock genl_mutex nbd_index_mutex &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex nbd_index_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex nbd_index_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock crngs.lock irq_context: 0 cb_lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key/1 &group->inotify_data.idr_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock key irq_context: 0 (wq_completion)nfc3_nci_rx_wq#227 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#226 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &rq->__lock cpu_asid_lock irq_context: 0 br_ioctl_mutex rtnl_mutex lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#541 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#541 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) pool_lock#2 irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#694 &rq->__lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 &mm->mmap_lock fs_reclaim &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#89 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#89 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy228 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock percpu_counters_lock irq_context: 0 (wq_completion)writeback (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex k-sk_lock-AF_INET &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rnp->exp_wq[0] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rcu_node_0 irq_context: 0 misc_mtx &dev->mutex &rcu_state.expedited_wq irq_context: 0 tomoyo_ss rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx &dev->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 key_construction_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &c->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu pool_lock#2 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy228 &rq->__lock irq_context: 0 (wq_completion)nfc20_nci_cmd_wq#13 irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#5 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#7 &sb->s_type->i_mutex_key#14 &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem pool_lock#2 irq_context: 0 sb_writers#7 batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &type->s_umount_key#30 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal &rcu_state.expedited_wq irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 sb_internal &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &rq->__lock irq_context: 0 &q->sysfs_dir_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx pcpu_alloc_mutex &cfs_rq->removed.lock irq_context: 0 misc_mtx pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 misc_mtx pcpu_alloc_mutex pool_lock#2 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &q->sysfs_dir_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock pool_lock#2 irq_context: 0 &tty->legacy_mutex &port->open_wait irq_context: 0 &tty->legacy_mutex &tty->ldisc_sem &buf->lock irq_context: 0 &tty->legacy_mutex &tty->ldisc_sem &buf->lock &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem vmap_area_lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex &x->wait#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex quarantine_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 binderfs_minors_mutex irq_context: 0 binderfs_minors_mutex binderfs_minors.xa_lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex &n->list_lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex &n->list_lock &c->lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#18 irq_context: 0 &sb->s_type->i_mutex_key#18 rename_lock.seqcount irq_context: 0 key_types_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 batched_entropy_u8.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)events (debug_obj_work).work &cfs_rq->removed.lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &base->lock irq_context: 0 &pool->lock &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc20_nci_rx_wq#13 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events (debug_obj_work).work &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 &q->debugfs_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &q->debugfs_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &q->debugfs_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock genl_mutex nbd_index_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex nbd_index_mutex &n->list_lock &c->lock irq_context: 0 &q->debugfs_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#694 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#695 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#697 irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#13 irq_context: 0 cb_lock genl_mutex deferred_probe_mutex &rq->__lock irq_context: 0 nl_table_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy228 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#87 irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#538 irq_context: 0 misc_mtx &dev->mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &rq->__lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#85 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &lock->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem purge_vmap_area_lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem purge_vmap_area_lock pool_lock#2 irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem purge_vmap_area_lock &pcp->lock &zone->lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem purge_vmap_area_lock &____s->seqcount irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem pool_lock#2 irq_context: 0 &tty->ldisc_sem tty_ldiscs_lock irq_context: 0 &tty->ldisc_sem &obj_hash[i].lock irq_context: 0 &tty->ldisc_sem pool_lock#2 irq_context: 0 (work_completion)(&tty->SAK_work) irq_context: 0 &dev->mutex rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 (work_completion)(&tty->hangup_work) irq_context: 0 tty_mutex (work_completion)(&buf->work) irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &c->lock irq_context: 0 uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#85 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#536 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#542 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &base->lock &obj_hash[i].lock irq_context: 0 init_user_ns.keyring_sem rcu_read_lock rcu_node_0 irq_context: 0 init_user_ns.keyring_sem rcu_read_lock &rq->__lock irq_context: 0 nfc_devlist_mutex &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_lock_key#8 bit_wait_table + i irq_context: 0 &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tt.last_changeset_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &rq->__lock irq_context: 0 &f->f_pos_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &lock->wait_lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_node_0 irq_context: 0 &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 &q->sysfs_dir_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &rq->__lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &obj_hash[i].lock pool_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock cpu_asid_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#542 irq_context: 0 init_user_ns.keyring_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 &base->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#539 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#537 irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock rcu_read_lock rcu_node_0 irq_context: 0 kn->active#46 &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key/1 &____s->seqcount irq_context: 0 sb_writers#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock rcu_read_lock &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 &p->lock &of->mutex kn->active#4 &rcu_state.expedited_wq irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#537 &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#4 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#232 irq_context: 0 &p->lock &of->mutex kn->active#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex fs_reclaim rcu_node_0 irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &root->kernfs_rwsem irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 tty_mutex &tty->legacy_mutex console_lock bus_type_sem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sk_lock-AF_INET remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#232 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock sysfs_symlink_target_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex subsys mutex#40 &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex subsys mutex#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &pcp->lock &zone->lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &root->kernfs_rwsem irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &dev->power.lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock dpm_list_mtx irq_context: 0 tty_mutex &tty->legacy_mutex console_lock req_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &p->pi_lock irq_context: 0 key_types_sem root_key_user.cons_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mount_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss mount_lock rcu_read_lock rename_lock.seqcount irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &p->pi_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex &obj_hash[i].lock pool_lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#228 irq_context: 0 kn->active#4 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 kn->active#4 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 kn->active#49 &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock fs_reclaim irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex lock kernfs_idr_lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 tty_mutex rcu_read_lock &pool->lock irq_context: 0 tty_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 tty_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 rcu_read_lock rcu_node_0 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events key_gc_work sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &p->pi_lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &dev->mutex &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock pool_lock#2 irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock &mm->page_table_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock &anon_vma->rwsem irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock &anon_vma->rwsem &mm->page_table_lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &____s->seqcount irq_context: 0 cb_lock batched_entropy_u8.lock irq_context: 0 &disk->open_mutex stock_lock irq_context: 0 &disk->open_mutex key irq_context: 0 br_ioctl_mutex rtnl_mutex bus_type_sem irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 tk_core.seq.seqcount irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &c->lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu pool_lock#2 irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 misc_mtx nfc_devlist_mutex bus_type_sem &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex bus_type_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#7 irq_context: 0 nfc_devlist_mutex &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &ct->lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock &____s->seqcount#9 irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nfnl_subsys_ipset &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nfnl_subsys_ipset &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 key_types_sem &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex sysfs_symlink_target_lock irq_context: 0 key_types_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 batched_entropy_u8.lock irq_context: 0 kn->active#4 kfence_freelist_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#2 rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#2 rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &____s->seqcount#2 irq_context: 0 &dev->mutex &root->kernfs_rwsem &base->lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &____s->seqcount irq_context: 0 br_ioctl_mutex rtnl_mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (switchdev_blocking_notif_chain).rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (switchdev_blocking_notif_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &cfs_rq->removed.lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex &n->list_lock irq_context: 0 &type->s_umount_key#55 &rq->__lock irq_context: 0 sb_writers#3 quarantine_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sk_lock-AF_INET &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ret->b_state_lock &journal->j_list_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NETLINK stock_lock irq_context: 0 sk_lock-AF_NETLINK &____s->seqcount#2 irq_context: 0 sk_lock-AF_NETLINK &f->f_lock irq_context: 0 &type->s_umount_key#30 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_NETLINK &f->f_lock fasync_lock irq_context: 0 &fc->uapi_mutex &rq->__lock irq_context: 0 wq_pool_mutex wq_pool_mutex.wait_lock irq_context: 0 wq_pool_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle irq_context: softirq (&pool->idle_timer) &pool->lock irq_context: softirq (&pool->idle_timer) &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&pool->idle_timer) &pool->lock &base->lock irq_context: softirq (&pool->idle_timer) &pool->lock &base->lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &vlan_netdev_addr_lock_key/1 _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock tk_core.seq.seqcount irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 &obj_hash[i].lock irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem percpu_counters_lock irq_context: 0 sb_writers#7 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &rq->__lock irq_context: 0 misc_mtx &dev->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock rcu_read_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex console_lock console_srcu console_owner irq_context: 0 cb_lock genl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 cb_lock genl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 &dev->mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &meta->lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem free_vmap_area_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fill_pool_map-wait-type-override &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER krc.lock &base->lock &obj_hash[i].lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem free_vmap_area_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 tty_mutex batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 tty_mutex kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 kfence_freelist_lock irq_context: 0 &tty->legacy_mutex &lock->wait_lock irq_context: 0 pcpu_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &tty->legacy_mutex &port->mutex &rq->__lock irq_context: 0 &tty->legacy_mutex &tty->ldisc_sem &buf->lock &rq->__lock irq_context: 0 &group->mark_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &group->mark_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->legacy_mutex &tty->ldisc_sem &buf->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_node_0 irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_retransmit_timer) slock-AF_INET#2 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dpm_list_mtx &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)events (debug_obj_work).work &meta->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &dev->power.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex dpm_list_mtx irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim &cfs_rq->removed.lock irq_context: 0 tty_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 tty_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 tty_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex remove_cache_srcu irq_context: 0 tty_mutex remove_cache_srcu quarantine_lock irq_context: 0 &f->f_owner.lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->work_lock &base->lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rq->__lock &cfs_rq->removed.lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock irq_context: 0 &fq->mq_flush_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &____s->seqcount irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq security/integrity/ima/ima_queue_keys.c:35 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#228 &rq->__lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 init_user_ns.keyring_sem &rq->__lock irq_context: 0 init_user_ns.keyring_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle pcpu_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#228 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock quarantine_lock irq_context: 0 sk_lock-AF_X25 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 &dev->mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 &tty->files_lock irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 stock_lock irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &pcp->lock &zone->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 misc_mtx pool_lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock quarantine_lock irq_context: 0 br_ioctl_mutex rtnl_mutex subsys mutex#20 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &____s->seqcount irq_context: 0 &dev->mutex &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 &dev->mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex kernfs_idr_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 br_ioctl_mutex rtnl_mutex dev_hotplug_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 pernet_ops_rwsem nl_table_wait.lock &p->pi_lock irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 pernet_ops_rwsem fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem &rq->__lock cpu_asid_lock irq_context: 0 br_ioctl_mutex rtnl_mutex dev_base_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rq->__lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex input_pool.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &base->lock &obj_hash[i].lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh &r->producer_lock#3 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &wq->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &tbl->lock irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock irq_context: 0 br_ioctl_mutex rtnl_mutex nl_table_lock irq_context: 0 br_ioctl_mutex rtnl_mutex nl_table_wait.lock irq_context: 0 br_ioctl_mutex rtnl_mutex failover_lock irq_context: 0 br_ioctl_mutex rtnl_mutex proc_subdir_lock irq_context: 0 cb_lock &dir->lock#2 irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 br_ioctl_mutex rtnl_mutex proc_subdir_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock irq_context: 0 &vma->vm_lock->lock &obj_hash[i].lock pool_lock irq_context: 0 kn->active#4 &rq->__lock cpu_asid_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &____s->seqcount irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex (switchdev_blocking_notif_chain).rwsem irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_node_0 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rhashtable_bucket irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock nl_table_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock &obj_hash[i].lock irq_context: 0 init_user_ns.keyring_sem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &p->lock &rcu_state.expedited_wq irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock nl_table_wait.lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock batched_entropy_u8.lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock kfence_freelist_lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &meta->lock irq_context: 0 &p->lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &p->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &p->lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tasklist_lock rcu_read_lock &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_node_0 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock rhashtable_bucket irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex dpm_list_mtx &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sched_map-wait-type-override &rq->__lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &pnettable->lock irq_context: 0 init_user_ns.keyring_sem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock &n->list_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &p->pi_lock irq_context: 0 tomoyo_ss rcu_read_lock stock_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock &n->list_lock &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex smc_ib_devices.mutex irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 cb_lock genl_mutex leds_list_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock cpu_asid_lock irq_context: 0 namespace_sem remove_cache_srcu irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &fq->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 init_user_ns.keyring_sem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 namespace_sem remove_cache_srcu quarantine_lock irq_context: 0 namespace_sem remove_cache_srcu &c->lock irq_context: 0 namespace_sem remove_cache_srcu &n->list_lock irq_context: 0 namespace_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex rcu_node_0 irq_context: 0 tty_mutex &cfs_rq->removed.lock irq_context: 0 namespace_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 tty_mutex quarantine_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 namespace_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#543 irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#18 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &xa->xa_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#543 &rq->__lock irq_context: 0 kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#543 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#5 pcpu_lock irq_context: 0 &type->i_mutex_dir_key#5 percpu_counters_lock irq_context: 0 &type->i_mutex_dir_key#5 pcpu_lock stock_lock irq_context: 0 kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 key_types_sem root_key_user.cons_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem root_key_user.cons_lock root_key_user.cons_lock.wait_lock irq_context: 0 key_types_sem root_key_user.cons_lock.wait_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#543 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#227 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#540 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &cfs_rq->removed.lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 &q->debugfs_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock quarantine_lock irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#538 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex remove_cache_srcu irq_context: 0 nfc_devlist_mutex &dev->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)nfc19_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#16 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#978 irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 irq_context: 0 (wq_completion)wg-kex-wg0#193 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#191 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc22_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc25_nci_rx_wq#14 irq_context: 0 key_types_sem &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &base->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#115 &devlink_port->type_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex.wait_lock irq_context: 0 &data->open_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex triggers_list_lock &rq->__lock irq_context: softirq (&timer.timer) &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc42_nci_cmd_wq#6 irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc42_nci_cmd_wq#6 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 key_types_sem &p->pi_lock &rq->__lock irq_context: 0 key_types_sem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_nat_proto_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 sk_lock-AF_CAIF console_lock console_srcu console_owner_lock irq_context: 0 key_types_sem &type->lock_class#2 root_key_user.lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 &mm->mmap_lock fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy228 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#90 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#90 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#90 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#118 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#118 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)bond0#118 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1000 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1000 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#689 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#689 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#689 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#686 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#684 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &tb->tb6_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#684 &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &obj_hash[i].lock pool_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &ifa->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#312 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#984 irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex quarantine_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem cgroup_threadgroup_rwsem.waiters.lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 cb_lock rcu_read_lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#984 irq_context: 0 &sb->s_type->i_lock_key#16 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_lock_key#16 &dentry->d_lock irq_context: 0 &resv_map->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rfkill_global_mutex pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#690 irq_context: softirq &(&bat_priv->tt.work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 key_types_sem &type->lock_class fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 &dentry->d_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 kn->active#4 fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 kn->active#4 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 kn->active#4 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->lock &base->lock irq_context: 0 &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 &hdev->lock &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)gid-cache-wq &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem &type->lock_class fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 tty_mutex rcu_read_lock &rq->__lock irq_context: 0 tty_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->lock &xa->xa_lock#17 irq_context: 0 &hdev->lock &xa->xa_lock#17 pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 &hdev->lock &list->lock#7 irq_context: 0 &hdev->lock rcu_read_lock &pool->lock irq_context: 0 &hdev->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &hdev->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &hdev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &hdev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &conn->lock#2 irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &obj_hash[i].lock irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &base->lock irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &base->lock &obj_hash[i].lock irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock irq_context: 0 &hdev->lock sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->lock slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &ei->socket.wq.wait irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock &meta->lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_PACKET rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &rq->__lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_QIPCRTR irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock pool_lock#2 irq_context: 0 &mm->mmap_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_QIPCRTR slock-AF_QIPCRTR irq_context: 0 sk_lock-AF_QIPCRTR clock-AF_QIPCRTR irq_context: 0 &mm->mmap_lock &anon_vma->rwsem key irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem percpu_counters_lock irq_context: 0 slock-AF_QIPCRTR irq_context: 0 rfkill_global_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock quarantine_lock irq_context: 0 &f->f_pos_lock key irq_context: 0 &f->f_pos_lock pcpu_lock irq_context: 0 &f->f_pos_lock percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim &cfs_rq->removed.lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim &obj_hash[i].lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim pool_lock#2 irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 vmap_purge_lock rcu_node_0 irq_context: 0 vmap_purge_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 key irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 pcpu_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 percpu_counters_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 stock_lock irq_context: 0 uevent_sock_mutex fs_reclaim irq_context: 0 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 uevent_sock_mutex rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#3 stock_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock &rq->__lock cpu_asid_lock irq_context: 0 &type->i_mutex_dir_key/1 &group->inotify_data.idr_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key/1 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock pool_lock#2 irq_context: 0 unix_gc_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 gdp_mutex &rq->__lock irq_context: 0 gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal &rq->__lock irq_context: 0 sb_internal &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#977 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#3 sb_internal stock_lock irq_context: 0 sb_writers#3 sb_internal key irq_context: 0 sb_writers#3 sb_internal pcpu_lock irq_context: 0 sb_writers#3 sb_internal percpu_counters_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#563 irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#993 irq_context: 0 &type->s_umount_key#46/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &type->s_umount_key#46/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &base->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &conn->lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &base->lock irq_context: softirq rcu_read_lock &br->multicast_lock quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 cb_lock genl_mutex deferred_probe_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 rcu_read_lock &sighand->siglock stock_lock irq_context: 0 &x->wait#27 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 stock_lock irq_context: 0 &ei->xattr_sem irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key &c->lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key &n->list_lock &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &____s->seqcount irq_context: 0 &ei->xattr_sem mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock irq_context: 0 sb_writers#3 jbd2_handle &pcp->lock &zone->lock irq_context: 0 sb_writers#3 jbd2_handle &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &mapping->i_mmap_rwsem irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock pool_lock#2 irq_context: 0 &po->bind_lock irq_context: 0 (wq_completion)wg-kex-wg0#194 irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &journal->j_state_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#233 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &q->sysfs_dir_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex rcu_read_lock (console_sem).lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu pool_lock#2 irq_context: 0 rtnl_mutex fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_QIPCRTR &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_QIPCRTR &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 &xt[i].mutex free_vmap_area_lock quarantine_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock stock_lock irq_context: 0 &hdev->lock &lock->wait_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock pcpu_lock stock_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 cb_lock genl_mutex triggers_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#205 irq_context: 0 (wq_completion)events_unbound (reaper_work).work quarantine_lock irq_context: 0 &dev->mutex triggers_list_lock triggers_list_lock.wait_lock irq_context: 0 &dev->mutex triggers_list_lock &rq->__lock irq_context: 0 &dev->mutex triggers_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex triggers_list_lock.wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 key_types_sem keyring_serialise_link_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#987 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#457 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#457 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#180 irq_context: 0 kn->active#48 fs_reclaim &rq->__lock irq_context: 0 kn->active#48 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#48 &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#233 &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 quarantine_lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc42_nci_cmd_wq#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#28 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#180 irq_context: 0 rtnl_mutex netpoll_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&blkg->free_work) &q->blkcg_mutex &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 pool_lock#2 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &____s->seqcount irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem init_mm.page_table_lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &____s->seqcount#2 irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#2 namespace_sem &c->lock irq_context: 0 &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock irq_context: 0 lock kernfs_idr_lock &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#764 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex &____s->seqcount#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &____s->seqcount#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work quarantine_lock irq_context: 0 &disk->open_mutex rcu_read_lock rcu_node_0 irq_context: 0 &disk->open_mutex rcu_read_lock &rq->__lock irq_context: 0 &disk->open_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex stock_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc44_nci_tx_wq#4 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#695 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#103 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem stock_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc17_nci_rx_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#966 irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &base->lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 bio_slab_lock &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &n->list_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &n->list_lock &c->lock irq_context: 0 &type->i_mutex_dir_key#4 iattr_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 bio_slab_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &obj_hash[i].lock pool_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &q->queue_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 uevent_sock_mutex.wait_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#449 irq_context: 0 &sb->s_type->i_mutex_key#3 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: softirq rcu_read_lock rcu_read_lock &meta->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#446 irq_context: 0 dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#446 &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 tty_mutex &tty->legacy_mutex &rq->__lock cpu_asid_lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &n->list_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &cfs_rq->removed.lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 rcu_node_0 irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &base->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock &base->lock &obj_hash[i].lock irq_context: 0 put_task_map-wait-type-override#2 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#446 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#178 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem pcpu_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock pool_lock#2 irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &base->lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#692 irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 rtnl_mutex rcu_read_lock (console_sem).lock &p->pi_lock &rq->__lock irq_context: 0 &p->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &group->inotify_data.idr_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &group->inotify_data.idr_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 sk_lock-AF_CAIF console_lock console_srcu console_owner irq_context: 0 sk_lock-AF_CAIF console_lock console_srcu console_owner &port_lock_key irq_context: 0 sk_lock-AF_CAIF console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#966 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#804 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &child->perf_event_mutex &rq->__lock cpu_asid_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex rcu_read_lock (console_sem).lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock pool_lock#2 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#175 irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 misc_mtx remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ep->mtx fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &ep->mtx fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ep->mtx fill_pool_map-wait-type-override &c->lock irq_context: 0 &ep->mtx fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex.wait_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex.wait_lock irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex _xmit_ETHER &local->filter_lock &____s->seqcount irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &rdev->wiphy.mtx lweventlist_lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#4 stock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 key irq_context: 0 sb_writers#7 &of->mutex kn->active#4 pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 percpu_counters_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 pcpu_lock stock_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock quarantine_lock irq_context: 0 kn->active#48 &____s->seqcount#2 irq_context: 0 kn->active#48 &____s->seqcount irq_context: 0 sb_writers#8 remove_cache_srcu irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 sb_writers#8 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#8 remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &u->bindlock &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 stock_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 key irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pcpu_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 percpu_counters_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 pcpu_lock stock_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->work_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->work_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->work_lock &base->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &wb->work_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &u->bindlock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#804 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#805 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 fs_reclaim irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &lock->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &xt[i].mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx quarantine_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 pool_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 stock_lock irq_context: 0 &dev->mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 &dev->mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex (inetaddr_chain).rwsem (inetaddr_chain).rwsem.wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#807 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#807 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 rtnl_mutex team->team_lock_key#118 irq_context: 0 rtnl_mutex team->team_lock_key#118 fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rcu_state.expedited_wq irq_context: 0 rtnl_mutex team->team_lock_key#118 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#118 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#118 net_rwsem irq_context: 0 rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock &local->rx_path_lock quarantine_lock irq_context: 0 rcu_read_lock &vma->vm_lock->lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem batched_entropy_u8.lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock rcu_read_lock &____s->seqcount irq_context: 0 &q->sysfs_dir_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#185 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem key irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem percpu_counters_lock irq_context: 0 &sb->s_type->i_lock_key#4 bit_wait_table + i irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 &mm->mmap_lock &anon_vma->rwsem pcpu_lock stock_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#2 rename_lock.seqcount irq_context: 0 &type->i_mutex_dir_key#2 &dentry->d_lock &wq#2 irq_context: 0 &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &sbi->s_orphan_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_raw_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex quarantine_lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 sk_lock-AF_INET6 slock-AF_INET6 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ndev->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_long &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_long &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&hsr->announce_timer) rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ep->mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#194 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &ret->b_state_lock bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#192 irq_context: 0 key_types_sem asymmetric_key_parsers_sem &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &obj_hash[i].lock pool_lock irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 (&timer.timer) irq_context: 0 rtnl_mutex &xa->xa_lock#4 &n->list_lock irq_context: 0 rtnl_mutex &xa->xa_lock#4 &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 &dev->mutex rfkill_global_mutex stock_lock irq_context: 0 &dev->mutex rfkill_global_mutex key irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &pcp->lock &zone->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu quarantine_lock irq_context: 0 &dev->mutex rfkill_global_mutex pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_conntrack_mutex &cfs_rq->removed.lock irq_context: 0 misc_mtx rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start stock_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 misc_mtx &dev->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 sk_lock-AF_INET batched_entropy_u8.lock crngs.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) stock_lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 key irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &rq->__lock irq_context: 0 cb_lock remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex fs_reclaim &rq->__lock irq_context: 0 &data->open_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &xa->xa_lock#15 &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock pcpu_lock irq_context: 0 smc_v4_hashinfo.lock irq_context: 0 &hdev->req_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock sb_writers#5 rcu_node_0 irq_context: softirq &(&gc_work->dwork)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 tty_mutex &tty->legacy_mutex fs_reclaim irq_context: 0 tty_mutex &tty->legacy_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tty_mutex &tty->legacy_mutex pool_lock#2 irq_context: 0 tty_mutex &tty->legacy_mutex &c->lock irq_context: 0 tty_mutex &tty->legacy_mutex stock_lock irq_context: 0 tty_mutex &tty->legacy_mutex tty_ldiscs_lock irq_context: 0 tty_mutex &tty->legacy_mutex &obj_hash[i].lock irq_context: 0 tty_mutex &tty->legacy_mutex &k->list_lock irq_context: 0 tty_mutex &tty->legacy_mutex &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex &k->k_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &o_tty->termios_rwsem/1 irq_context: 0 &tty->legacy_mutex &tty->read_wait irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &tty->legacy_mutex &tty->write_wait irq_context: 0 &tty->legacy_mutex &tty->ctrl.lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 &kernfs_locks->open_file_mutex[count] rcu_node_0 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#317 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock jbd2_handle &journal->j_wait_updates irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 sb_internal jbd2_handle &obj_hash[i].lock pool_lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock init_task.mems_allowed_seq.seqcount irq_context: 0 &tty->legacy_mutex redirect_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->files_lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 tasklist_lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &buf->lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->write_wait irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->read_wait irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 vmap_area_lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 &obj_hash[i].lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 purge_vmap_area_lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 purge_vmap_area_lock pool_lock#2 irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 pool_lock#2 irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 &rq->__lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem tty_ldiscs_lock irq_context: 0 (wq_completion)events free_ipc_work &obj_hash[i].lock pool_lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &obj_hash[i].lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem pool_lock#2 irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ctrl.lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex &____s->seqcount#2 irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &____s->seqcount#2 irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss pcpu_lock stock_lock irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem vmap_area_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex pool_lock#2 irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem &obj_hash[i].lock irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem purge_vmap_area_lock irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &x->wait#11 irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock batched_entropy_u8.lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock kfence_freelist_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex fs_reclaim irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem purge_vmap_area_lock pool_lock#2 irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem pool_lock#2 irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &obj_hash[i].lock pool_lock irq_context: 0 misc_mtx &meta->lock irq_context: 0 &mm->mmap_lock sb_writers#5 &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag6_mutex nf_hook_mutex &c->lock irq_context: 0 rtnl_mutex &rnp->exp_wq[2] irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 tty_ldiscs_lock irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 (work_completion)(&buf->work) irq_context: 0 (wq_completion)events (work_completion)(&tty->hangup_work)#2 &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu pcpu_lock stock_lock irq_context: 0 &tty->ldisc_sem/1 irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock irq_context: 0 pernet_ops_rwsem cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &wg->device_update_lock &handshake->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim stock_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim pcpu_lock irq_context: 0 kn->active#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim percpu_counters_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#696 irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 &tty->legacy_mutex tty_mutex.wait_lock irq_context: 0 &tty->legacy_mutex &p->pi_lock irq_context: 0 &tty->legacy_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &tty->legacy_mutex &p->pi_lock &rq->__lock irq_context: 0 uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET key irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock &ep->wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->legacy_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->legacy_mutex &rq->__lock irq_context: 0 &tty->legacy_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.barrier_mutex rcu_state.barrier_mutex.wait_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex &n->list_lock irq_context: 0 tty_mutex &tty->legacy_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock percpu_counters_lock irq_context: 0 key_types_sem remove_cache_srcu irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 tty_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 sb_writers#7 &mm->mmap_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 &sb->s_type->i_mutex_key#8 mapping.invalidate_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 fs_reclaim stock_lock irq_context: 0 fs_reclaim pcpu_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#8 mmu_notifier_invalidate_range_start irq_context: 0 &sb->s_type->i_mutex_key#8 jbd2_handle irq_context: 0 &sb->s_type->i_mutex_key#8 jbd2_handle mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events key_gc_work &meta->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex lock kernfs_idr_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex lock kernfs_idr_lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#8 jbd2_handle &ei->i_raw_lock irq_context: 0 &sb->s_type->i_mutex_key#8 jbd2_handle &journal->j_wait_updates irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &base->lock &obj_hash[i].lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 irq_context: 0 (wq_completion)events key_gc_work kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#697 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc16_nci_cmd_wq#18 irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 stock_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &dentry->d_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock rename_lock.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc45_nci_cmd_wq#4 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#233 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss rcu_read_lock mount_lock.seqcount irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#233 irq_context: 0 &hdev->req_lock (wq_completion)hci4#4 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tomoyo_ss &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &dentry->d_lock &wq#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &sbinfo->stat_lock irq_context: 0 (wq_completion)events (debug_obj_work).work rcu_node_0 irq_context: 0 (wq_completion)events (debug_obj_work).work &rcu_state.expedited_wq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 pool_lock#2 irq_context: 0 (wq_completion)events (debug_obj_work).work &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &c->lock irq_context: 0 misc_mtx &x->wait#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) rcu_node_0 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc40_nci_rx_wq#8 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock percpu_counters_lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 namespace_sem pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &local->active_txq_lock[i] irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx (work_completion)(&sta->drv_deliver_wk) irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &cfs_rq->removed.lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#201 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#201 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#197 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#196 irq_context: 0 &hdev->lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#75 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#75 irq_context: 0 console_lock console_srcu &rq->__lock irq_context: 0 console_lock console_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#73 irq_context: 0 &hdev->lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#72 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex kfence_freelist_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &xa->xa_lock#15 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &xa->xa_lock#15 pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &obj_hash[i].lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex &obj_hash[i].lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#494 irq_context: 0 (wq_completion)nfc40_nci_rx_wq#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc32_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#494 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#491 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &n->list_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#489 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &sem->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &n->list_lock &c->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &sb->s_type->i_lock_key#5 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &s->s_inode_list_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 tk_core.seq.seqcount irq_context: 0 sb_writers &type->i_mutex_dir_key#2 batched_entropy_u32.lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &xattrs->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &simple_offset_xa_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &xattrs->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 smack_known_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 smack_known_lock &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &sb->s_type->i_lock_key#5 &dentry->d_lock irq_context: 0 sk_lock-AF_ALG &n->list_lock irq_context: 0 sk_lock-AF_ALG &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#495 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#495 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#495 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#495 irq_context: 0 tty_mutex fs_reclaim &rq->__lock irq_context: 0 &bdi->cgwb_release_mutex &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#492 irq_context: 0 &hdev->lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) &hwstats->hwsdev_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#490 irq_context: 0 &hdev->lock &xa->xa_lock#17 &n->list_lock irq_context: 0 &hdev->lock &xa->xa_lock#17 &n->list_lock &c->lock irq_context: 0 &hdev->lock remove_cache_srcu irq_context: 0 &hdev->lock remove_cache_srcu quarantine_lock irq_context: 0 &hdev->lock remove_cache_srcu &c->lock irq_context: 0 &hdev->lock remove_cache_srcu &n->list_lock irq_context: 0 &hdev->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#822 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#76 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#76 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#76 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#76 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock pcpu_lock stock_lock irq_context: 0 rcu_read_lock &sighand->siglock batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &ndev->req_lock &wq->mutex &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock &wq->mutex &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &rq->__lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &n->list_lock irq_context: 0 &dev->mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock sb_writers#4 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#74 irq_context: 0 &type->s_umount_key#30 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#73 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#32 irq_context: 0 &sb->s_type->i_mutex_key#8 &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: softirq &(&forw_packet_aggr->delayed_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &tbl->lock init_task.mems_allowed_seq.seqcount irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#32 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#31 irq_context: 0 &sb->s_type->i_mutex_key#8 &wb->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rnp->exp_wq[0] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#8 &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &meta->lock irq_context: 0 sb_writers#3 fs_reclaim irq_context: 0 sb_writers#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_read_lock &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock subsys mutex#24 irq_context: 0 tty_mutex &tty->legacy_mutex console_lock subsys mutex#24 &k->k_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 rcu_read_lock &____s->seqcount#4 irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &n->list_lock &c->lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx net_rwsem &rq->__lock irq_context: 0 sb_writers#3 &zone->lock irq_context: 0 sb_writers#3 &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &mm->mmap_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#30 irq_context: 0 sb_writers#3 &mm->mmap_lock &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex &c->lock irq_context: 0 sb_writers#3 &p->alloc_lock irq_context: 0 sb_writers#3 rcu_read_lock &____s->seqcount#3 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex kfence_freelist_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &base->lock irq_context: 0 misc_mtx nfc_devlist_mutex subsys mutex#39 &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex subsys mutex#39 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex subsys mutex#39 &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex subsys mutex#39 &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex subsys mutex#39 pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#6 pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss rcu_read_lock rcu_node_0 irq_context: softirq rcu_callback rcu_read_lock &q->mq_freeze_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 rcu_read_lock rcu_read_lock mount_lock.seqcount irq_context: 0 misc_mtx (wq_completion)nfc4_nci_tx_wq#74 irq_context: 0 (wq_completion)events_unbound (stats_flush_dwork).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#66 binderfs_minors_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rcu_state.expedited_wq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback rcu_read_lock &q->mq_freeze_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#66 binderfs_minors_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#75 irq_context: 0 &type->s_umount_key#40 &rq->__lock irq_context: 0 &type->s_umount_key#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PACKET key irq_context: 0 sk_lock-AF_PACKET pcpu_lock irq_context: 0 sk_lock-AF_PACKET percpu_counters_lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (crypto_chain).rwsem quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rcu_state.expedited_wq irq_context: 0 sb_writers#3 rcu_read_lock rcu_read_lock rcu_read_lock rename_lock.seqcount irq_context: 0 &dev->mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#191 irq_context: 0 dev_addr_sem &rq->__lock irq_context: 0 dev_addr_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#118 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#118 &tn->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#8 &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_lock_key &xa->xa_lock#8 pool_lock#2 irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &ndev->lock &ifa->lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock batched_entropy_u32.lock crngs.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 &n->list_lock irq_context: 0 (work_completion)(&smc->connect_work) irq_context: 0 sk_lock-AF_SMC irq_context: 0 sk_lock-AF_SMC slock-AF_SMC irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_SMC smc_v4_hashinfo.lock irq_context: 0 slock-AF_SMC irq_context: 0 &smc->clcsock_release_lock irq_context: 0 &smc->clcsock_release_lock k-sk_lock-AF_INET irq_context: 0 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET#2 irq_context: 0 &smc->clcsock_release_lock k-sk_lock-AF_INET k-clock-AF_INET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dir->lock#2 &base->lock &obj_hash[i].lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &cfs_rq->removed.lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex.wait_lock irq_context: 0 misc_mtx cpu_hotplug_lock &p->pi_lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem rcu_read_lock &rq->__lock irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &smc->clcsock_release_lock k-sk_lock-AF_INET k-slock-AF_INET#2 &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock k-slock-AF_INET#2 irq_context: 0 &smc->clcsock_release_lock pool_lock#2 irq_context: 0 &smc->clcsock_release_lock &dir->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &smc->clcsock_release_lock &obj_hash[i].lock irq_context: 0 &smc->clcsock_release_lock stock_lock irq_context: 0 &smc->clcsock_release_lock &sb->s_type->i_lock_key#8 irq_context: 0 &smc->clcsock_release_lock &xa->xa_lock#8 irq_context: 0 rtnl_mutex remove_cache_srcu key irq_context: 0 rtnl_mutex remove_cache_srcu pcpu_lock irq_context: 0 rtnl_mutex remove_cache_srcu percpu_counters_lock irq_context: 0 &smc->clcsock_release_lock &fsnotify_mark_srcu irq_context: 0 sk_lock-AF_SMC clock-AF_SMC irq_context: 0 rtnl_mutex team->team_lock_key#118 _xmit_ETHER irq_context: 0 (wq_completion)wg-kex-wg2#191 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#191 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 bt_proto_lock sco_sk_list.lock irq_context: 0 prog_idr_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &msk->pm.lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex lock kernfs_idr_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 irq_context: 0 &fc->uapi_mutex remove_cache_srcu &rq->__lock irq_context: 0 &dir->lock#2 quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) quarantine_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &n->list_lock irq_context: 0 &fc->uapi_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 rtnl_mutex team->team_lock_key#118 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#118 &dir->lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex team->team_lock_key#118 input_pool.lock irq_context: 0 key_types_sem remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (crypto_chain).rwsem &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 rcu_node_0 irq_context: 0 &smc->clcsock_release_lock quarantine_lock irq_context: 0 &type->s_umount_key#73 &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#73 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 &wb->work_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 &wb->work_lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 &wb->work_lock &base->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 &wb->work_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex quarantine_lock irq_context: 0 file_rwsem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 misc_mtx wq_pool_mutex &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 stock_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events fqdir_free_work &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 cb_lock genl_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 rtnl_mutex (console_sem).lock &p->pi_lock irq_context: 0 rtnl_mutex (console_sem).lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex (console_sem).lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem remove_cache_srcu irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim pool_lock#2 irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem remove_cache_srcu quarantine_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem remove_cache_srcu &c->lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem remove_cache_srcu &n->list_lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &lock->wait_lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_node_0 irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &cfs_rq->removed.lock irq_context: 0 kn->active#14 &kernfs_locks->open_file_mutex[count] &obj_hash[i].lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start key irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 tomoyo_ss mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &p->lock stock_lock irq_context: 0 &p->lock key irq_context: 0 &p->lock pcpu_lock irq_context: 0 &p->lock percpu_counters_lock irq_context: 0 &p->lock pcpu_lock stock_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &type->i_mutex_dir_key#4 iattr_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &____s->seqcount#2 irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 lock#4 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu quarantine_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &c->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &n->list_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#47 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock lock kernfs_idr_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[2] irq_context: 0 rtnl_mutex uevent_sock_mutex &rcu_state.expedited_wq irq_context: 0 rtnl_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 rtnl_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &ssp->srcu_sup->srcu_cb_mutex pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock kfence_freelist_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &rq->__lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &journal->j_state_lock &journal->j_wait_transaction_locked irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &____s->seqcount irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#186 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#825 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex pool_lock irq_context: 0 rtnl_mutex fs_reclaim key irq_context: 0 rtnl_mutex fs_reclaim pcpu_lock irq_context: 0 rtnl_mutex fs_reclaim percpu_counters_lock irq_context: 0 rtnl_mutex fs_reclaim pool_lock#2 irq_context: 0 namespace_sem pcpu_alloc_mutex.wait_lock irq_context: 0 namespace_sem &p->pi_lock irq_context: 0 namespace_sem &p->pi_lock &rq->__lock irq_context: 0 namespace_sem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#21/1 pcpu_alloc_mutex.wait_lock irq_context: 0 &type->s_umount_key#21/1 &p->pi_lock irq_context: 0 (work_completion)(&local->tx_work) &rq->__lock irq_context: 0 (work_completion)(&local->tx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 kn->active#4 fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 sb_writers#3 sb_internal &cfs_rq->removed.lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock kfence_freelist_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN &____s->seqcount#2 irq_context: 0 sk_lock-AF_CAN &____s->seqcount irq_context: 0 sb_writers#3 &f->f_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#8 &xa->xa_lock#15 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#8 &obj_hash[i].lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &xa->xa_lock#8 stock_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 &rq->__lock irq_context: softirq (&net->sctp.addr_wq_timer) &net->sctp.addr_wq_lock kfence_freelist_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex.wait_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu &c->lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex &p->pi_lock irq_context: 0 uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &mm->mmap_lock fs_reclaim irq_context: 0 sb_writers#3 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &mm->mmap_lock &____s->seqcount irq_context: 0 sb_writers#3 &mm->mmap_lock stock_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock krc.lock &base->lock irq_context: softirq rcu_callback put_task_map-wait-type-override css_set_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock &____s->seqcount irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem &rq->__lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#8 irq_context: 0 sb_writers#3 &mm->mmap_lock &sb->s_type->i_lock_key irq_context: 0 sb_writers#3 &mm->mmap_lock &info->lock irq_context: 0 sb_writers#3 &mm->mmap_lock lock#4 irq_context: 0 key_types_sem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#52 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &pcp->lock &zone->lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rq->__lock cpu_asid_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx &dev->mutex &pcp->lock &zone->lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &cfs_rq->removed.lock irq_context: 0 &ep->mtx fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem rcu_read_lock &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#8 stock_lock irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#8 pool_lock#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rq->__lock cpu_asid_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem stock_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex rcu_state.exp_wake_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnn->pndevs.lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 tomoyo_ss remove_cache_srcu pool_lock#2 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem key irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 &dentry->d_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem percpu_counters_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pcpu_lock stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 &lru->node[i].lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 rcu_read_lock &dentry->d_lock irq_context: 0 bt_proto_lock &obj_hash[i].lock pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#63 &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex &ht->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx pcpu_alloc_mutex rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_CAN &priv->lock &c->lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &x->wait#2 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sb_writers#3 sb_internal jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&w->w) nfc_devlist_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers &type->i_mutex_dir_key#2 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 rtnl_mutex team->team_lock_key#118 rcu_read_lock &ndev->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem key irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem percpu_counters_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem pcpu_lock stock_lock irq_context: 0 &ep->mtx &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock stock_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock fs_reclaim &rq->__lock irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#118 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#118 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#118 nl_table_wait.lock irq_context: 0 &disk->open_mutex pcpu_lock irq_context: 0 &disk->open_mutex percpu_counters_lock irq_context: 0 &disk->open_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#118 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tt.commit_lock &bat_priv->tvlv.container_list_lock quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex team->team_lock_key#118 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex team->team_lock_key#118 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#118 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq drivers/net/wireguard/ratelimiter.c:20 rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key#118 &in_dev->mc_tomb_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex sched_map-wait-type-override &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#118 &im->lock irq_context: 0 sb_writers &type->i_mutex_dir_key/1 tomoyo_ss &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex.wait_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &p->pi_lock irq_context: 0 sk_lock-AF_NETLINK &f->f_lock fasync_lock &new->fa_lock irq_context: softirq (&peer->timer_retransmit_handshake) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_NETLINK &f->f_lock fasync_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK &f->f_lock fasync_lock pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock krc.lock &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#118 cbs_list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex &idev->mc_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex subsys mutex#40 &rq->__lock irq_context: 0 &dev->mutex subsys mutex#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work ima_keys_lock irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work &obj_hash[i].lock irq_context: 0 (wq_completion)events (ima_keys_delayed_work).work pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&(&rdev->dfs_update_channels_wk)->work) rtnl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key#118 &ndev->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 &vma->vm_lock->lock batched_entropy_u8.lock crngs.lock irq_context: 0 rtnl_mutex &tn->lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tn->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &c->lock irq_context: 0 rtnl_mutex &tn->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&group->avgs_work)->work) &group->avgs_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fill_pool_map-wait-type-override &rq->__lock &cfs_rq->removed.lock irq_context: 0 init_user_ns.keyring_sem rcu_node_0 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex batched_entropy_u8.lock crngs.lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 sk_lock-AF_PPPOX chan_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem pool_lock#2 irq_context: 0 sk_lock-AF_PPPOX &obj_hash[i].lock irq_context: 0 sk_lock-AF_PPPOX &x->wait#2 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#77 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#202 irq_context: 0 cb_lock rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 key_types_sem &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#202 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#198 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#197 irq_context: 0 &fc->uapi_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#496 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#496 irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#493 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#493 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#493 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#491 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_writers#5 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock sb_writers#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex subsys mutex#39 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_state_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &mm->mmap_lock sb_writers#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &journal->j_state_lock &journal->j_wait_transaction_locked &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock fs_reclaim &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->notification_waitq &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx pcpu_alloc_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#78 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock kfence_freelist_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) lock pidmap_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 &xt[i].mutex fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 irq_context: 0 loop_validate_mutex &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &xt[i].mutex remove_cache_srcu pcpu_lock stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock fs_reclaim irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock &c->lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock pool_lock#2 irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock root_key_user.lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock key_user_lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock crngs.lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock key_serial_lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex keyring_name_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock key_construction_mutex pool_lock#2 irq_context: 0 key_types_sem keyring_serialise_link_lock irq_context: 0 (wq_completion)events key_gc_work irq_context: 0 (wq_completion)events key_gc_work key_serial_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)events key_gc_work &obj_hash[i].lock irq_context: 0 (wq_completion)events key_gc_work &x->wait#2 irq_context: 0 (wq_completion)events key_gc_work sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events key_gc_work &rq->__lock irq_context: 0 (wq_completion)events key_gc_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events key_gc_work keyring_name_lock irq_context: 0 (wq_completion)events key_gc_work rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events key_gc_work rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events key_gc_work pool_lock#2 irq_context: 0 (wq_completion)events key_gc_work root_key_user.lock irq_context: 0 &dev->mutex device_links_lock &rq->__lock irq_context: 0 &dev->mutex device_links_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events key_gc_work &rq->__lock cpu_asid_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#51 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &bat_priv->tt.changes_list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &bat_priv->tt.changes_list_lock &obj_hash[i].lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock &rq->__lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &bat_priv->tt.changes_list_lock pool_lock#2 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock pool_lock#2 irq_context: 0 loop_validate_mutex &lo->lo_mutex &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock &____s->seqcount#2 irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#78 irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &____s->seqcount#2 irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#76 irq_context: 0 sb_writers#3 fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev_addr_list_lock_key#2/1 _xmit_ETHER fill_pool_map-wait-type-override pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 sb_lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle percpu_counters_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#75 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#75 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#75 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx rfkill_global_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 rcu_read_lock pcpu_lock stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: softirq &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 key_types_sem kfence_freelist_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 fs_reclaim irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem batched_entropy_u8.lock irq_context: 0 rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: softirq &(&bat_priv->dat.work)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: softirq (&rxnet->peer_keepalive_timer) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 &dentry->d_lock irq_context: 0 sb_writers#3 jbd2_handle stock_lock irq_context: 0 sb_writers#3 jbd2_handle key irq_context: 0 sb_writers#3 jbd2_handle pcpu_lock irq_context: 0 sb_writers#3 jbd2_handle percpu_counters_lock irq_context: 0 sb_writers#3 jbd2_handle pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 &lru->node[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 rcu_read_lock &dentry->d_lock irq_context: 0 &dev->mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex nf_hook_mutex &rq->__lock irq_context: 0 rtnl_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#113 irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex console_owner_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex console_owner irq_context: 0 cb_lock genl_mutex rtnl_mutex console_lock console_srcu console_owner_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex console_lock console_srcu console_owner irq_context: 0 cb_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nl_table_lock nl_table_wait.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key krc.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &bridge_netdev_addr_lock_key irq_context: softirq rcu_callback pcpu_lock stock_lock irq_context: 0 sk_lock-AF_PPPOX &rq->__lock irq_context: 0 sk_lock-AF_PPPOX &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 fs_reclaim &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ht->mutex quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 irq_context: 0 &p->lock &of->mutex kn->active#4 fs_reclaim pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#203 irq_context: 0 &type->s_umount_key#30 &rcu_state.expedited_wq irq_context: 0 &type->s_umount_key#30 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->s_umount_key#30 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex.wait_lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#203 irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 tomoyo_ss rcu_read_lock key irq_context: 0 tomoyo_ss rcu_read_lock pcpu_lock irq_context: 0 tomoyo_ss rcu_read_lock percpu_counters_lock irq_context: 0 tomoyo_ss rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)events pcpu_balance_work &p->pi_lock irq_context: 0 &type->s_umount_key#29 &rq->__lock irq_context: 0 key_types_sem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock pool_lock#2 irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&pool->idle_timer) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock &base->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &pool->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &dev->mutex &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 fs_reclaim irq_context: 0 &dev->mutex kn->active#4 irq_context: 0 &dev->mutex kn->active#4 &root->deactivate_waitq irq_context: 0 &dev->mutex kn->active#4 &rq->__lock irq_context: 0 &dev->mutex kn->active#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 &xt[i].mutex rcu_read_lock stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock &____s->seqcount irq_context: 0 &xt[i].mutex rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx cpu_hotplug_lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &tty->legacy_mutex &port->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) irq_context: 0 rtnl_mutex gdp_mutex &cfs_rq->removed.lock irq_context: softirq (&peer->timer_persistent_keepalive) init_task.mems_allowed_seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex j1939_netdev_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex j1939_netdev_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock &rq->__lock irq_context: 0 rtnl_mutex gdp_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 misc_mtx nfc_devlist_mutex &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnettable->lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &pnettable->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback stock_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 pernet_ops_rwsem rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#13 &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 &type->s_umount_key#43 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &rq->__lock cpu_asid_lock irq_context: 0 &type->s_umount_key#43 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 &type->i_mutex_dir_key#3 stock_lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc27_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc27_nci_rx_wq#13 irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)nfc27_nci_tx_wq#13 irq_context: 0 (wq_completion)wg-kex-wg2#192 irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#12 irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)nfc3_nci_rx_wq#199 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#198 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#497 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#497 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#494 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy225 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#958 irq_context: 0 &dev->mutex uevent_sock_mutex quarantine_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#492 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#564 irq_context: 0 key_types_sem remove_cache_srcu &n->list_lock irq_context: softirq &(&conn->disc_work)->timer irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#185 irq_context: 0 &child->perf_event_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rq->__lock cpu_asid_lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#204 irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock sb_writers#14 irq_context: 0 &f->f_pos_lock sb_writers#14 &rq->__lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &n->list_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#14 fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#14 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#14 pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#14 &mm->mmap_lock irq_context: 0 &f->f_pos_lock sb_writers#14 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock sb_writers#14 &c->lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET slock-AF_INET#2 &tcp_hashinfo.bhash[i].lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#958 &rq->__lock irq_context: 0 &hdev->req_lock (wq_completion)hci4#4 &rq->__lock irq_context: 0 &hdev->req_lock (wq_completion)hci4#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock sched_map-wait-type-override &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#959 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#959 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#959 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock batched_entropy_u8.lock irq_context: 0 &data->open_mutex quarantine_lock irq_context: 0 (wq_completion)hci1#7 irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &n->list_lock irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 sk_lock-AF_INET6 tcpv6_prot_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#959 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#953 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#953 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &hdev->req_lock &hdev->lock quarantine_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex _xmit_ETHER console_owner irq_context: 0 (wq_completion)nfc2_nci_rx_wq#954 &rq->__lock irq_context: softirq (&app->periodic_timer) &app->lock#2 fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&app->periodic_timer) &app->lock#2 fill_pool_map-wait-type-override &c->lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock cpu_asid_lock irq_context: softirq (&app->periodic_timer) &app->lock#2 fill_pool_map-wait-type-override &n->list_lock irq_context: softirq (&app->periodic_timer) &app->lock#2 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&app->periodic_timer) &app->lock#2 fill_pool_map-wait-type-override pool_lock irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex devnet_rename_sem stock_lock irq_context: 0 rtnl_mutex devnet_rename_sem key irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock key irq_context: 0 &ndev->req_lock pcpu_lock irq_context: 0 &ndev->req_lock percpu_counters_lock irq_context: 0 &ndev->req_lock pcpu_lock stock_lock irq_context: 0 &ndev->req_lock pool_lock#2 irq_context: 0 rtnl_mutex devnet_rename_sem pcpu_lock irq_context: 0 rtnl_mutex devnet_rename_sem percpu_counters_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &xa->xa_lock#15 &n->list_lock irq_context: 0 rtnl_mutex devnet_rename_sem pcpu_lock stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &xa->xa_lock#15 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#971 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#971 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#971 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#962 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#962 irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#955 irq_context: 0 kn->active#18 fs_reclaim &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#358 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem device_links_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#949 irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#963 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#963 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#963 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &br->lock lweventlist_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock &p->pi_lock irq_context: softirq &(&hdev->cmd_timer)->timer rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#14 &rq->__lock irq_context: 0 sk_lock-AF_KCM irq_context: 0 sk_lock-AF_KCM slock-AF_KCM irq_context: 0 sk_lock-AF_KCM clock-AF_KCM irq_context: 0 tty_mutex &tty->legacy_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 slock-AF_KCM irq_context: 0 (work_completion)(&kcm->tx_work) irq_context: 0 cb_lock genl_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 cb_lock genl_mutex triggers_list_lock &rq->__lock irq_context: 0 pernet_ops_rwsem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 fs_reclaim pool_lock#2 irq_context: 0 sk_lock-AF_CAN remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &mux->rx_lock rlock-AF_KCM irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#15 irq_context: 0 sb_writers#3 jbd2_handle &ret->b_state_lock bit_wait_table + i irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#770 irq_context: 0 sk_lock-AF_INET batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)bond0#112 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#777 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock quarantine_lock irq_context: 0 &hdev->req_lock &hdev->lock &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex.wait_lock irq_context: 0 &type->s_umount_key#73 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#15 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&p->wq) purge_vmap_area_lock kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock krc.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#963 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#956 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#953 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#967 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#565 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex crngs.lock irq_context: 0 &fc->uapi_mutex krc.lock irq_context: 0 &fc->uapi_mutex batched_entropy_u8.lock irq_context: 0 &fc->uapi_mutex kfence_freelist_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex krc.lock &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex krc.lock &base->lock irq_context: 0 &fc->uapi_mutex krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#565 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 pcpu_alloc_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 shrinker_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 list_lrus_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 sb_lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex pcpu_alloc_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#565 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#565 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#562 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#560 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock quarantine_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 (console_sem).lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 console_lock console_srcu console_owner_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rnp->exp_wq[0] irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 console_lock console_srcu console_owner irq_context: 0 kn->active#16 remove_cache_srcu irq_context: 0 kn->active#16 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#16 remove_cache_srcu &c->lock irq_context: 0 kn->active#16 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#16 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 console_lock console_srcu console_owner &port_lock_key irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 console_lock console_srcu console_owner console_owner_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex rcu_node_0 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &dentry->d_lock fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#967 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#960 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#954 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#102 irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#206 irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#203 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 &x->wait#23 irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &____s->seqcount irq_context: 0 &fc->uapi_mutex unnamed_dev_ida.xa_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#2 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#829 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#831 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 remove_cache_srcu irq_context: 0 sb_writers#4 remove_cache_srcu quarantine_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#14 &rq->__lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#875 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#864 irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#12 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem device_links_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#790 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock batched_entropy_u8.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock kfence_freelist_lock irq_context: softirq rcu_callback fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_callback fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_callback fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#785 irq_context: 0 &mm->mmap_lock remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#782 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#791 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#791 irq_context: 0 &fc->uapi_mutex list_lrus_mutex irq_context: 0 &fc->uapi_mutex &xa->xa_lock#15 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#786 irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#783 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#204 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#964 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#964 &rq->__lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &fc->uapi_mutex bit_wait_table + i irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &xa->xa_lock#15 &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex &xa->xa_lock#15 &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 bit_wait_table + i irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 bit_wait_table + i &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 bit_wait_table + i &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &lock->wait_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#68 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC kfence_freelist_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#68 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#68 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex &rnp->exp_wq[1] irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#792 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#792 irq_context: 0 key_types_sem remove_cache_srcu &obj_hash[i].lock irq_context: 0 key_types_sem remove_cache_srcu pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag4_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc33_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc33_nci_cmd_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc33_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc33_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc17_nci_rx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc17_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc17_nci_tx_wq#14 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock batched_entropy_u8.lock irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#115 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#358 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#350 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock (&timer.timer) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#566 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#566 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#566 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#566 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#787 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#787 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#136 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#140 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock (worker)->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_TIPC &tn->nametbl_lock &service->lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#359 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#785 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#785 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#140 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#140 &rq->__lock irq_context: 0 (wq_completion)nfc17_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET/1 irq_context: 0 sk_lock-AF_INET/1 slock-AF_INET#2 irq_context: 0 sk_lock-AF_INET/1 rlock-AF_INET irq_context: 0 sk_lock-AF_INET/1 &list->lock#19 irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 key irq_context: 0 cb_lock genl_mutex rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock kernfs_idr_lock &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#7 tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex cpu_hotplug_lock wq_pool_mutex &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx hrtimer_bases.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &br->hash_lock &____s->seqcount#2 irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET/1 irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET/1 &sctp_ep_hashtable[i].lock irq_context: 0 &data->open_mutex subsys mutex#74 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET/1 &obj_hash[i].lock irq_context: 0 (wq_completion)kblockd (work_completion)(&(&q->requeue_work)->work) rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fill_pool_map-wait-type-override pool_lock irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET/1 pool_lock#2 irq_context: 0 &pipe->mutex/1 pcpu_lock irq_context: 0 &pipe->mutex/1 percpu_counters_lock irq_context: 0 kn->active#49 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 rcu_read_lock &dentry->d_lock irq_context: 0 &pipe->mutex/1 pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET/1 clock-AF_INET irq_context: 0 (wq_completion)nfc18_nci_cmd_wq#16 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) cgroup_threadgroup_rwsem rcu_node_0 irq_context: 0 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &im->lock init_task.mems_allowed_seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc33_nci_rx_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc33_nci_rx_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#136 &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&sdp->work) &rq->__lock cpu_asid_lock irq_context: 0 &root->kernfs_iattr_rwsem stock_lock irq_context: 0 &root->kernfs_iattr_rwsem key irq_context: 0 &root->kernfs_iattr_rwsem pcpu_lock irq_context: 0 &root->kernfs_iattr_rwsem percpu_counters_lock irq_context: 0 &root->kernfs_iattr_rwsem pcpu_lock stock_lock irq_context: 0 &root->kernfs_iattr_rwsem &____s->seqcount irq_context: 0 &fc->uapi_mutex remove_cache_srcu quarantine_lock irq_context: 0 &fc->uapi_mutex remove_cache_srcu &c->lock irq_context: 0 &fc->uapi_mutex remove_cache_srcu &n->list_lock irq_context: 0 &fc->uapi_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &u->iolock stock_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 remove_cache_srcu pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#787 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#784 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#374 irq_context: 0 &tty->ldisc_sem &tty->termios_rwsem kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#965 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#359 irq_context: 0 &xt[i].mutex remove_cache_srcu &meta->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1001 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#136 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#766 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#352 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &dev->tx_global_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#801 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#374 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#366 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#768 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#562 irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#103 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#10 irq_context: 0 &f->f_pos_lock sb_writers#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#48 irq_context: 0 &f->f_pos_lock sb_writers#4 sysctl_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#96 irq_context: 0 &f->f_pos_lock sb_writers#4 fs_reclaim irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex uevent_sock_mutex.wait_lock irq_context: 0 &f->f_pos_lock sb_writers#4 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nfc3_nci_tx_wq#364 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock rcu_read_lock &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &bgl->locks[i].lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &mm->mmap_lock irq_context: 0 &f->f_pos_lock sb_writers#4 &mm->mmap_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &folio_wait_table[i] irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock irq_context: 0 kn->active#14 &____s->seqcount#2 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock &cfs_rq->removed.lock irq_context: 0 key_types_sem keyring_serialise_link_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &obj_hash[i].lock irq_context: 0 &p->lock &of->mutex kn->active#18 &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#57 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_sockopt_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#563 irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock ptlock_ptr(ptdesc)#2 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#141 irq_context: 0 (wq_completion)wg-kex-wg0#213 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#141 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc17_nci_cmd_wq#12 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#965 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#958 irq_context: 0 &type->s_umount_key#57 &x->wait#23 irq_context: 0 &type->s_umount_key#57 shrinker_mutex irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem quarantine_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 key#14 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#12 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) put_task_map-wait-type-override#3 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#141 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_tx_wq#12 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#952 irq_context: 0 &mm->mmap_lock remove_cache_srcu key irq_context: 0 &type->s_umount_key#57 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#57 rename_lock.seqcount irq_context: 0 pernet_ops_rwsem fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) put_task_map-wait-type-override#3 pool_lock#2 irq_context: 0 pernet_ops_rwsem fs_reclaim &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) put_task_map-wait-type-override#3 stock_lock irq_context: 0 kn->active#50 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#57 &dentry->d_lock irq_context: 0 &type->s_umount_key#57 rcu_read_lock &dentry->d_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &type->s_umount_key#57 &sb->s_type->i_lock_key#16 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_node_0 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) rcu_read_lock &c->lock irq_context: 0 &type->s_umount_key#57 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#57 &xa->xa_lock#8 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC &rxnet->local_mutex &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&peer->timer_persistent_keepalive) &pcp->lock &zone->lock irq_context: softirq (&peer->timer_persistent_keepalive) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex sched_map-wait-type-override &rq->__lock irq_context: softirq (&peer->timer_send_keepalive) &____s->seqcount#2 irq_context: softirq (&peer->timer_send_keepalive) &pcp->lock &zone->lock irq_context: softirq (&peer->timer_send_keepalive) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &mapping->i_mmap_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&peer->timer_send_keepalive) &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override pool_lock#2 irq_context: 0 &type->s_umount_key#57 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#57 &dentry->d_lock/1 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &vlan_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem fs_reclaim stock_lock irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#11 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx (console_sem).lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) pool_lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &rcu_state.expedited_wq irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rq->__lock cpu_asid_lock irq_context: 0 &u->iolock &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &u->lock rcu_read_lock &p->pi_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex stock_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &c->lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#141 irq_context: 0 &u->lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex stack_depot_init_mutex irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx gdp_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex crngs.lock irq_context: 0 cb_lock genl_mutex (inetaddr_chain).rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex (inetaddr_chain).rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &c->lock irq_context: 0 &u->lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex krc.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &dir->lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex quarantine_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex gdp_mutex.wait_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[1] irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock _xmit_ETHER fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh key#20 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &xa->xa_lock#15 &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#801 &rq->__lock irq_context: 0 &disk->open_mutex &lock->wait_lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&ndev_work->work) devices_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &caifn->caifdevs.lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#9 &n->list_lock irq_context: 0 sb_writers#9 &n->list_lock &c->lock irq_context: 0 nf_hook_mutex &rq->__lock irq_context: 0 nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cgroup_threadgroup_rwsem &obj_hash[i].lock pool_lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 sb_writers &sb->s_type->i_mutex_key#4 rcu_read_lock &rq->__lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq &(&net->ipv6.addr_chk_work)->timer rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#407 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#884 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#884 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#884 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#13 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex stock_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex key irq_context: 0 rtnl_mutex rcu_state.exp_mutex pcpu_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex percpu_counters_lock irq_context: 0 rtnl_mutex rcu_state.exp_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 &pipe->wr_wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#12 irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 &type->i_mutex_dir_key#4 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc18_nci_rx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#10 irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 rtnl_mutex team->team_lock_key#118 sysfs_symlink_target_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &sb->s_type->i_mutex_key#13/4 rcu_read_lock pool_lock#2 irq_context: 0 rtnl_mutex &idev->mc_lock _xmit_ETHER &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#5 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc21_nci_cmd_wq#10 irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#13 irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#770 &rq->__lock irq_context: 0 (wq_completion)nfc28_nci_cmd_wq#11 irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)nfc28_nci_rx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#15 irq_context: 0 pernet_ops_rwsem fs_reclaim key irq_context: 0 rtnl_mutex team->team_lock_key#118 lock irq_context: 0 pernet_ops_rwsem fs_reclaim pcpu_lock irq_context: 0 rtnl_mutex team->team_lock_key#118 lock kernfs_idr_lock irq_context: 0 &type->s_umount_key#73 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#4 pcpu_lock stock_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#118 &root->kernfs_rwsem irq_context: 0 pernet_ops_rwsem fs_reclaim percpu_counters_lock irq_context: 0 pernet_ops_rwsem fs_reclaim pcpu_lock stock_lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 &child->perf_event_mutex rcu_node_0 irq_context: 0 &child->perf_event_mutex &rcu_state.expedited_wq irq_context: 0 &child->perf_event_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &child->perf_event_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &child->perf_event_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim &cfs_rq->removed.lock irq_context: 0 rtnl_mutex uevent_sock_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 &dev->mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO rcu_read_lock rcu_node_0 irq_context: 0 pernet_ops_rwsem fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)nfc26_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#628 irq_context: 0 pernet_ops_rwsem rtnl_mutex fs_reclaim pool_lock#2 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#628 irq_context: 0 &type->s_umount_key#73 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#73 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock key irq_context: 0 (wq_completion)nfc2_nci_rx_wq#625 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &c->lock irq_context: 0 sk_lock-AF_X25 fs_reclaim irq_context: 0 sk_lock-AF_X25 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sk_lock-AF_X25 stock_lock irq_context: 0 sk_lock-AF_X25 &f->f_lock irq_context: 0 sk_lock-AF_X25 &f->f_lock fasync_lock irq_context: 0 sk_lock-AF_X25 &f->f_lock fasync_lock &new->fa_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#623 irq_context: 0 tty_mutex &tty->legacy_mutex rcu_node_0 irq_context: 0 tty_mutex &tty->legacy_mutex &rcu_state.expedited_wq irq_context: 0 misc_mtx (wq_completion)nfc48_nci_tx_wq#3 irq_context: 0 rtnl_mutex team->team_lock_key#118 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#20 &n->list_lock irq_context: 0 kn->active#20 &n->list_lock &c->lock irq_context: 0 kn->active#20 &rq->__lock irq_context: 0 kn->active#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim rcu_node_0 irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &rcu_state.expedited_wq irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 kn->active#20 &cfs_rq->removed.lock irq_context: 0 kn->active#20 &obj_hash[i].lock irq_context: 0 kn->active#20 pool_lock#2 irq_context: 0 rtnl_mutex &br->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex vlan_ioctl_mutex.wait_lock irq_context: 0 vlan_ioctl_mutex &rq->__lock irq_context: 0 vlan_ioctl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc48_nci_rx_wq#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events kernfs_notify_work &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &cfs_rq->removed.lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu &obj_hash[i].lock irq_context: 0 kn->active#18 &n->list_lock irq_context: 0 kn->active#18 &n->list_lock &c->lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex kfence_freelist_lock irq_context: 0 (wq_completion)events kernfs_notify_work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &fsnotify_mark_srcu pool_lock#2 irq_context: 0 sb_writers#3 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_internal jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_internal jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (crypto_chain).rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 stock_lock irq_context: 0 sk_lock-AF_X25 &f->f_lock fasync_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu quarantine_lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu &c->lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu &n->list_lock irq_context: 0 &mm->mmap_lock sb_writers#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#770 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock rlock-AF_PACKET irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rnp->exp_wq[3] irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override &rq->__lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex.wait_lock irq_context: 0 fs_reclaim &rcu_state.expedited_wq irq_context: 0 fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &base->lock irq_context: 0 &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#770 irq_context: 0 kn->active#18 &rq->__lock irq_context: 0 kn->active#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 fs_reclaim irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 pcpu_alloc_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 shrinker_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 list_lrus_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &xa->xa_lock#15 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 tasklist_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 tasklist_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 stock_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lock#6 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lock#6 kcov_remote_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->lock krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &sta->rate_ctrl_lock krc.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &rdev->bss_lock &obj_hash[i].lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem stock_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &obj_hash[i].lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem key irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pcpu_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem percpu_counters_lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_CAIF irq_context: 0 sk_lock-AF_CAIF slock-AF_CAIF irq_context: 0 sk_lock-AF_CAIF &obj_hash[i].lock irq_context: 0 slock-AF_CAIF irq_context: 0 rlock-AF_CAIF irq_context: 0 sk_lock-AF_CAIF &this->info_list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_CAIF (console_sem).lock irq_context: 0 sk_lock-AF_CAIF &x->wait#2 irq_context: 0 sk_lock-AF_CAIF &rq->__lock irq_context: 0 sk_lock-AF_CAIF &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAIF &ei->socket.wq.wait irq_context: 0 sk_lock-AF_CAIF clock-AF_CAIF irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock kfence_freelist_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &sb->s_type->i_lock_key#36 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex rfkill_global_mutex remove_cache_srcu irq_context: 0 &data->open_mutex rfkill_global_mutex remove_cache_srcu quarantine_lock irq_context: 0 &data->open_mutex rfkill_global_mutex remove_cache_srcu &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &tb->tb6_lock &meta->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &s->s_inode_list_lock irq_context: 0 &q->sysfs_lock &obj_hash[i].lock pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &sb->s_type->i_lock_key#36 &dentry->d_lock irq_context: 0 &data->open_mutex rfkill_global_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &data->open_mutex rfkill_global_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 &x->wait#20 &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rcu_node_0 irq_context: 0 crypto_alg_sem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem rcu_node_0 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &q->sysfs_dir_lock &rcu_state.expedited_wq irq_context: 0 &q->sysfs_dir_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &q->sysfs_dir_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &rq->__lock irq_context: 0 rtnl_mutex gdp_mutex &rq->__lock irq_context: 0 rtnl_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex gdp_mutex gdp_mutex.wait_lock irq_context: 0 nfc_devlist_mutex gdp_mutex &rq->__lock irq_context: 0 nfc_devlist_mutex gdp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock rcu_node_0 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &rcu_state.expedited_wq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem quarantine_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override &c->lock irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override &n->list_lock irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->mq_freeze_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss &obj_hash[i].lock pool_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 crypto_alg_sem &obj_hash[i].lock irq_context: 0 crypto_alg_sem pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#204 irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &simple_offset_xa_lock &obj_hash[i].lock pool_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 irq_context: 0 sb_writers#5 &fsnotify_mark_srcu &rq->__lock irq_context: 0 sb_writers#5 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &root->kernfs_rwsem quarantine_lock irq_context: 0 &hdev->req_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 &xt[i].mutex rcu_read_lock key irq_context: 0 &xt[i].mutex rcu_read_lock pcpu_lock irq_context: 0 &xt[i].mutex rcu_read_lock percpu_counters_lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key/1 &simple_offset_xa_lock &n->list_lock &c->lock irq_context: 0 pernet_ops_rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 rtnl_mutex dev_addr_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 vlan_ioctl_mutex.wait_lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#75 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#75 &x->wait#23 irq_context: 0 &type->s_umount_key#75 shrinker_mutex irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &xt[i].mutex fs_reclaim &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start rcu_node_0 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#75 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#770 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#770 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#765 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#363 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock key irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#75 pool_lock#2 irq_context: 0 &type->s_umount_key#75 rename_lock.seqcount irq_context: 0 &type->s_umount_key#75 &dentry->d_lock irq_context: 0 &type->s_umount_key#75 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#75 &sb->s_type->i_lock_key#36 irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock &tb->tb6_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex nf_conntrack_mutex.wait_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex.wait_lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#75 &s->s_inode_list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 data_sockets.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &base->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu rcu_node_0 irq_context: 0 cb_lock rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#978 &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#978 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 &____s->seqcount#2 irq_context: 0 &net->xdp.lock irq_context: 0 &xs->map_list_lock irq_context: 0 &xs->mutex irq_context: 0 clock-AF_XDP irq_context: 0 &dev->mutex deferred_probe_mutex &rq->__lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock remove_cache_srcu &c->lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock remove_cache_srcu &n->list_lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &pcp->lock &zone->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sk_lock-AF_ISDN irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &____s->seqcount#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#952 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#952 &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex fs_reclaim &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex sb_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 tty_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &meta->lock irq_context: 0 pernet_ops_rwsem __ip_vs_app_mutex &n->list_lock &c->lock irq_context: 0 &data->open_mutex rfkill_global_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex sb_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc18_nci_rx_wq#16 irq_context: 0 &type->s_umount_key#75 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#75 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#75 &dentry->d_lock/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &xa->xa_lock#15 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 lweventlist_lock &dir->lock#2 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[0] irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &____s->seqcount#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock rcu_node_0 irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#968 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#978 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &n->list_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 &n->list_lock &c->lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &pnettable->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex fs_reclaim irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_alloc_mutex &____s->seqcount irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu &c->lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu &n->list_lock irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_RXRPC remove_cache_srcu pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#448 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#364 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#364 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#771 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#773 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#773 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &base->lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#74/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rdev->bss_lock &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) irq_context: 0 tty_mutex stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#44 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#44 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#44 &dentry->d_lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (reaper_work).work &rq->__lock irq_context: 0 (reaper_work).work &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#3 sb_internal jbd2_handle fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &dir->lock &meta->lock irq_context: 0 &dir->lock kfence_freelist_lock irq_context: 0 cb_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex stock_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &p->pi_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &tsk->futex_exit_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex percpu_counters_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (debug_obj_work).work kfence_freelist_lock irq_context: 0 sk_lock-AF_CAN stock_lock irq_context: 0 sk_lock-AF_CAN &f->f_lock irq_context: 0 sk_lock-AF_CAN &f->f_lock fasync_lock irq_context: 0 kn->active#14 remove_cache_srcu &c->lock irq_context: 0 kn->active#14 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#14 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#14 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#14 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &n->list_lock &c->lock irq_context: 0 sk_lock-AF_CAN &f->f_lock fasync_lock &new->fa_lock irq_context: 0 sk_lock-AF_CAN &f->f_lock fasync_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &nbd->config_lock pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock &bdev->bd_size_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &q->queue_lock irq_context: 0 &p->lock &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(sdp, lock) irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_ISDN &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#366 irq_context: 0 cb_lock genl_mutex &nbd->config_lock set->srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &net->ipv6.ip6addrlbl_table.lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#776 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rnp->exp_wq[3] irq_context: 0 cb_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#766 irq_context: 0 sk_lock-AF_ISDN slock-AF_ISDN irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#968 irq_context: 0 sk_lock-AF_ISDN clock-AF_ISDN irq_context: 0 (wq_completion)nfc2_nci_rx_wq#961 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#102 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#366 &rq->__lock irq_context: 0 tty_mutex &rq->__lock irq_context: 0 tty_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#777 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock batched_entropy_u8.lock crngs.lock irq_context: 0 sb_writers#7 kn->active#4 &kernfs_locks->open_file_mutex[count] &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc36_nci_cmd_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#778 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#778 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#779 irq_context: 0 (wq_completion)bond0#112 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 rtnl_mutex team->team_lock_key#118 &____s->seqcount#2 irq_context: 0 &type->s_umount_key#64 &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem rcu_node_0 irq_context: 0 cb_lock genl_mutex &nbd->config_lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock stock_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 tty_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#105 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#205 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#969 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#229 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &meta->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock kfence_freelist_lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#41 irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#701 irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_CAIF &rq->__lock cpu_asid_lock irq_context: 0 cgroup_threadgroup_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->monitor_work)->work) rcu_callback &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#200 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 vmap_purge_lock free_vmap_area_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim rcu_node_0 irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock &base->lock irq_context: 0 (wq_completion)events_long (work_completion)(&br->mcast_gc_work) krc.lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)events key_gc_work &type->lock_class batched_entropy_u8.lock irq_context: 0 (wq_completion)events key_gc_work &type->lock_class kfence_freelist_lock irq_context: 0 (wq_completion)events key_gc_work &type->lock_class remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)events key_gc_work &type->lock_class remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)events key_gc_work &type->lock_class remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events key_gc_work &type->lock_class remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 cgroup_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &x->wait#3 irq_context: 0 cb_lock genl_mutex &nbd->config_lock set->srcu irq_context: 0 pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock &c->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &n->list_lock irq_context: 0 &vma->vm_lock->lock &base->lock irq_context: 0 &vma->vm_lock->lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &obj_hash[i].lock pool_lock irq_context: 0 &fc->uapi_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &vma->vm_lock->lock &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &fc->uapi_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &fc->uapi_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#199 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim &rcu_state.expedited_wq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#498 irq_context: 0 rtnl_mutex dev_addr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#112 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#112 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#112 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#369 irq_context: 0 sb_writers#3 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#3 remove_cache_srcu fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 fs_reclaim &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 smack_known_lock pool_lock#2 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#780 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock key irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#780 &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#955 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#331 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 pcpu_alloc_mutex irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#331 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#335 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 shrinker_mutex irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#112 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) &ndev->lock batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 list_lrus_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &xa->xa_lock#15 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 &hdev->lock &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 stock_lock irq_context: 0 kn->active#46 &n->list_lock irq_context: 0 kn->active#46 &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#118 &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nfc2_nci_tx_wq#766 &rq->__lock irq_context: 0 &fc->uapi_mutex &____s->seqcount#2 irq_context: 0 &fc->uapi_mutex &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &sb->s_type->i_lock_key#29 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &s->s_inode_list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &sb->s_type->i_lock_key#29 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock (kmod_concurrent_max).lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &meta->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu kfence_freelist_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_lock &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#68 sb_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &type->s_umount_key#68 sb_mutex fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#68 sb_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &type->s_umount_key#68 sb_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &xa->xa_lock#15 &obj_hash[i].lock pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock &x->wait#17 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock running_helpers_waitq.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &dentry->d_lock irq_context: 0 &p->lock &of->mutex kn->active#14 rcu_node_0 irq_context: 0 &p->lock &of->mutex kn->active#14 &rq->__lock irq_context: 0 &p->lock &of->mutex kn->active#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#64 irq_context: 0 &type->s_umount_key#64 &x->wait#23 irq_context: 0 &type->s_umount_key#64 shrinker_mutex irq_context: 0 &type->s_umount_key#64 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#64 rename_lock.seqcount irq_context: 0 &type->s_umount_key#64 &dentry->d_lock irq_context: 0 &type->s_umount_key#64 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#64 &sb->s_type->i_lock_key#29 irq_context: 0 &type->s_umount_key#64 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#64 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#64 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#64 &dentry->d_lock/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &xa->xa_lock#15 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock bpf_preload_lock.wait_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock.wait_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &p->pi_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &p->pi_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#748 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#191 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#228 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#444 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#444 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#450 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#450 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#450 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#173 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#173 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#450 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#447 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#169 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#445 irq_context: 0 sb_writers#3 sb_internal quarantine_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 put_task_map-wait-type-override &obj_hash[i].lock irq_context: 0 put_task_map-wait-type-override pool_lock#2 irq_context: 0 put_task_map-wait-type-override stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &xa->xa_lock#15 &n->list_lock irq_context: 0 &p->lock &of->mutex kn->active#4 stock_lock irq_context: 0 &p->lock &of->mutex kn->active#4 key irq_context: 0 &p->lock &of->mutex kn->active#4 pcpu_lock irq_context: 0 &p->lock &of->mutex kn->active#4 percpu_counters_lock irq_context: 0 &p->lock &of->mutex kn->active#4 pcpu_lock stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 &xa->xa_lock#15 &n->list_lock &c->lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC remove_cache_srcu irq_context: 0 cb_lock nlk_cb_mutex-GENERIC remove_cache_srcu quarantine_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock key irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock percpu_counters_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock pcpu_lock stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock &obj_hash[i].lock pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock &cfs_rq->removed.lock irq_context: 0 &tsk->futex_exit_mutex &obj_hash[i].lock pool_lock irq_context: softirq (&p->timer) irq_context: softirq (&p->timer) &br->multicast_lock irq_context: softirq (&p->timer) &br->multicast_lock pool_lock#2 irq_context: softirq (&p->timer) &br->multicast_lock &dir->lock#2 irq_context: softirq (&p->timer) &br->multicast_lock deferred_lock irq_context: 0 &vma->vm_lock->lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: softirq (&p->timer) &br->multicast_lock rcu_read_lock &pool->lock irq_context: softirq (&p->timer) &br->multicast_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: softirq (&p->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: softirq (&p->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: softirq (&p->timer) &br->multicast_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&p->timer) &br->multicast_lock nl_table_lock irq_context: softirq (&p->timer) &br->multicast_lock &obj_hash[i].lock irq_context: softirq (&p->timer) &br->multicast_lock nl_table_wait.lock irq_context: softirq (&p->timer) &br->multicast_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: softirq (&p->timer) &br->multicast_lock &base->lock irq_context: softirq (&p->timer) &br->multicast_lock &base->lock &obj_hash[i].lock irq_context: softirq (&mp->timer) &br->multicast_lock pool_lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock &c->lock irq_context: softirq (&mp->timer) &br->multicast_lock &dir->lock#2 irq_context: softirq (&mp->timer) &br->multicast_lock deferred_lock irq_context: softirq (&mp->timer) &br->multicast_lock nl_table_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &sbi->s_orphan_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&mp->timer) &br->multicast_lock &obj_hash[i].lock irq_context: softirq (&mp->timer) &br->multicast_lock nl_table_wait.lock irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 remove_cache_srcu irq_context: 0 sk_lock-AF_INET6 k-sk_lock-AF_INET6/1 remove_cache_srcu quarantine_lock irq_context: softirq (&mp->timer) &br->multicast_lock &n->list_lock irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &tb->tb6_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#168 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &dev->mutex rfkill_global_mutex percpu_counters_lock irq_context: 0 &dev->mutex rfkill_global_mutex pcpu_lock stock_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&mp->timer) &br->multicast_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu pool_lock#2 irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#451 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#451 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &aux->poke_mutex irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 map_idr_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 map_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 map_idr_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work)#2 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#90 irq_context: 0 &type->s_umount_key#64 &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#64 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#970 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#498 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#498 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events key_gc_work &type->lock_class &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#498 irq_context: 0 &fsnotify_mark_srcu &rcu_state.expedited_wq irq_context: 0 &fsnotify_mark_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#495 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#493 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#69/1 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events key_gc_work rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) pcpu_lock stock_lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#970 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#14 remove_cache_srcu irq_context: 0 &f->f_pos_lock sb_writers#14 remove_cache_srcu quarantine_lock irq_context: 0 &f->f_pos_lock sb_writers#14 remove_cache_srcu &c->lock irq_context: 0 &f->f_pos_lock sb_writers#14 remove_cache_srcu &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#14 remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#205 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &sem->wait_lock irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#205 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bond0#116 irq_context: 0 (wq_completion)bond0#116 (work_completion)(&(&slave->notify_work)->work) irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#5 &base->lock irq_context: 0 sb_writers#5 &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#971 irq_context: 0 &type->s_umount_key#64 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&(&ssp->srcu_sup->work)->work) &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 pcpu_alloc_mutex &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#971 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#88 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu pcpu_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu percpu_counters_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#544 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx quarantine_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh &meta->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh kfence_freelist_lock irq_context: 0 &kcov->lock kcov_remote_lock &____s->seqcount#2 irq_context: 0 &kcov->lock kcov_remote_lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#544 irq_context: 0 rtnl_mutex team->team_lock_key#116 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#116 cbs_list_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem rcu_node_0 irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) cpu_hotplug_lock jump_label_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)afs (work_completion)(&net->fs_manager) pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#39/1 bpf_preload_lock &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#32 &c->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#748 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#176 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#176 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#541 irq_context: 0 (wq_completion)events key_gc_work &type->lock_class &____s->seqcount#2 irq_context: 0 (wq_completion)events key_gc_work &type->lock_class &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#118 &n->list_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex &rq->__lock cpu_asid_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pcpu_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem percpu_counters_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &p->pi_lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 &ret->b_state_lock &journal->j_list_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock fs_reclaim irq_context: 0 cb_lock genl_mutex &nbd->config_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#118 &n->list_lock &c->lock irq_context: 0 (wq_completion)bond0#116 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#116 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#203 irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &simple_offset_xa_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#137 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#204 irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex nl_table_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#206 irq_context: 0 rtnl_mutex subsys mutex#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex subsys mutex#20 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#969 irq_context: 0 slock-AF_ISDN irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#176 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#962 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#766 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#775 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#775 &rq->__lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex dev_addr_sem &br->lock &br->hash_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)rcu_gp &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#972 irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &rq->__lock &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &xa->xa_lock#4 &pcp->lock &zone->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &xa->xa_lock#4 &____s->seqcount irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 tty_mutex &tty->legacy_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tty_mutex &tty->legacy_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#775 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex key irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex pcpu_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex percpu_counters_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex pcpu_lock stock_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex pool_lock#2 irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex &cfs_rq->removed.lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cgroup_threadgroup_rwsem freezer_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#4 &dentry->d_lock &wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &idev->mc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#775 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#775 irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &lock->wait_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &rq->__lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &xa->xa_lock#4 pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#770 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#767 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#767 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#767 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &base->lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#771 irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex purge_vmap_area_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &pipe->mutex/1 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#768 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#772 irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#10 irq_context: 0 s_shared_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex &lock->wait_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock &lock->wait_lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex &nbd->config_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex purge_vmap_area_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events pcpu_balance_work pcpu_alloc_mutex purge_vmap_area_lock pool_lock#2 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#769 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#773 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#10 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &____s->seqcount irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#568 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#568 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events key_gc_work quarantine_lock irq_context: 0 cb_lock rcu_read_lock stock_lock irq_context: 0 cb_lock rcu_read_lock key irq_context: 0 cb_lock rcu_read_lock pcpu_lock irq_context: 0 cb_lock rcu_read_lock percpu_counters_lock irq_context: 0 rtnl_mutex bus_type_sem &rq->__lock irq_context: 0 rtnl_mutex bus_type_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 &sb->s_type->i_mutex_key#10 stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &vma->vm_lock->lock pool_lock#2 irq_context: 0 &q->sysfs_lock mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &q->sysfs_dir_lock remove_cache_srcu irq_context: 0 &q->sysfs_dir_lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 nf_conntrack_mutex &obj_hash[i].lock pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock quarantine_lock irq_context: 0 sb_internal remove_cache_srcu irq_context: 0 &disk->open_mutex &p->pi_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu irq_context: 0 &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#33 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#33 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#33 &dentry->d_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock kfence_freelist_lock irq_context: 0 sb_internal remove_cache_srcu quarantine_lock irq_context: 0 sb_internal remove_cache_srcu &c->lock irq_context: 0 &q->queue_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sched_map-wait-type-override pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 pcpu_alloc_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 shrinker_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 list_lrus_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 sb_lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &xa->xa_lock#15 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_lock_key#23 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &s->s_inode_list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 sk_lock-AF_INET fill_pool_map-wait-type-override &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers fill_pool_map-wait-type-override pool_lock irq_context: 0 &fsnotify_mark_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx rfkill_global_mutex &____s->seqcount#2 irq_context: 0 pernet_ops_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock lock#4 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &xa->xa_lock#15 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &n->list_lock irq_context: 0 &fsnotify_mark_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &n->list_lock &c->lock irq_context: 0 &disk->open_mutex &p->pi_lock &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_internal remove_cache_srcu &n->list_lock irq_context: 0 &disk->open_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &____s->seqcount#2 irq_context: 0 sb_internal remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_internal remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_internal remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_tx_wq#770 irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock &meta->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu key irq_context: 0 sk_lock-AF_INET &tcp_hashinfo.bhash[i].lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu pcpu_lock stock_lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem ipvs->est_mutex fs_reclaim &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc26_nci_rx_wq#11 irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 rtnl_mutex &wq->mutex &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sighand->siglock quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &lock->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex (console_sem).lock irq_context: 0 &fc->uapi_mutex console_lock console_srcu console_owner_lock irq_context: 0 &fc->uapi_mutex console_lock console_srcu console_owner irq_context: 0 &fc->uapi_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 &fc->uapi_mutex console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)wg-kex-wg2#186 irq_context: 0 sk_lock-AF_INET6 fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex smc_ib_devices.mutex &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &zone->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &n->list_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &zone->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] &n->list_lock irq_context: 0 kn->active#16 &kernfs_locks->open_file_mutex[count] &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 pcpu_alloc_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 shrinker_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 list_lrus_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 sb_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &xa->xa_lock#15 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &obj_hash[i].lock pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &sb->s_type->i_lock_key#34 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &s->s_inode_list_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock stock_lock irq_context: 0 vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 vmap_purge_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock key irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock pcpu_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock percpu_counters_lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock pcpu_lock stock_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle key irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle pcpu_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle percpu_counters_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 &dev->mutex kn->active#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &disk->open_mutex nbd_index_mutex &nbd->config_lock &cfs_rq->removed.lock irq_context: 0 &disk->open_mutex &nbd->config_lock stock_lock irq_context: 0 &disk->open_mutex &nbd->config_lock key irq_context: 0 &disk->open_mutex &nbd->config_lock pcpu_lock irq_context: 0 &disk->open_mutex &nbd->config_lock percpu_counters_lock irq_context: 0 &disk->open_mutex &nbd->config_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 rcu_read_lock rcu_read_lock_bh &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#367 irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#367 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#359 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#357 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)kblockd (work_completion)(&q->timeout_work) &tags->lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 irq_context: 0 &mm->mmap_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock rcu_read_lock &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 sb_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#779 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#596 irq_context: 0 &dev->mutex dpm_list_mtx rcu_node_0 irq_context: 0 &fc->uapi_mutex v9fs_trans_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 fill_pool_map-wait-type-override &c->lock irq_context: 0 misc_mtx hrtimer_bases.lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 tasklist_lock &sighand->siglock &sighand->signalfd_wqh &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &xa->xa_lock#15 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &s->s_inode_list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &sb->s_type->i_mutex_key#9 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem quarantine_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 &lru->node[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &sb->s_type->i_lock_key#34 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#71 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock &br->hash_lock quarantine_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock &____s->seqcount irq_context: 0 &vma->vm_lock->lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#14 &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#702 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#323 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#323 &rq->__lock irq_context: 0 (wq_completion)events drain_vmap_work vmap_purge_lock free_vmap_area_lock init_mm.page_table_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#703 irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 fs_reclaim &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#323 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#6 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: softirq (&peer->timer_persistent_keepalive) batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 pcpu_alloc_mutex irq_context: 0 (wq_completion)events (work_completion)(&(&krcp->krw_arr[i].rcu_work)->work) rcu_callback rcu_node_0 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#18 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 shrinker_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &pcp->lock &zone->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 list_lrus_mutex irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 tty_mutex &tty->legacy_mutex console_lock lock kernfs_idr_lock &c->lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &sem->wait_lock irq_context: 0 tty_mutex &tty->legacy_mutex console_lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 sb_lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock rcu_read_lock &rq->__lock irq_context: 0 cb_lock genl_mutex &nbd->config_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex bpf_devs_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_node_0 irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 pcpu_alloc_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &type->s_umount_key#71 &x->wait#23 irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock &sem->wait_lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_RXRPC quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex netpoll_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &p->pi_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_X25 &sem->wait_lock irq_context: 0 sk_lock-AF_X25 &p->pi_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 shrinker_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 list_lrus_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 sb_lock irq_context: 0 rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rcu_state.exp_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rcu_state.exp_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rnp->exp_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &rnp->exp_wq[3] irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_state.exp_mutex.wait_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &xa->xa_lock#15 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &xa->xa_lock#15 &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &base->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &____s->seqcount#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &pcp->lock &zone->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &____s->seqcount irq_context: 0 rtnl_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 &type->s_umount_key#71 shrinker_mutex irq_context: 0 &type->s_umount_key#71 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &pcp->lock &zone->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock _xmit_ETHER &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_X25 &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_X25 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_lock_key#37 irq_context: 0 &type->s_umount_key#30 rcu_read_lock &obj_hash[i].lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &s->s_inode_list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 cb_lock genl_mutex rfkill_global_mutex triggers_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 tk_core.seq.seqcount irq_context: 0 &xt[i].mutex sched_map-wait-type-override &rq->__lock irq_context: 0 &xt[i].mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_lock_key#37 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &type->s_umount_key#71 rename_lock.seqcount irq_context: 0 &type->s_umount_key#71 &dentry->d_lock irq_context: 0 &type->s_umount_key#71 rcu_read_lock &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx lock kernfs_idr_lock &____s->seqcount irq_context: softirq rcu_read_lock &rq->__lock &obj_hash[i].lock irq_context: softirq rcu_read_lock &rq->__lock &base->lock irq_context: softirq rcu_read_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &xa->xa_lock#15 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &xa->xa_lock#15 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock pool_lock#2 irq_context: 0 &type->s_umount_key#71 &sb->s_type->i_lock_key#34 irq_context: 0 &type->s_umount_key#71 &s->s_inode_list_lock irq_context: 0 cb_lock &rcu_state.expedited_wq irq_context: 0 cb_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 cb_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 kn->active#49 &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem stock_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem key irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pcpu_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem percpu_counters_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem pcpu_lock stock_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &n->list_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem &n->list_lock &c->lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem.waiters.lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#71 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#71 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#71 &dentry->d_lock/1 irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &rq->__lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem cpuset_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 rename_lock.seqcount irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock &meta->lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock batched_entropy_u8.lock irq_context: softirq (&hsr->announce_timer) rcu_read_lock &hsr->seqnr_lock rcu_read_lock kfence_freelist_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock stock_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock key irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock pcpu_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem stock_lock irq_context: 0 &fsnotify_mark_srcu &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem key irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &dev->mutex batched_entropy_u8.lock irq_context: 0 &dev->mutex kfence_freelist_lock irq_context: 0 &dev->mutex &meta->lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 stock_lock irq_context: 0 &p->lock &of->mutex kn->active#4 &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 pool_lock#2 irq_context: 0 sk_lock-AF_INET &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_index_ida.xa_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock percpu_counters_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_lock_key#35 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 rcu_read_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET &msk->pm.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &sb->s_type->i_lock_key#37 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &n->list_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx lweventlist_lock &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &s->s_inode_list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 rcu_read_lock iunique_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &sb->s_type->i_lock_key#37 &dentry->d_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &c->lock irq_context: 0 (wq_completion)nfc36_nci_cmd_wq#5 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &n->list_lock &c->lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 rename_lock.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 rcu_read_lock &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 rename_lock.seqcount irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex rcu_node_0 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rcu_state.expedited_wq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex deferred_probe_mutex &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex deferred_probe_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_PPPOX irq_context: 0 sk_lock-AF_PPPOX slock-AF_PPPOX irq_context: 0 sk_lock-AF_PPPOX &pn->hash_lock irq_context: 0 sk_lock-AF_PPPOX clock-AF_PPPOX irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 pool_lock#2 irq_context: 0 slock-AF_PPPOX irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 mmu_notifier_invalidate_range_start irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#123 irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->dat.work)->work) &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_PPPOX pool_lock irq_context: 0 (wq_completion)events key_gc_work rcu_node_0 irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 pcpu_alloc_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 pcpu_alloc_mutex &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 shrinker_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 list_lrus_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 percpu_counters_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 crngs.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &sbinfo->stat_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &xa->xa_lock#15 irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &sb->s_type->i_lock_key irq_context: 0 &dev->mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &s->s_inode_list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 batched_entropy_u32.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &sb->s_type->i_lock_key &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#47 irq_context: 0 &type->s_umount_key#47 &x->wait#23 irq_context: 0 &type->s_umount_key#47 shrinker_mutex irq_context: 0 &type->s_umount_key#47 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#47 rename_lock.seqcount irq_context: 0 &type->s_umount_key#47 &dentry->d_lock irq_context: 0 &type->s_umount_key#47 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#47 &sb->s_type->i_lock_key irq_context: 0 &type->s_umount_key#47 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#47 &sbinfo->stat_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rcu_read_lock rcu_node_0 irq_context: 0 &type->s_umount_key#47 &xa->xa_lock#8 irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->s_umount_key#47 pool_lock#2 irq_context: 0 &type->s_umount_key#47 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#47 &simple_offset_xa_lock irq_context: 0 cb_lock console_lock console_srcu console_owner &port_lock_key irq_context: 0 cb_lock console_lock console_srcu console_owner console_owner_lock irq_context: 0 &type->s_umount_key#47 &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#47 &rq->__lock irq_context: 0 &type->s_umount_key#47 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#47 &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#47 percpu_counters_lock irq_context: 0 &type->s_umount_key#47 pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key/1 &xa->xa_lock#15 &c->lock irq_context: 0 sb_writers#3 &mm->mmap_lock lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 key irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &n->list_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &n->list_lock &c->lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#8 &____s->seqcount irq_context: 0 sb_writers#3 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &____s->seqcount#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#3 &mm->mmap_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &mm->mmap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &mm->mmap_lock &xa->xa_lock#8 &c->lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &mm->mmap_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 (work_completion)(&msk->work) irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_BLUETOOTH-BTPROTO_SCO irq_context: 0 &sb->s_type->i_mutex_key#10 clock-AF_BLUETOOTH irq_context: 0 &sb->s_type->i_mutex_key#10 sco_sk_list.lock irq_context: 0 &sb->s_type->i_mutex_key#9 key irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 rcu_read_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &sb->s_type->i_mutex_key#9 pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#9 percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#9 pcpu_lock stock_lock irq_context: 0 sb_writers#3 &mm->mmap_lock pool_lock#2 irq_context: 0 sb_writers#3 &mm->mmap_lock ptlock_ptr(ptdesc)#2 irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &meta->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 pcpu_alloc_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 shrinker_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 list_lrus_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &xa->xa_lock#15 irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &sb->s_type->i_lock_key#16 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &s->s_inode_list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &sb->s_type->i_lock_key#16 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#18/1 &dentry->d_lock irq_context: 0 sb_writers#3 &sb->s_type->i_mutex_key#8 &ei->i_data_sem &ei->i_es_lock key#8 irq_context: 0 &sbi->s_writepages_rwsem &____s->seqcount#2 irq_context: 0 &sbi->s_writepages_rwsem &c->lock irq_context: 0 &sbi->s_writepages_rwsem &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET stock_lock irq_context: 0 sk_lock-AF_INET key irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET pcpu_lock irq_context: 0 sk_lock-AF_INET percpu_counters_lock irq_context: 0 sk_lock-AF_INET pcpu_lock stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 &sb->s_type->i_lock_key#37 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 &s->s_inode_list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&aux->work) per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sbi->s_md_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)inet_frag_wq (work_completion)(&fqdir->destroy_work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &ei->i_prealloc_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &mapping->i_private_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem key#3 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &pa->pa_lock#2 irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 lock#4 &lruvec->lru_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fib_info_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fib_info_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fib_info_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &(ei->i_block_reservation_lock) key#15 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#65 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_SCO &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &xa->xa_lock#8 key#11 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock key#11 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &base->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &base->lock &obj_hash[i].lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 rcu_read_lock iunique_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &retval->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock tk_core.seq.seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &nvmeq->sq_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &____s->seqcount#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 &sb->s_type->i_lock_key#37 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#783 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &base->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sn->rpc_client_lock irq_context: 0 &u->iolock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem rename_lock.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem rcu_read_lock &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 irq_context: 0 ebt_mutex &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 ebt_mutex &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 ebt_mutex &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 rename_lock.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &pcp->lock &zone->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &sb->s_type->i_lock_key#37 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &s->s_inode_list_lock irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 rcu_read_lock iunique_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &sb->s_type->i_lock_key#37 &dentry->d_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx remove_cache_srcu irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &type->i_mutex_dir_key#4 &root->kernfs_rwsem pool_lock#2 irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx remove_cache_srcu pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#77 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_node_0 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 pcpu_alloc_mutex &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#27/1 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#77 &x->wait#23 irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 &type->s_umount_key#71 &rq->__lock irq_context: 0 &type->s_umount_key#71 &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#71 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#77 shrinker_mutex irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_state.exp_mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &type->s_umount_key#77 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#779 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 sb_internal jbd2_handle &ei->i_data_sem &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc19_nci_tx_wq#10 irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &sb->s_type->i_mutex_key#3 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) &ndev->lock fill_pool_map-wait-type-override pool_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 &f->f_pos_lock sb_writers#14 &mm->mmap_lock rcu_read_lock ptlock_ptr(ptdesc)#2 irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&port->bc_work) rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &base->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#779 irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#703 irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#774 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#771 irq_context: 0 &data->open_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#368 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#368 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#138 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#138 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#360 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock irq_context: 0 rtnl_mutex team->team_lock_key#118 lweventlist_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#700 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#201 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#200 irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 cb_lock stock_lock irq_context: 0 cb_lock key irq_context: 0 cb_lock pcpu_lock irq_context: 0 cb_lock percpu_counters_lock irq_context: 0 cb_lock pcpu_lock stock_lock irq_context: 0 sb_writers#8 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#700 &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#358 irq_context: 0 (wq_completion)nfc26_nci_tx_wq#16 irq_context: 0 sk_lock-AF_INET rcu_read_lock rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 pcpu_alloc_mutex &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock &base->lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex udc_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock pool_lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex udc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex rcu_read_lock rcu_node_0 irq_context: 0 &mm->mmap_lock &anon_vma->rwsem rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#9 batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#9 kfence_freelist_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex &rnp->exp_wq[0] irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 sb_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sn->rpc_client_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 &mm->mmap_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem key irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem pcpu_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem percpu_counters_lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) &bat_priv->mcast.mla_lock kfence_freelist_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &sb->s_type->i_mutex_key#20 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rcu_state.expedited_wq irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &sb->s_type->i_mutex_key#20 &pipe->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &sb->s_type->i_mutex_key#20 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &sb->s_type->i_mutex_key#20 (work_completion)(&(&pipe->queue_timeout)->work) irq_context: 0 &disk->open_mutex &nbd->config_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem sysctl_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &disk->open_mutex &rq->__lock cpu_asid_lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &fsnotify_mark_srcu irq_context: 0 (wq_completion)writeback &rq->__lock irq_context: 0 sk_lock-AF_CAN &pcp->lock &zone->lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &sb->s_type->i_lock_key#37 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &xa->xa_lock#8 irq_context: 0 pernet_ops_rwsem rtnl_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex j1939_netdev_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (reaper_work).work &ACCESS_PRIVATE(ssp->srcu_sup, lock) rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex j1939_netdev_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle rcu_read_lock &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#358 &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_raw_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock pool_lock#2 irq_context: 0 &dev->mutex deferred_probe_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &retval->lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &c->lock irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock &base->lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#358 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#136 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#134 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#780 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#30 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &n->list_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock rcu_node_0 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock key#7 irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock kfence_freelist_lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &dentry->d_lock pool_lock#2 irq_context: 0 &type->s_umount_key#77 rename_lock.seqcount irq_context: 0 &type->s_umount_key#77 &rq->__lock irq_context: 0 &type->s_umount_key#77 &dentry->d_lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->s_umount_key#77 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#77 &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#77 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&cache_cleaner)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#780 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#775 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#772 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#369 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#771 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#766 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#763 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#772 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#772 irq_context: 0 &type->s_umount_key#77 &sb->s_type->i_lock_key#37 irq_context: 0 &type->s_umount_key#77 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#77 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#77 pool_lock#2 irq_context: 0 &type->s_umount_key#77 &fsnotify_mark_srcu irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#77 &dentry->d_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#77 &dentry->d_lock pool_lock#2 irq_context: 0 &type->s_umount_key#77 &obj_hash[i].lock pool_lock irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex rcu_node_0 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 rcu_read_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &____s->seqcount#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&s->destroy_work) kfence_freelist_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &n->list_lock &c->lock irq_context: 0 rcu_read_lock &f->f_owner.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20 &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sb->s_type->i_mutex_key#20/1 &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#77 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#77 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem rcu_read_lock rcu_node_0 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem rcu_read_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock (work_completion)(flush) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &x->wait#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_highpri (work_completion)(flush) irq_context: 0 &dev->mutex kernfs_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_highpri (work_completion)(flush) &list->lock#12 irq_context: 0 (wq_completion)events_highpri (work_completion)(&barr->work) irq_context: 0 (wq_completion)events_highpri (work_completion)(&barr->work) &x->wait#10 irq_context: 0 (wq_completion)events_highpri (work_completion)(&barr->work) &x->wait#10 &p->pi_lock irq_context: 0 (wq_completion)events_highpri (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_highpri (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#76/1 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &type->s_umount_key#30 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &type->s_umount_key#30 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#30 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &n->list_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#767 irq_context: 0 (wq_completion)hci0#5 irq_context: 0 &sig->cred_guard_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex rcu_read_lock pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#70/1 &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 jbd2_handle &____s->seqcount#2 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex cgroup_mutex irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &fc->uapi_mutex cgroup_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 tomoyo_ss rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &simple_offset_xa_lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#337 irq_context: 0 rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#3 &dentry->d_lock pool_lock#2 irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_NETLINK rcu_read_lock pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle lock#4 &lruvec->lru_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &____s->seqcount irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock quarantine_lock irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#8 irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#764 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#356 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#354 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#773 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#773 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#176 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#976 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &macsec_netdev_addr_lock_key/1 &obj_hash[i].lock pool_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#177 irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy230 irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 &group->mark_mutex fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &group->mark_mutex fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &group->mark_mutex fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_owner_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx console_owner irq_context: 0 (wq_completion)nfc2_nci_rx_wq#768 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#765 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#365 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#365 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#357 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#355 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#355 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &s->s_inode_list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_lock_key#35 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &dentry->d_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock stock_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock key irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#355 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#337 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &mm->mmap_lock &sem->wait_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &mm->mmap_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &mm->mmap_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 init_user_ns.keyring_sem init_user_ns.keyring_sem.wait_lock irq_context: 0 init_user_ns.keyring_sem.wait_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem batched_entropy_u8.lock crngs.lock irq_context: 0 cgroup_threadgroup_rwsem pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN &priv->lock quarantine_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &ret->b_state_lock bit_wait_table + i &p->pi_lock irq_context: 0 &q->debugfs_mutex &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 &xt[i].mutex fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_INET &mm->mmap_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET &mm->mmap_lock pool_lock#2 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#366 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex triggers_list_lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#366 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#366 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#366 irq_context: 0 &tty->legacy_mutex/1 irq_context: 0 &tty->legacy_mutex/1 &tty->ldisc_sem irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock genl_mutex &pernet->lock irq_context: 0 &tty->legacy_mutex/1 &tty->files_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->legacy_mutex/1 tasklist_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#358 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock quarantine_lock irq_context: 0 &tty->legacy_mutex/1 tasklist_lock &sighand->siglock irq_context: 0 &tty->legacy_mutex/1 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 stock_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->legacy_mutex/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#358 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#356 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#774 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#774 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#774 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#774 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#774 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#774 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &meta->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock kfence_freelist_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock stock_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &sb->s_type->i_mutex_key#3 &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#769 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#369 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#369 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#361 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#359 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#139 irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 purge_vmap_area_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#561 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#244 irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->files_lock &f->f_lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->files_lock &f->f_lock fasync_lock irq_context: 0 &tty->legacy_mutex/1 &f->f_lock irq_context: 0 &tty->legacy_mutex/1 &f->f_lock fasync_lock irq_context: 0 &tty->legacy_mutex/1 &tty->read_wait irq_context: 0 &tty->legacy_mutex/1 &tty->write_wait irq_context: 0 &tty->legacy_mutex/1 &tty->ctrl.lock irq_context: 0 &tty->legacy_mutex/1 &obj_hash[i].lock irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem &pcp->lock &zone->lock irq_context: 0 &tty->ldisc_sem &tty->ldisc_sem/1 &tty->termios_rwsem &____s->seqcount irq_context: 0 (wq_completion)events free_ipc_work &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 purge_vmap_area_lock &____s->seqcount irq_context: 0 &hdev->req_lock &hdev->lock &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#139 &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#137 irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (crypto_chain).rwsem &____s->seqcount#2 irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#139 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#139 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#137 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#135 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#781 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#781 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#776 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#773 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#49 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#49 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#49 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#47 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#46 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#782 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#8 stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#782 &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#8 pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#782 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#782 irq_context: 0 sk_lock-AF_INET6 device_spinlock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#777 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#774 irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &xa->xa_lock#8 &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem lock#4 irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#778 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#775 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#775 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#775 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#784 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#784 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#779 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#779 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#779 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#776 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &nf_conntrack_locks[i] batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#776 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#370 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#370 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#362 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#362 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#362 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#360 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &meta->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx kfence_freelist_lock irq_context: 0 key_types_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 key_types_sem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &meta->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &obj_hash[i].lock irq_context: 0 &hdev->lock &____s->seqcount#2 irq_context: 0 &hdev->lock &pcp->lock &zone->lock irq_context: 0 &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_tx_wq#360 &rq->__lock irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 &data->open_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex subsys mutex#20 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#785 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#785 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#780 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#780 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#780 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->legacy_mutex &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#777 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#980 &cfs_rq->removed.lock irq_context: 0 sb_writers#9 remove_cache_srcu irq_context: 0 sb_writers#9 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#9 remove_cache_srcu &c->lock irq_context: 0 sb_writers#9 remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_CAN &list->lock#21 irq_context: 0 sk_lock-AF_CAN &priv->active_session_list_lock irq_context: 0 sk_lock-AF_CAN hrtimer_bases.lock irq_context: 0 sk_lock-AF_CAN hrtimer_bases.lock tk_core.seq.seqcount irq_context: 0 sk_lock-AF_CAN hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock &priv->lock irq_context: softirq rcu_read_lock rcu_read_lock &priv->active_session_list_lock irq_context: softirq rcu_read_lock rcu_read_lock hrtimer_bases.lock irq_context: softirq rcu_read_lock rcu_read_lock hrtimer_bases.lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock hrtimer_bases.lock tk_core.seq.seqcount irq_context: softirq rcu_read_lock rcu_read_lock &priv->j1939_socks_lock irq_context: softirq rcu_read_lock rcu_read_lock &priv->j1939_socks_lock &jsk->filters_lock irq_context: softirq rcu_read_lock rcu_read_lock &list->lock#21 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &tty->legacy_mutex &____s->seqcount irq_context: softirq &priv->j1939_socks_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sk_lock-AF_ALG &dir->lock irq_context: 0 rtnl_mutex subsys mutex#20 &rq->__lock irq_context: 0 sk_lock-AF_INET6/1 irq_context: 0 sk_lock-AF_INET6/1 &rq->__lock irq_context: 0 sk_lock-AF_INET6/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6/1 slock-AF_INET6 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &base->lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &base->lock &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET6/1 rlock-AF_INET6 irq_context: 0 sk_lock-AF_INET6/1 &list->lock#19 irq_context: 0 &net->sctp.addr_wq_lock irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET6/1 irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET6/1 &sctp_ep_hashtable[i].lock irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET6/1 &obj_hash[i].lock irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET6/1 pool_lock#2 irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET6/1 &obj_hash[i].lock pool_lock irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET6/1 clock-AF_INET6 irq_context: 0 &net->sctp.addr_wq_lock slock-AF_INET6/1 &____s->seqcount irq_context: 0 &hdev->req_lock &____s->seqcount#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &____s->seqcount#2 irq_context: softirq &priv->j1939_socks_lock &jsk->filters_lock irq_context: softirq &list->lock#21 irq_context: softirq rcu_read_lock rcu_read_lock &priv->j1939_socks_lock pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex sb_writers#3 &rq->__lock &cfs_rq->removed.lock irq_context: softirq rcu_read_lock rcu_read_lock &priv->j1939_socks_lock rlock-AF_CAN irq_context: softirq rcu_read_lock rcu_read_lock &jsk->sk_session_queue_lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu irq_context: 0 sk_lock-AF_ALG remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start rcu_node_0 irq_context: 0 sk_lock-AF_ALG remove_cache_srcu &c->lock irq_context: 0 sk_lock-AF_CAN &jsk->waitq irq_context: softirq rcu_callback rlock-AF_CAN irq_context: softirq rcu_read_lock_bh _xmit_NONE#2 irq_context: softirq rcu_read_lock_bh _xmit_NONE#2 rcu_read_lock pool_lock#2 irq_context: softirq rcu_read_lock_bh _xmit_NONE#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) &ndev->lock &base->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &base->lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &base->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: softirq (&ndev->rs_timer) &ndev->lock &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_read_lock_bh _xmit_NONE#2 rcu_read_lock tk_core.seq.seqcount irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_owner_lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_NETROM console_owner irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sk_lock-AF_PPPOX &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu &n->list_lock irq_context: 0 sk_lock-AF_ALG remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu &c->lock irq_context: softirq rcu_read_lock_bh _xmit_NONE#2 rcu_read_lock &list->lock#12 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock &n->list_lock irq_context: softirq rcu_read_lock_bh _xmit_NONE#2 rcu_read_lock &c->lock irq_context: softirq rcu_read_lock rcu_read_lock &jsk->waitq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: softirq rcu_read_lock_bh _xmit_NONE#2 rcu_read_lock &____s->seqcount#2 irq_context: softirq rcu_read_lock_bh _xmit_NONE#2 rcu_read_lock &____s->seqcount irq_context: softirq rcu_read_lock rcu_read_lock &priv->j1939_socks_lock &c->lock irq_context: 0 &type->s_umount_key#66 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq rcu_read_lock rcu_read_lock &priv->j1939_socks_lock &____s->seqcount#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem &n->list_lock &c->lock irq_context: 0 key_types_sem init_user_ns.keyring_sem rcu_read_lock &rq->__lock irq_context: 0 key_types_sem init_user_ns.keyring_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &vma->vm_lock->lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem &ipvlan->addrs_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem &rq->__lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &rq->__lock irq_context: softirq rcu_read_lock rcu_read_lock &priv->j1939_socks_lock &____s->seqcount irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) pool_lock#2 irq_context: softirq rcu_read_lock_bh _xmit_NONE#2 rcu_read_lock &n->list_lock irq_context: softirq rcu_read_lock_bh _xmit_NONE#2 rcu_read_lock &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#786 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#786 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 &dentry->d_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#967 irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock key irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock pcpu_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->lock quarantine_lock irq_context: 0 (wq_completion)events (work_completion)(&ht->run_work) &ht->mutex remove_cache_srcu &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#781 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock stock_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock rcu_read_lock pcpu_lock stock_lock irq_context: 0 cb_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#778 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#371 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &batadv_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &batadv_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &batadv_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_hook_mutex kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->lock#2 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &____s->seqcount#2 irq_context: 0 kn->active#48 &kernfs_locks->open_file_mutex[count] &____s->seqcount irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cgroup_mutex.wait_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex stock_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex key irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex pcpu_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex percpu_counters_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex pcpu_lock stock_lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex pool_lock#2 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex &cfs_rq->removed.lock irq_context: 0 nf_nat_proto_mutex nf_nat_proto_mutex.wait_lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 map_idr_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#985 irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#985 irq_context: 0 nf_nat_proto_mutex.wait_lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 krc.lock &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &batadv_netdev_addr_lock_key/1 krc.lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rlock-AF_PACKET irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key/1 tomoyo_ss quarantine_lock irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 &f->f_pos_lock sb_writers#4 &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 &f->f_pos_lock sb_writers#4 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#5 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &lock->wait_lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx sched_map-wait-type-override &pool->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &lock->wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#93 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->timeout_work)->work) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#2 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx fill_pool_map-wait-type-override &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex stock_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock irq_context: 0 &sig->cred_guard_mutex key irq_context: 0 &sig->cred_guard_mutex pcpu_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &local->handle_wake_tx_queue_lock hwsim_radio_lock init_task.mems_allowed_seq.seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rlock-AF_CAN irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 &rnp->exp_wq[1] irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock &tbl->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &lock->wait_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &p->pi_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock kfence_freelist_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &c->lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &meta->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1009 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1009 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1009 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy213 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#461 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#461 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 &n->list_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#978 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &simple_offset_xa_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#7 remove_cache_srcu stock_lock irq_context: 0 sb_writers#7 remove_cache_srcu key irq_context: 0 sb_writers#7 remove_cache_srcu pcpu_lock irq_context: 0 sb_writers#7 remove_cache_srcu percpu_counters_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &pcp->lock &zone->lock irq_context: 0 sb_writers#7 remove_cache_srcu pcpu_lock stock_lock irq_context: 0 sb_writers#7 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#210 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock key irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock pcpu_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock percpu_counters_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock pcpu_lock stock_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &____s->seqcount irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)events fqdir_free_work rcu_state.barrier_mutex pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#4 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#4 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem net_rwsem pool_lock#2 irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) quarantine_lock irq_context: 0 sk_lock-AF_NFC batched_entropy_u8.lock irq_context: 0 sk_lock-AF_CAN slock-AF_CAN &sk->sk_lock.wq irq_context: 0 slock-AF_CAN &sk->sk_lock.wq irq_context: 0 slock-AF_CAN &sk->sk_lock.wq &p->pi_lock irq_context: 0 slock-AF_CAN &sk->sk_lock.wq &p->pi_lock &rq->__lock irq_context: 0 slock-AF_CAN &sk->sk_lock.wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal batched_entropy_u8.lock crngs.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &base->lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 rcu_read_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss key irq_context: 0 &sig->cred_guard_mutex tomoyo_ss pcpu_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#244 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#240 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#239 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#567 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#451 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy222 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &meta->lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#463 irq_context: 0 sk_lock-AF_NFC kfence_freelist_lock irq_context: softirq (&app->join_timer) batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#567 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#567 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#567 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#565 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#563 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#93 irq_context: 0 rtnl_mutex smc_ib_devices.mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 &sb->s_type->i_lock_key#35 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 &s->s_inode_list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 rcu_read_lock iunique_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 tk_core.seq.seqcount irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &tty->termios_rwsem &rq->__lock irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &tty->termios_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 &sb->s_type->i_lock_key#35 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &xa->xa_lock#8 &c->lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &xa->xa_lock#8 &n->list_lock irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &xa->xa_lock#8 &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex stock_lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex &net->ipv6.fib6_gc_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &fc->uapi_mutex mnt_id_ida.xa_lock irq_context: 0 &fc->uapi_mutex pcpu_alloc_mutex irq_context: 0 &fc->uapi_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 &fc->uapi_mutex &dentry->d_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 &fc->uapi_mutex mount_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex mount_lock mount_lock.seqcount irq_context: 0 &fc->uapi_mutex ucounts_lock irq_context: 0 &fc->uapi_mutex &newf->file_lock irq_context: 0 &fc->uapi_mutex rcu_read_lock &dentry->d_lock irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#73 irq_context: 0 &type->s_umount_key#73 &x->wait#23 irq_context: 0 &type->s_umount_key#73 shrinker_mutex irq_context: 0 &type->s_umount_key#73 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#73 nfsd_mutex irq_context: 0 &type->s_umount_key#73 rename_lock.seqcount irq_context: 0 &type->s_umount_key#73 &dentry->d_lock irq_context: 0 &type->s_umount_key#73 &dentry->d_lock &dentry->d_lock/1 irq_context: 0 &type->s_umount_key#73 &sb->s_type->i_lock_key#35 irq_context: 0 &type->s_umount_key#73 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#73 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#73 pool_lock#2 irq_context: 0 kn->active#16 &____s->seqcount#2 irq_context: 0 kn->active#16 &____s->seqcount irq_context: 0 &type->s_umount_key#73 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#73 &dentry->d_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &type->s_umount_key#73 &dentry->d_lock pool_lock#2 irq_context: 0 &type->s_umount_key#73 &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#73 &rq->__lock irq_context: 0 &type->s_umount_key#73 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal &rq->__lock cpu_asid_lock irq_context: 0 &map->freeze_mutex irq_context: 0 &type->s_umount_key#73 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#73 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#73 &dentry->d_lock/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &n->list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#73 &fsnotify_mark_srcu &rq->__lock irq_context: 0 &type->s_umount_key#73 &fsnotify_mark_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &xa->xa_lock#15 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &____s->seqcount#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 rcu_read_lock rcu_node_0 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &____s->seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#73 &dentry->d_lock rcu_read_lock &sb->s_type->i_lock_key#35 &p->pi_lock irq_context: 0 &type->s_umount_key#73 &dentry->d_lock rcu_read_lock &sb->s_type->i_lock_key#35 &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#73 &dentry->d_lock rcu_read_lock &sb->s_type->i_lock_key#35 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#73 rcu_read_lock rcu_node_0 irq_context: 0 &type->s_umount_key#73 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &type->s_umount_key#73 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->s_umount_key#73 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->s_umount_key#73 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#73 rcu_read_lock &rq->__lock irq_context: 0 &type->s_umount_key#73 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock percpu_counters_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 remove_cache_srcu irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#978 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#978 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#135 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 &____s->seqcount#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 &sb->s_type->i_mutex_key#19 &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_INET6 &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#371 irq_context: 0 &sb->s_type->i_mutex_key#9 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#993 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#972 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) quarantine_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) remove_cache_srcu irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex kfence_freelist_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex console_lock console_srcu console_owner &port_lock_key irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &n->list_lock &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &base->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) irq_context: 0 rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#14 mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#14 smack_known_lock irq_context: 0 &f->f_pos_lock sb_writers#14 smack_known_lock mmu_notifier_invalidate_range_start irq_context: 0 &f->f_pos_lock sb_writers#14 smack_known_lock &c->lock irq_context: 0 &f->f_pos_lock sb_writers#14 smack_known_lock pool_lock#2 irq_context: 0 &f->f_pos_lock sb_writers#14 smack_known_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#14 smack_known_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#14 smack_known_lock.wait_lock irq_context: 0 &f->f_pos_lock sb_writers#14 &p->pi_lock irq_context: 0 &f->f_pos_lock sb_writers#14 &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#14 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#14 smk_net6addr_lock irq_context: 0 &f->f_pos_lock sb_writers#14 smk_net6addr_lock fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#14 smk_net6addr_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 &f->f_pos_lock sb_writers#14 smk_net6addr_lock pool_lock#2 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#455 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#455 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy221 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &rcu_state.expedited_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#465 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#465 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy221 irq_context: 0 sb_writers#3 batched_entropy_u8.lock crngs.lock irq_context: 0 &f->f_pos_lock sb_writers#14 smack_known_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 kn->active#18 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc22_nci_rx_wq#12 &rq->__lock irq_context: 0 &type->s_umount_key#66 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 pcpu_alloc_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &c->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &wb->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem &wb->list_lock &sb->s_type->i_lock_key#22 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &bdi->cgwb_release_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &list->lock#5 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 alg_types_sem &rq->__lock irq_context: 0 alg_types_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#73 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 &sbi->s_writepages_rwsem jbd2_handle &ei->i_data_sem &ei->i_es_lock &c->lock irq_context: 0 &sbi->s_writepages_rwsem &rq_wait->wait irq_context: 0 &sbi->s_writepages_rwsem &__ctx->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 &type->s_umount_key#73 fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rlock-AF_NETLINK irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock &ei->socket.wq.wait irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 shrinker_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex fs_reclaim irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex stock_lock irq_context: 0 br_ioctl_mutex rtnl_mutex nf_hook_mutex pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex cpu_hotplug_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rlock-AF_NETLINK irq_context: 0 kn->active#53 fs_reclaim irq_context: 0 &sb->s_type->i_mutex_key#9 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 kn->active#53 fs_reclaim &rq->__lock irq_context: 0 kn->active#53 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#53 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#53 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 kn->active#53 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &root->kernfs_rwsem &base->lock &obj_hash[i].lock irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 kn->active#53 &kernfs_locks->open_file_mutex[count] &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh _xmit_ETHER#2 rcu_read_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock pool_lock#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &sbi->s_writepages_rwsem rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ei->i_data_sem &ei->i_prealloc_lock &pa->pa_lock#2 irq_context: 0 &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &ei->i_data_sem &obj_hash[i].lock irq_context: 0 &ei->i_data_sem &____s->seqcount irq_context: 0 &ei->i_data_sem pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 list_lrus_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &xa->xa_lock#15 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 stock_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex (work_completion)(&(&br->gc_work)->work) irq_context: 0 sb_writers#3 &xa->xa_lock#8 irq_context: 0 sb_writers#3 &mapping->i_private_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &sb->s_type->i_lock_key#22 &xa->xa_lock#8 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 lock#4 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &idev->mc_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 lock#4 &lruvec->lru_lock irq_context: 0 sb_writers#3 lock#5 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex kernfs_idr_lock irq_context: 0 sb_writers#3 &lruvec->lru_lock irq_context: 0 sb_writers#3 &sb->s_type->i_lock_key#22 &xa->xa_lock#8 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem pool_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work rcu_state.barrier_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &obj_hash[i].lock irq_context: 0 nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#73 fill_pool_map-wait-type-override pool_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 &data->open_mutex rfkill_global_mutex &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &____s->seqcount irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[0] irq_context: 0 br_ioctl_mutex rtnl_mutex dev_base_lock &xa->xa_lock#4 irq_context: 0 br_ioctl_mutex rtnl_mutex dev_base_lock &xa->xa_lock#4 &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex dev_base_lock &xa->xa_lock#4 pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex cpu_hotplug_lock &list->lock#12 irq_context: 0 br_ioctl_mutex rtnl_mutex bpf_devs_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &in_dev->mc_tomb_lock irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex sysctl_lock krc.lock irq_context: 0 br_ioctl_mutex rtnl_mutex class irq_context: 0 br_ioctl_mutex rtnl_mutex (&tbl->proxy_timer) irq_context: 0 br_ioctl_mutex rtnl_mutex &base->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &ul->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &net->xdp.lock irq_context: 0 br_ioctl_mutex rtnl_mutex mirred_list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &nft_net->commit_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &ent->pde_unload_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &net->ipv6.addrconf_hash_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &ndev->lock irq_context: 0 br_ioctl_mutex rtnl_mutex &ndev->lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key krc.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_query_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_query_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex (work_completion)(&(&idev->mc_report_work)->work) irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_report_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &idev->mc_lock krc.lock irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip4_mc_router_timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip4_other_query.timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip4_other_query.delay_timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip4_own_query.timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip6_mc_router_timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip6_other_query.timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip6_other_query.delay_timer) irq_context: 0 br_ioctl_mutex rtnl_mutex (&brmctx->ip6_own_query.timer) irq_context: 0 br_ioctl_mutex rtnl_mutex &pnn->pndevs.lock irq_context: 0 br_ioctl_mutex rtnl_mutex &pnn->routes.lock irq_context: 0 br_ioctl_mutex rtnl_mutex target_list_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &bridge_netdev_addr_lock_key irq_context: 0 br_ioctl_mutex rtnl_mutex &br->multicast_lock irq_context: 0 br_ioctl_mutex rtnl_mutex (work_completion)(&br->mcast_gc_work) irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex irq_context: 0 &dev->mutex leds_list_lock &rq->__lock irq_context: 0 &dev->mutex leds_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex &x->wait#24 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex pcpu_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &x->wait#2 irq_context: 0 br_ioctl_mutex rtnl_mutex (work_completion)(&ht->run_work) irq_context: 0 br_ioctl_mutex rtnl_mutex &ht->mutex irq_context: 0 br_ioctl_mutex rtnl_mutex &ht->mutex &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &ht->mutex pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex &ht->mutex &____s->seqcount irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &____s->seqcount#2 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_read_lock &ei->socket.wq.wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &meta->lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &____s->seqcount#2 irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex &____s->seqcount irq_context: 0 br_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 br_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 br_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 br_ioctl_mutex rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 br_ioctl_mutex rtnl_mutex subsys mutex#20 &k->k_lock klist_remove_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &sem->wait_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &p->pi_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 br_ioctl_mutex rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#4 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex rtnl_mutex deferred_probe_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex device_links_lock irq_context: 0 br_ioctl_mutex rtnl_mutex mmu_notifier_invalidate_range_start irq_context: 0 br_ioctl_mutex rtnl_mutex uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 br_ioctl_mutex rtnl_mutex cpu_hotplug_lock xps_map_mutex irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[3] irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex rcu_state.barrier_lock &obj_hash[i].lock irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex &x->wait#24 irq_context: 0 sk_lock-AF_BLUETOOTH-BTPROTO_HCI &____s->seqcount irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex &rq->__lock irq_context: 0 br_ioctl_mutex rcu_state.barrier_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 br_ioctl_mutex dev_base_lock irq_context: 0 br_ioctl_mutex lweventlist_lock irq_context: 0 br_ioctl_mutex stock_lock irq_context: 0 br_ioctl_mutex &obj_hash[i].lock irq_context: 0 br_ioctl_mutex pool_lock#2 irq_context: 0 br_ioctl_mutex krc.lock irq_context: 0 br_ioctl_mutex &dir->lock#2 irq_context: 0 br_ioctl_mutex &dir->lock#2 &obj_hash[i].lock irq_context: 0 br_ioctl_mutex &dir->lock#2 pool_lock#2 irq_context: 0 br_ioctl_mutex netdev_unregistering_wq.lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &cfs_rq->removed.lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex pool_lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.barrier_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: softirq net/core/link_watch.c:31 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 sb_writers#4 tomoyo_ss &____s->seqcount#2 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &sb->s_type->i_lock_key#22 &xa->xa_lock#8 pool_lock#2 irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &lruvec->lru_lock irq_context: 0 sb_writers#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &sb->s_type->i_lock_key#22 &xa->xa_lock#8 &obj_hash[i].lock pool_lock irq_context: 0 &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#53 &c->lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[1] irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_state.exp_mutex &rnp->exp_wq[2] irq_context: 0 br_ioctl_mutex &obj_hash[i].lock pool_lock irq_context: 0 &xt[i].mutex &mm->mmap_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem kernfs_idr_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex &root->kernfs_rwsem kernfs_idr_lock pool_lock#2 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#363 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#363 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#363 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#361 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#787 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#787 &cfs_rq->removed.lock irq_context: 0 namespace_sem fs_reclaim &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#787 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#787 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#782 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#782 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#782 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#779 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#372 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#372 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#364 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#364 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#364 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem irq_context: 0 (wq_completion)nfc3_nci_tx_wq#362 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#373 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#373 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#373 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#373 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#365 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#363 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &ei->i_prealloc_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#788 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#788 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#783 irq_context: 0 (wq_completion)nfc22_nci_rx_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc22_nci_tx_wq#12 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#12 &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &ei->i_raw_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc14_nci_tx_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc15_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#105 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#629 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem quarantine_lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock pool_lock#2 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &ei->i_es_lock &sbi->s_es_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &c->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &n->list_lock &c->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem pool_lock#2 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock &sbi->s_md_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &sb->s_type->i_lock_key#22 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem key#3 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &ret->b_state_lock &journal->j_list_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &journal->j_revoke_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle tk_core.seq.seqcount irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_prealloc_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &journal->j_list_lock irq_context: 0 (wq_completion)ext4-rsv-conversion (work_completion)(&ei->i_rsv_conversion_work) &folio_wait_table[i] &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &ei->i_data_sem &bgl->locks[i].lock pool_lock#2 irq_context: 0 &type->s_umount_key#57 pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#629 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#780 irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#140 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#140 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#21/1 list_lrus_mutex &rq->__lock irq_context: 0 &type->s_umount_key#21/1 list_lrus_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex &wg->device_update_lock fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#93 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#93 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#93 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &sb->s_type->i_lock_key#3 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &s->s_inode_list_lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#91 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#789 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#789 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#89 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#205 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &sb->s_type->i_lock_key#3 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#40 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#784 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#768 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#768 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#768 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#763 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#760 irq_context: 0 (wq_completion)wg-kex-wg0#207 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#118 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#118 (console_sem).lock irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc20_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc20_nci_rx_wq#10 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)events &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 rtnl_mutex team->team_lock_key#118 console_lock console_srcu console_owner_lock irq_context: 0 &type->s_umount_key#40 &x->wait#23 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex batched_entropy_u32.lock irq_context: 0 rtnl_mutex team->team_lock_key#118 console_lock console_srcu console_owner irq_context: 0 (wq_completion)nfc20_nci_tx_wq#10 irq_context: hardirq &fq->mq_flush_lock bit_wait_table + i &p->pi_lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key#118 console_lock console_srcu console_owner &port_lock_key irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#629 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#630 irq_context: 0 rtnl_mutex team->team_lock_key#118 console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#118 _xmit_ETHER &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#630 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#627 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#639 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#288 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#670 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#670 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#678 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#684 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#781 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#140 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#138 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#206 irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 &type->s_umount_key#40 shrinker_mutex irq_context: 0 &type->s_umount_key#40 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#972 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#972 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq &(&ovs_net->masks_rebalance)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1003 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex fs_reclaim irq_context: 0 &ndev->req_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &xa->xa_lock#4 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 (work_completion)(&local->sdreq_timeout_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh &hsr->seqnr_lock rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_tx_wq#972 &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#972 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#452 irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 &sb->s_type->i_mutex_key#10 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 file_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#452 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#444 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#441 irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &dev->power.lock irq_context: 0 &mm->mmap_lock fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc11_nci_rx_wq#12 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex pcpu_alloc_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1005 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex &rq->__lock cpu_asid_lock irq_context: 0 (work_completion)(&tty->SAK_work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#986 irq_context: 0 &type->lock_class &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#986 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#979 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#973 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#104 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#208 irq_context: 0 &vma->vm_lock->lock remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 &vma->vm_lock->lock remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 &vma->vm_lock->lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#207 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#987 irq_context: 0 &type->s_umount_key#40 rename_lock.seqcount irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#987 irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &rq->__lock irq_context: 0 wq_pool_attach_mutex wq_pool_attach_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex.wait_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 wq_pool_attach_mutex.wait_lock irq_context: 0 wq_pool_attach_mutex &cfs_rq->removed.lock irq_context: 0 wq_pool_attach_mutex &obj_hash[i].lock irq_context: 0 wq_pool_attach_mutex pool_lock#2 irq_context: 0 wq_pool_attach_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#40 &dentry->d_lock irq_context: 0 &type->s_umount_key#40 rcu_read_lock &dentry->d_lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock batched_entropy_u8.lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock kfence_freelist_lock irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &meta->lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#40 &sb->s_type->i_lock_key#3 irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#9 &sb->s_type->i_lock_key#23 &dentry->d_lock &wq#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock fs_reclaim &rq->__lock irq_context: 0 key_types_sem &type->lock_class keyring_serialise_link_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 &dev->mutex uevent_sock_mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#96 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#4 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#200 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#200 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#499 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#499 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#496 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#494 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#79 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#79 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#79 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#79 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#77 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#76 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#76 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#500 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#500 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#497 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#495 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#206 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#206 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#206 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#206 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#202 irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1007 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#466 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#86 irq_context: 0 link_idr_lock irq_context: 0 rtnl_mutex fs_reclaim stock_lock irq_context: 0 &po->bind_lock ptype_lock irq_context: 0 fanout_mutex irq_context: 0 clock-AF_PACKET irq_context: 0 rtnl_mutex dev_addr_sem fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&w->work)#2 &rq->__lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex &br->lock &br->hash_lock rcu_read_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex quarantine_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#201 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#501 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#501 irq_context: 0 rtnl_mutex fs_reclaim pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#498 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#496 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#496 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#496 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#80 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#80 irq_context: 0 &type->s_umount_key#40 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#40 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#40 pool_lock#2 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#78 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#27 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#27 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#77 irq_context: 0 nfc_devlist_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#207 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#207 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#203 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#203 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#203 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#192 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1016 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1016 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#188 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#188 &rq->__lock irq_context: softirq (&icsk->icsk_delack_timer) slock-AF_INET#2 &____s->seqcount#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#202 irq_context: 0 &data->open_mutex lock kernfs_idr_lock &____s->seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#3 irq_context: 0 &type->s_umount_key#40 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#40 &dentry->d_lock/1 irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem rcu_read_lock &obj_hash[i].lock pool_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &xa->xa_lock#15 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &xa->xa_lock#15 &n->list_lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 &sighand->siglock &n->list_lock irq_context: 0 &sighand->siglock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci4#4 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) &list->lock#5 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 sk_lock-AF_INET6 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#362 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#362 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#362 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex uevent_sock_mutex &base->lock irq_context: 0 rtnl_mutex uevent_sock_mutex &base->lock &obj_hash[i].lock irq_context: 0 &type->s_umount_key#73 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)nfc4_nci_tx_wq#133 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#133 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#769 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &xa->xa_lock#15 &n->list_lock &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &base->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#769 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#764 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#761 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#761 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#761 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#761 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#761 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#3/1 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#762 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#363 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#363 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#363 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#355 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#136 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_INET6 slock-AF_INET6 key#23 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#980 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#974 irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#211 irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)mm_percpu_wq (work_completion)(&(({ do { const void *__vpp_verify = (typeof((&vmstat_work) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((&vmstat_work))) *)((&vmstat_work))); (typeof((typeof(*((&vmstat_work))) *)((&vmstat_work)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); }))->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#136 &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1030 irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)nfc4_nci_tx_wq#136 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#353 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#139 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#137 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#793 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#793 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#793 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#793 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#788 irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem sched_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#801 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#801 irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &hdev->lock remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#541 &rq->__lock irq_context: softirq mm/memcontrol.c:679 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#466 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#466 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#466 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#458 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#454 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#467 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#467 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#467 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#467 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#459 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#455 irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#472 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#472 &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#455 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &journal->j_list_lock quarantine_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#455 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->managed_work)->work) &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#937 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1010 irq_context: 0 &sig->cred_guard_mutex remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#468 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#211 irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 &sb->s_type->i_mutex_key#10 unix_gc_lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)hci0#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1012 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &mapping->i_private_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#451 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#788 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#788 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#785 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq rcu_callback rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq rcu_callback rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: softirq rcu_callback rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc40_nci_cmd_wq#6 irq_context: 0 rtnl_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#451 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &base->lock irq_context: 0 (wq_completion)nfc36_nci_tx_wq#6 irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 rtnl_mutex &br->hash_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 &mm->mmap_lock sb_writers#5 &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock sb_writers#5 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1012 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#462 irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#462 &rq->__lock irq_context: 0 &type->s_umount_key#30 (wq_completion)ext4-rsv-conversion &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1013 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &n->list_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#456 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#448 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#446 irq_context: 0 (wq_completion)hci3#5 irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#794 irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#794 irq_context: 0 (wq_completion)hci3#6 irq_context: softirq (&timer) rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) &list->lock#5 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#174 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#174 irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)nfc3_nci_rx_wq#170 irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 rtnl_mutex gdp_mutex &n->list_lock irq_context: 0 rtnl_mutex gdp_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem stock_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 cb_lock genl_mutex rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount#2 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_power_efficient (gc_work).work &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1013 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 (netlink_chain).rwsem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc11_nci_tx_wq#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#106 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &entry->crc_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#207 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#208 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#13 irq_context: 0 namespace_sem remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#13 irq_context: 0 (wq_completion)wg-kex-wg0#209 irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 rtnl_mutex &idev->mc_lock &dev_addr_list_lock_key#3 &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 &xt[i].mutex remove_cache_srcu kfence_freelist_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#179 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1016 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1016 &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: softirq &(&tbl->gc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu &n->list_lock irq_context: 0 &hdev->req_lock &hdev->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)nfc2_nci_rx_wq#789 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#730 irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex percpu_counters_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex pcpu_lock stock_lock irq_context: 0 ebt_mutex rcu_read_lock &rq->__lock irq_context: 0 ebt_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#720 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#720 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#730 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#739 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &tbl->lock nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#750 irq_context: 0 &type->i_mutex_dir_key#2 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#2 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#750 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 sb_writers#3 jbd2_handle rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)cfg80211 (work_completion)(&rdev->event_work) &rdev->wiphy.mtx &rq->__lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &cfs_rq->removed.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#750 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &u->iolock &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 bt_proto_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock batched_entropy_u8.lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET#2 &obj_hash[i].lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#930 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#3 &____s->seqcount#2 irq_context: 0 &type->i_mutex_dir_key#5 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#750 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nfc2_nci_tx_wq#956 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#970 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&aux->work) purge_vmap_area_lock kfence_freelist_lock irq_context: 0 &f->f_pos_lock sb_writers#14 &n->list_lock irq_context: 0 &f->f_pos_lock sb_writers#14 &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#943 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#757 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#471 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1042 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (work_completion)(&local->tx_work) &cfs_rq->removed.lock irq_context: 0 (work_completion)(&local->tx_work) &obj_hash[i].lock irq_context: 0 (work_completion)(&local->tx_work) pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#590 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#590 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx.wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#597 irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq#14 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex remove_cache_srcu &____s->seqcount irq_context: 0 &fc->uapi_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#256 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1232 irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex stock_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex key irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex pcpu_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex percpu_counters_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#151 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#433 irq_context: 0 pidmap_lock &obj_hash[i].lock irq_context: 0 &rq->__lock cid_lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override batched_entropy_u8.lock irq_context: 0 tasklist_lock &sighand->siglock &(&sig->stats_lock)->lock &____s->seqcount#4 pidmap_lock fill_pool_map-wait-type-override kfence_freelist_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#440 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#441 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#445 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#702 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#169 irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle smack_known_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] rcu_node_0 irq_context: 0 rtcdev_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &____s->seqcount irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1016 &obj_hash[i].lock irq_context: 0 sk_lock-AF_INET rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &base->lock irq_context: 0 &type->i_mutex_dir_key#4 &base->lock &obj_hash[i].lock irq_context: 0 &p->lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq (&peer->timer_send_keepalive) init_task.mems_allowed_seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 rtnl_mutex uevent_sock_mutex stock_lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: softirq &(&conn->disc_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#786 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#786 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#786 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#375 irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 &f->f_pos_lock sb_writers#14 smk_net6addr_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#375 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#375 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#375 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#367 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#365 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#365 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#48 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#46 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#45 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#767 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#767 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#762 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#759 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#759 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#759 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#361 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#361 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#353 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#351 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#365 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#795 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#795 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#790 irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 &pipe->wr_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1014 irq_context: 0 key_types_sem root_key_user.cons_lock &____s->seqcount#2 irq_context: 0 key_types_sem root_key_user.cons_lock &pcp->lock &zone->lock irq_context: 0 key_types_sem root_key_user.cons_lock &____s->seqcount irq_context: 0 &sig->cred_guard_mutex sb_writers#3 mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 key_types_sem init_user_ns.keyring_sem rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1014 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1014 irq_context: 0 kn->active#14 remove_cache_srcu irq_context: 0 kn->active#14 remove_cache_srcu &rq->__lock irq_context: 0 kn->active#14 remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1006 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1000 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1000 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#472 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#472 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#464 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#464 &rq->__lock irq_context: 0 (wq_completion)events key_gc_work &type->lock_class remove_cache_srcu &c->lock irq_context: 0 (wq_completion)events key_gc_work &type->lock_class remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#787 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#452 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#746 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#142 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#142 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#140 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem kn->active#4 irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#464 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#460 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#185 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#138 irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 fh->state->lock &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#27 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#23 &p->pi_lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#23 &p->pi_lock &rq->__lock irq_context: 0 rcu_read_lock &dentry->d_lock &sb->s_type->i_lock_key#23 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#63 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#376 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#376 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#796 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#625 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#796 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#796 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &meta_group_info[i]->alloc_sem pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#796 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &____s->seqcount#2 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers &type->i_mutex_dir_key#2 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &br->hash_lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#791 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_CAN j1939_netdev_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#791 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#791 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &ret->b_state_lock bit_wait_table + i &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &ret->b_state_lock bit_wait_table + i &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 clock-AF_INET irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#143 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#368 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#143 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#368 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#143 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#368 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#788 irq_context: 0 rcu_read_lock rhashtable_bucket irq_context: 0 clock-AF_NETLINK irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#209 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#207 irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock krc.lock &base->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 pcpu_alloc_mutex irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 pcpu_alloc_mutex pcpu_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 cb_lock &xa->xa_lock#16 irq_context: 0 genl_sk_destructing_waitq.lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#368 &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu rcu_node_0 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#11 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc23_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc23_nci_rx_wq#11 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_rx_wq#368 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc23_nci_tx_wq#11 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#9 &rq->__lock irq_context: 0 (wq_completion)hci5 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#143 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 &f->f_pos_lock sb_writers#4 batched_entropy_u8.lock irq_context: 0 &f->f_pos_lock sb_writers#4 kfence_freelist_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#4 &meta->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 br_ioctl_mutex rtnl_mutex rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &xt[i].mutex remove_cache_srcu rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &ei->i_data_sem mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock stock_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock key irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock pcpu_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock percpu_counters_lock irq_context: 0 cb_lock genl_mutex uevent_sock_mutex rcu_read_lock pcpu_lock stock_lock irq_context: 0 sk_lock-AF_CAN &priv->lock &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1017 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1017 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1008 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1002 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#96 irq_context: 0 (wq_completion)wg-kex-wg0#217 irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#95 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &mm->mmap_lock stock_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 &ep->mtx fs_reclaim &cfs_rq->removed.lock irq_context: 0 rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#366 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#143 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#143 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#141 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#139 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#797 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc24_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc24_nci_rx_wq#9 irq_context: 0 (wq_completion)nfc24_nci_tx_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc26_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc26_nci_rx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc26_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#9 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &list->lock#5 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 cb_lock rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#193 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &base->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#106 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1019 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#100 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13/4 &simple_offset_xa_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex pool_lock#2 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#452 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 shrinker_mutex irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 &pl->lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &xa->xa_lock#8 &pl->lock key#12 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex quarantine_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 &data->open_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)nfc2_nci_rx_wq#449 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#797 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#792 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#789 irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) &list->lock#5 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock rcu_read_lock &pool->lock irq_context: 0 &hdev->req_lock quarantine_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock (work_completion)(flush) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock &x->wait#10 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#199 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg2#200 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 remove_cache_srcu &____s->seqcount irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock rcu_node_0 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1020 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#473 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#100 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#199 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#200 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#64 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#64 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#62 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#103 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#201 irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &n->list_lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) rcu_state.exp_wake_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#202 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#185 irq_context: softirq rcu_read_lock rcu_read_lock slock-AF_INET/1 rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock &sch->q.lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#185 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#185 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock kfence_freelist_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock &tbl->lock &meta->lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#65 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#63 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#188 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#188 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#186 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#186 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#186 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#47 &n->list_lock irq_context: 0 kn->active#47 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &root->kernfs_rwsem &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 kn->active#20 remove_cache_srcu irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 kn->active#20 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#20 remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 kn->active#20 remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex uevent_sock_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &n->list_lock irq_context: 0 rtnl_mutex &idev->mc_lock &bridge_netdev_addr_lock_key &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1021 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1021 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1012 irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1006 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#474 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#474 irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1022 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1022 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1022 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1022 irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1023 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1023 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1014 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#475 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#184 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#184 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1024 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1024 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1026 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1026 &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &bat_priv->forw_bat_list_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1026 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#464 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#464 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1027 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1027 irq_context: 0 kn->active#20 remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1017 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1017 &rq->__lock irq_context: 0 &mm->mmap_lock fs_reclaim rcu_node_0 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &net->xdp.lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &net->xdp.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#465 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1028 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1028 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1018 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1012 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#478 irq_context: 0 &dev->mutex kn->active#4 &rq->__lock cpu_asid_lock irq_context: 0 &p->lock &of->mutex &root->deactivate_waitq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1030 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1030 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1030 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1020 irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1014 irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock krc.lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock krc.lock &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) devices_rwsem rcu_node_0 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1031 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1031 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1021 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1032 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1032 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1032 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#479 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#479 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#471 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1033 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#472 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem kfence_freelist_lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem &meta->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#468 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1034 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1034 &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1034 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1034 irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1035 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1035 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1035 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1035 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1036 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1036 irq_context: 0 binderfs_minors_mutex &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#482 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#482 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#482 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#482 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#474 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#470 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1027 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1027 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1027 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1038 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock stock_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock key irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock pcpu_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock percpu_counters_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1039 irq_context: 0 proto_tab_lock rcu_read_lock &p->pi_lock irq_context: 0 proto_tab_lock rcu_read_lock &p->pi_lock &rq->__lock irq_context: 0 proto_tab_lock rcu_read_lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nf_hook_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1029 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1023 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1041 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1042 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#484 irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1043 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1043 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1043 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq &(&bat_priv->nc.work)->timer rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1043 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#191 irq_context: 0 cb_lock rtnl_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &pipe->mutex/1 sched_map-wait-type-override &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#191 &rq->__lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock#2 rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#207 irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#377 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#377 &rq->__lock irq_context: 0 (wq_completion)nfc29_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc29_nci_rx_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#377 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#377 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#369 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#144 irq_context: 0 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sk_lock-AF_INET6 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 rtnl_mutex (switchdev_blocking_notif_chain).rwsem &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#367 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#367 &rq->__lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex key irq_context: 0 cb_lock genl_mutex rfkill_global_mutex pcpu_lock irq_context: 0 cb_lock genl_mutex rfkill_global_mutex percpu_counters_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#596 irq_context: 0 &pipe->mutex/1 sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)nfc29_nci_rx_wq#9 &rq->__lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &rq->__lock cpu_asid_lock irq_context: 0 nf_hook_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)nfc29_nci_rx_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc29_nci_tx_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq#9 irq_context: 0 rtnl_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex wq_pool_mutex &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 nf_hook_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#6 &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#175 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#453 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#461 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 tty_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&data->fib_event_work) &data->fib_lock remove_cache_srcu kfence_freelist_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &base->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#461 irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#182 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#470 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#470 &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#471 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#188 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#190 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#190 &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#477 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#144 &rq->__lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci2#2 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (crypto_chain).rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (crypto_chain).rwsem &pcp->lock &zone->lock irq_context: 0 (crypto_chain).rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &base->lock irq_context: 0 &ep->mtx &mm->mmap_lock &cfs_rq->removed.lock irq_context: 0 &ep->mtx &mm->mmap_lock &obj_hash[i].lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx net_rwsem &c->lock irq_context: 0 sb_writers#9 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) batched_entropy_u8.lock crngs.lock irq_context: 0 sk_lock-AF_X25 &mm->mmap_lock rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 rtnl_mutex &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#103 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg2#103 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#92 irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#92 &rq->__lock irq_context: 0 &data->open_mutex rfkill_global_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#92 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#92 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#90 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#243 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#243 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#243 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#88 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#243 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#239 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#238 irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 &data->open_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#208 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &dev->tx_global_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#564 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#564 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#561 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#559 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#559 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#559 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&(&kfence_timer)->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &wg->device_update_lock rcu_state.exp_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem nf_nat_proto_mutex nf_hook_mutex &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#619 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#619 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) &list->lock#5 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#700 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#705 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#705 &rq->__lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#477 &rq->__lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#743 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#348 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#348 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#751 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#751 irq_context: 0 &type->s_umount_key#30 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &____s->seqcount#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#367 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#144 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#144 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#367 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#367 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#142 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#140 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#140 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#140 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#378 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#378 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#378 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 fs_reclaim &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &type->i_mutex_dir_key#3 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc30_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc30_nci_rx_wq#9 irq_context: 0 sk_lock-AF_CAN j1939_netdev_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)nfc30_nci_tx_wq#9 irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 cb_lock quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#71 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#482 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#378 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) stock_lock rcu_read_lock per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc30_nci_tx_wq#9 &rq->__lock irq_context: 0 (wq_completion)nfc30_nci_tx_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#370 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#368 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 tomoyo_ss remove_cache_srcu rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock irq_context: 0 kn->active#4 remove_cache_srcu rcu_node_0 irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) pcpu_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) &x->wait#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#798 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#798 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#793 irq_context: 0 pernet_ops_rwsem quarantine_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#790 irq_context: 0 pernet_ops_rwsem pcpu_alloc_mutex pcpu_alloc_mutex.wait_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &rcu_state.expedited_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&map->work) pool_lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &base->lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock rcu_node_0 irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 rtnl_mutex rcu_read_lock rcu_read_lock rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &kernfs_locks->open_file_mutex[count] &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &kernfs_locks->open_file_mutex[count] &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#104 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#925 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#799 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#799 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#794 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#791 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#791 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#800 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#800 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#795 irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#792 irq_context: 0 &sig->cred_guard_mutex &ei->xattr_sem rcu_read_lock pool_lock#2 irq_context: 0 tty_mutex fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#196 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#196 irq_context: 0 misc_mtx nfc_devlist_mutex uevent_sock_mutex quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#488 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#379 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#379 &rq->__lock irq_context: 0 &type->s_umount_key#61 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#379 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#379 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#371 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#369 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#13 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &____s->seqcount irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#278 irq_context: 0 (wq_completion)nfc18_nci_rx_wq#16 &rq->__lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)events_long (work_completion)(&(&ipvs->defense_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 &sb->s_type->i_mutex_key#10 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 kn->active#20 &kernfs_locks->open_file_mutex[count] fs_reclaim &rq->__lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock kfence_freelist_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 sk_lock-AF_INET remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: hardirq rcu_read_lock &xa->xa_lock#8 &obj_hash[i].lock irq_context: hardirq rcu_read_lock &xa->xa_lock#8 &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &obj_hash[i].lock pool_lock irq_context: 0 &sig->cred_guard_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 rtnl_mutex nf_hook_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->power_on) &n->list_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: hardirq rcu_read_lock &xa->xa_lock#8 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 misc_mtx &dev->mutex batched_entropy_u8.lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex kfence_freelist_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)nfc20_nci_rx_wq#13 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#66 fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)nfc20_nci_rx_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc20_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc33_nci_tx_wq#12 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#796 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#796 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#796 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#793 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci0#5 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 vlan_ioctl_mutex rtnl_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 misc_mtx (wq_completion)nfc43_nci_tx_wq#4 irq_context: 0 misc_mtx (wq_completion)nfc43_nci_rx_wq#4 irq_context: 0 misc_mtx (wq_completion)nfc43_nci_cmd_wq#4 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)bond0#101 irq_context: 0 (wq_completion)bond0#101 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex kernfs_idr_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bond0#101 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#101 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#101 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock percpu_counters_lock irq_context: softirq &jsk->sk_session_queue_lock irq_context: 0 sk_lock-AF_CAN &obj_hash[i].lock pool_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx bus_type_sem &rq->__lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) &list->lock#5 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci5#2 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci5#2 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#46 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->tx_work) &list->lock#5 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#447 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#175 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)nfc3_nci_rx_wq#171 irq_context: 0 misc_mtx nfc_devlist_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#131 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#131 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#753 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex sched_map-wait-type-override rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#7 &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock rcu_node_0 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 &mm->mmap_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#208 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#754 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#970 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#963 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#963 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#957 irq_context: 0 (wq_completion)bond0#116 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#116 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#116 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#964 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#958 irq_context: 0 rtnl_mutex team->team_lock_key#116 irq_context: 0 rtnl_mutex team->team_lock_key#116 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#116 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#116 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#101 irq_context: 0 rtnl_mutex team->team_lock_key#101 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#101 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#101 netpoll_srcu irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&gc_work->dwork)->work) &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#101 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#101 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#101 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#101 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#101 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#101 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#101 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#101 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#101 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 rtnl_mutex team->team_lock_key#101 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#101 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#101 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#101 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#101 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#101 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#101 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#101 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#101 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 lock irq_context: 0 rtnl_mutex team->team_lock_key#101 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#101 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#101 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#101 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#101 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#101 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#101 console_lock console_srcu console_owner console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 &obj_hash[i].lock pool_lock irq_context: 0 rtnl_mutex team->team_lock_key#101 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#101 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sk_lock-AF_CAN rcu_node_0 irq_context: 0 rtnl_mutex team->team_lock_key#101 rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#101 rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#116 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#116 &tn->lock irq_context: 0 sk_lock-AF_CAN fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#116 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#116 &dir->lock#2 irq_context: 0 &type->i_mutex_dir_key#5 batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#701 irq_context: 0 rtnl_mutex team->team_lock_key#116 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#116 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#116 &____s->seqcount#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#340 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#755 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#755 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#755 irq_context: 0 rtnl_mutex team->team_lock_key#116 &____s->seqcount irq_context: 0 rtnl_mutex team->team_lock_key#116 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#116 &pcp->lock &zone->lock irq_context: 0 rtnl_mutex team->team_lock_key#116 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#116 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 quarantine_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#116 remove_cache_srcu irq_context: 0 rtnl_mutex team->team_lock_key#116 remove_cache_srcu quarantine_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 remove_cache_srcu &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#116 remove_cache_srcu &n->list_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#973 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#973 irq_context: 0 &mm->mmap_lock remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#707 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#708 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#708 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#708 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#45 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#327 irq_context: 0 rtnl_mutex team->team_lock_key#116 remove_cache_srcu &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#116 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#116 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &xt[i].mutex remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rcu_state.expedited_wq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#943 irq_context: 0 sb_writers#7 &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci0#6 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#327 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#710 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#710 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#710 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#706 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#329 irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#329 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#713 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#330 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#330 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#324 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#322 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#715 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#331 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#331 &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#331 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#123 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#711 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#333 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#333 &rq->__lock irq_context: 0 (wq_completion)bond0#101 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#101 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#101 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#333 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#128 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 rcu_read_lock &ndev->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#47 irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sk_lock-AF_CAN remove_cache_srcu &____s->seqcount irq_context: 0 (wq_completion)nfc35_nci_cmd_wq#7 irq_context: 0 (wq_completion)nfc35_nci_rx_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#47 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#720 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#716 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#327 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#722 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#722 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#718 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#718 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#724 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#724 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#724 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#724 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#725 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#725 irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc35_nci_rx_wq#7 &rq->__lock irq_context: 0 (wq_completion)nfc35_nci_tx_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#245 irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#245 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#241 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#945 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#727 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#728 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#729 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#729 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#729 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#339 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#726 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#731 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#731 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#631 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#284 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#731 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#727 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex &tbl->lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#725 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#732 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#732 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#732 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#732 irq_context: 0 &ndev->req_lock (wq_completion)nfc42_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc42_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc42_nci_rx_wq#5 irq_context: 0 (wq_completion)nfc42_nci_tx_wq#5 irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim &obj_hash[i].lock irq_context: 0 pernet_ops_rwsem nf_ct_proto_mutex defrag6_mutex nf_hook_mutex fs_reclaim pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc41_nci_cmd_wq#5 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)nfc2_nci_rx_wq#728 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#341 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#733 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#733 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#633 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#633 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#285 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#635 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &meta->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 &tn->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &x->wait#2 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#729 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#730 irq_context: 0 (wq_completion)hci0#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#936 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#930 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#944 irq_context: 0 (wq_completion)nfc41_nci_cmd_wq#5 irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc41_nci_rx_wq#5 irq_context: 0 &ep->mtx fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &ep->mtx fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc41_nci_rx_wq#5 &rq->__lock irq_context: 0 (wq_completion)nfc41_nci_rx_wq#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc41_nci_tx_wq#5 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &journal->j_state_lock &journal->j_wait_commit &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&net->ipv6.addr_chk_work)->work) rtnl_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc40_nci_cmd_wq#6 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) &base->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start stock_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start key irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start percpu_counters_lock irq_context: 0 &mm->mmap_lock fs_reclaim mmu_notifier_invalidate_range_start pcpu_lock stock_lock irq_context: 0 &dev->mutex uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc40_nci_rx_wq#6 irq_context: 0 (wq_completion)nfc40_nci_rx_wq#6 &rq->__lock irq_context: 0 (wq_completion)nfc40_nci_rx_wq#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc40_nci_tx_wq#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc39_nci_cmd_wq#5 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex net_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &tn->lock irq_context: 0 (wq_completion)nfc39_nci_cmd_wq#5 irq_context: 0 &sig->cred_guard_mutex sb_writers#3 jbd2_handle &____s->seqcount#2 irq_context: 0 &mm->mmap_lock rcu_read_lock rcu_read_lock rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc39_nci_rx_wq#5 irq_context: 0 (wq_completion)nfc39_nci_rx_wq#5 &rq->__lock irq_context: 0 (wq_completion)nfc39_nci_rx_wq#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->mark_mutex stock_lock irq_context: 0 (wq_completion)nfc39_nci_tx_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc38_nci_cmd_wq#5 irq_context: 0 (wq_completion)events_unbound (work_completion)(&barr->work) &x->wait#10 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#925 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#940 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#940 irq_context: 0 (wq_completion)events_unbound (reaper_work).work stock_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#941 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci3#6 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci3#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)nfc38_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#170 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#453 irq_context: 0 &group->mark_mutex rcu_read_lock rcu_node_0 irq_context: 0 &group->mark_mutex rcu_read_lock &rq->__lock irq_context: 0 &group->mark_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock batched_entropy_u8.lock irq_context: 0 &group->mark_mutex lock &group->inotify_data.idr_lock kfence_freelist_lock irq_context: 0 rcu_read_lock &pool->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 &dentry->d_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 &sb->s_type->i_mutex_key#8 rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#453 &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#14 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &f->f_pos_lock sb_writers#14 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#453 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#453 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#240 irq_context: 0 misc_mtx (wq_completion)nfc35_nci_cmd_wq#8 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#246 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#242 irq_context: 0 &ndev->req_lock (wq_completion)nfc37_nci_cmd_wq#4 irq_context: 0 &xt[i].mutex remove_cache_srcu fill_pool_map-wait-type-override &c->lock irq_context: 0 kn->active#20 &pcp->lock &zone->lock irq_context: 0 &xt[i].mutex remove_cache_srcu fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc37_nci_cmd_wq#4 &rq->__lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc34_nci_rx_wq#9 irq_context: 0 (wq_completion)nfc29_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc26_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc22_nci_tx_wq#11 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#11 irq_context: 0 (wq_completion)bond0#110 irq_context: 0 wq_pool_attach_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 wq_pool_attach_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 wq_pool_attach_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->bla.work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#62 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#62 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#450 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#448 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#60 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#59 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#454 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#454 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#451 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss &mm->mmap_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock pool_lock#2 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock &ul->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#451 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#451 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#449 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#449 &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock rcu_node_0 irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 cb_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#455 irq_context: 0 (wq_completion)bond0#106 irq_context: 0 (wq_completion)bond0#106 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#106 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#106 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bond0#106 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#455 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#455 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#176 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#176 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#455 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#452 irq_context: 0 (wq_completion)bond0#106 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#106 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#106 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#106 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#106 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock &obj_hash[i].lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock &base->lock irq_context: 0 (wq_completion)rcu_gp (work_completion)(&rew->rew_work) &rq->__lock &base->lock &obj_hash[i].lock irq_context: softirq slock-AF_INET#2 rcu_read_lock rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#450 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#172 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#171 irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 tty_mutex remove_cache_srcu &c->lock irq_context: 0 tty_mutex remove_cache_srcu &n->list_lock irq_context: 0 tty_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 tty_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&hwstats->traffic_dw)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#215 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 fill_pool_map-wait-type-override &rq->__lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_X25 fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 &type->i_mutex_dir_key#4 stock_lock irq_context: 0 &type->i_mutex_dir_key#4 key irq_context: 0 &type->i_mutex_dir_key#4 pcpu_lock irq_context: 0 &type->i_mutex_dir_key#4 percpu_counters_lock irq_context: 0 &type->i_mutex_dir_key/1 &sb->s_type->i_mutex_key#4 &sem->wait_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock pcpu_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock percpu_counters_lock irq_context: 0 &type->i_mutex_dir_key#4 pcpu_lock stock_lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 tomoyo_ss remove_cache_srcu rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#218 irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq irq_context: 0 fs_reclaim mmu_notifier_invalidate_range_start &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#456 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#451 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#457 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#457 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#457 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &mm->mmap_lock sb_writers#3 rcu_node_0 irq_context: 0 &mm->mmap_lock sb_writers#3 &rcu_state.expedited_wq irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] kfence_freelist_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock sb_writers#3 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#457 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#454 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#452 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#177 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#177 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#173 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#172 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#4 sb_writers#7 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#63 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#63 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#63 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#61 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#60 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#458 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#458 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#455 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#453 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#459 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#459 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#459 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#459 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#456 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#454 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#454 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#454 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#178 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#178 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#174 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#173 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#64 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#64 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#62 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#61 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#460 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#460 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#457 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#455 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#458 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#456 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#179 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#179 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#175 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#175 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#175 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#174 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#174 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#174 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#174 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#174 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#65 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#65 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#63 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#62 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#462 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#462 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#459 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#457 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#463 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#463 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#463 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#463 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#460 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#458 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#180 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#180 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#176 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#176 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#176 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#175 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#464 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#464 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#461 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#459 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#459 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#459 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#66 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#66 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#64 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#181 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#181 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#63 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#177 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#176 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#176 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#176 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#465 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#465 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#465 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#465 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#462 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#460 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#466 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#466 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#466 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#466 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#463 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#461 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#461 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#461 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#67 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#67 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#65 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#64 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#64 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#64 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#467 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#467 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#464 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#462 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#182 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#178 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#177 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#468 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#468 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#468 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#468 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#465 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#463 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#463 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#463 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#183 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#183 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#179 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#178 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#68 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#68 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#66 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#65 irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#184 irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#184 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#184 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#184 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#469 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#469 irq_context: 0 (wq_completion)nfc38_nci_rx_wq#5 irq_context: 0 (wq_completion)nfc38_nci_tx_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc37_nci_cmd_wq#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc37_nci_cmd_wq#6 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc37_nci_cmd_wq#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc37_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc37_nci_rx_wq#6 irq_context: 0 (wq_completion)nfc37_nci_tx_wq#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc33_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc33_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc33_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc36_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc36_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc36_nci_rx_wq#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#11 irq_context: 0 &f->f_pos_lock sb_writers#14 smack_known_lock rcu_node_0 irq_context: 0 (wq_completion)nfc35_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc35_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc35_nci_tx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc34_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc34_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc34_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#11 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc32_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc32_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc32_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc31_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc31_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc31_nci_tx_wq#11 irq_context: 0 (wq_completion)nfc31_nci_tx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc31_nci_tx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq#12 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)nfc30_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc30_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc30_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc29_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc29_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc29_nci_rx_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc29_nci_rx_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc29_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc27_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc27_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc27_nci_rx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc27_nci_rx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc27_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc28_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc28_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc28_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc26_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc26_nci_tx_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc25_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc25_nci_cmd_wq#12 irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc25_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc25_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc24_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc24_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc24_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc23_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc23_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc23_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc20_nci_cmd_wq#14 irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&aux->work) rtnl_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc20_nci_rx_wq#14 irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#4 oom_adj_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc20_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc22_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc22_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc22_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc22_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc21_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc21_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc21_nci_rx_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc21_nci_rx_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc21_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc19_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc18_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc18_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc18_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc18_nci_tx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc18_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc17_nci_cmd_wq#12 irq_context: 0 nf_nat_proto_mutex cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc17_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc17_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc16_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc16_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc16_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc15_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc15_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#13 irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&sub_info->work) rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#17 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 pool_lock#2 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#13 irq_context: 0 sb_writers#7 fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_tx_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#20 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#20 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#20 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#20 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#380 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#380 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#372 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#372 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#370 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#50 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#50 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#48 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#48 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#47 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#145 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#145 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#145 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#145 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#143 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#141 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#141 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock quarantine_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#141 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#802 irq_context: 0 &mm->mmap_lock rcu_read_lock pool_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &vma->vm_lock->lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#802 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#802 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#802 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#797 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#797 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#797 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#797 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#794 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#803 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#944 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#128 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#756 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#756 &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &o_tty->termios_rwsem/1 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#757 &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#803 &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 tomoyo_ss quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &c->lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#514 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#514 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#217 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 pcpu_alloc_mutex &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#72/1 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem &p->pi_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5/1 &sb->s_type->i_mutex_key#13 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 rtnl_mutex nf_hook_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#33 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#541 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#539 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#34 irq_context: 0 (wq_completion)gid-cache-wq (work_completion)(&work->work) &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#555 rcu_node_0 irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)events_power_efficient (check_lifetime_work).work fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#803 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc17_nci_tx_wq#10 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#11 irq_context: 0 nfc_devlist_mutex &rcu_state.expedited_wq irq_context: 0 nfc_devlist_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#570 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#101 &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#570 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#570 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#249 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#803 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#798 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#795 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#804 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#804 irq_context: 0 sb_writers#3 remove_cache_srcu rcu_node_0 irq_context: 0 sb_writers#3 remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 sb_writers#3 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock key irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock pcpu_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock percpu_counters_lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &pcp->lock &zone->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &____s->seqcount irq_context: 0 sb_writers#3 remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#799 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#796 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#249 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#571 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal &rq->__lock cpu_asid_lock irq_context: 0 sk_lock-AF_CAN fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#572 irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 devlinks.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &xa->xa_lock#19 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 pcpu_alloc_mutex irq_context: 0 kn->active#4 &kernfs_locks->open_file_mutex[count] remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 pcpu_alloc_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 pcpu_alloc_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &sb->s_type->i_mutex_key#10 unix_gc_lock rlock-AF_UNIX irq_context: 0 &sb->s_type->i_mutex_key#10 unix_gc_lock unix_gc_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex remove_cache_srcu pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 sb_writers#7 kn->active#4 remove_cache_srcu rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex nl_table_wait.lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#4 uevent_sock_mutex nl_table_wait.lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pidmap_lock pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#111 irq_context: 0 (wq_completion)wg-crypt-wg0#111 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 stack_depot_init_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#485 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex pin_fs_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#192 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#192 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#190 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#187 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#193 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1046 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1046 irq_context: 0 sk_lock-AF_NETLINK &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &sb->s_type->i_mutex_key#9 rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &xa->xa_lock#4 irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#246 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#575 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#576 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#98 irq_context: 0 sb_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sb_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#990 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#990 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#979 irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#979 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#972 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy229 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#585 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#945 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#945 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#945 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#938 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: softirq (&p->timer) &br->multicast_lock &c->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex &tbl->lock &____s->seqcount#2 irq_context: 0 rtnl_mutex &tbl->lock &____s->seqcount irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#735 irq_context: 0 &sb->s_type->i_mutex_key#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#593 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#343 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#739 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#741 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#741 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#741 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#754 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#743 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#739 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#746 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#353 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#752 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#354 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#354 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#761 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#761 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#761 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#355 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#348 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &rnp->exp_wq[0] irq_context: 0 sb_writers#5 &type->i_mutex_dir_key#5 smack_known_lock quarantine_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#805 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#800 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#593 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#591 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#263 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#699 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->xattr_sem &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rcu_state.barrier_mutex rcu_state.barrier_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#264 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#99 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#599 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#265 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->id_addr_timer)->work) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &x->wait#2 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#270 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#611 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#735 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#736 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#737 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex net_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &tn->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) &obj_hash[i].lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex gdp_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex subsys mutex#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#343 &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex rcu_state.exp_mutex sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#734 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex devlinks.xa_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#937 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#937 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#937 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex failover_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex net_rwsem &rq->__lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &pipe->mutex/1 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem k-sk_lock-AF_INET6 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#931 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#344 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#736 irq_context: 0 rtnl_mutex team->team_lock_key#116 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#116 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#116 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#116 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 lock irq_context: 0 rtnl_mutex team->team_lock_key#116 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 lock kernfs_idr_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#116 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#116 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#116 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#116 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#112 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#112 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#112 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#112 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#112 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#797 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &ipvlan->addrs_lock krc.lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#806 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#806 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#801 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#798 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#798 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#807 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#807 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#802 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#802 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#802 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#799 irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex bpf_devs_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex bpf_devs_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#118 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#118 &devlink_port->type_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1005 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1005 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1005 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#997 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#991 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1006 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1006 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1006 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1006 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#998 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#992 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#992 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#992 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy214 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy214 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1007 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#999 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#993 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy214 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem triggers_list_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#460 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#460 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#460 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#460 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#460 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#460 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#452 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#449 irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#118 irq_context: 0 rtnl_mutex (inetaddr_chain).rwsem fs_reclaim &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex subsys mutex#20 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex failover_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 &dev->mutex dpm_list_mtx &rq->__lock irq_context: 0 &dev->mutex dpm_list_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx wq_pool_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &sb->s_type->i_mutex_key#3 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex uevent_sock_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#101 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#101 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#101 &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#101 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#101 &devlink_port->type_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX slock-AF_PPPOX irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1008 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1008 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1008 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1008 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1000 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1000 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1000 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#994 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy213 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1009 irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX &pn->hash_lock irq_context: 0 &sb->s_type->i_mutex_key#10 sk_lock-AF_PPPOX clock-AF_PPPOX irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_PPPOX irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1001 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1001 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#995 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#995 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#995 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#453 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#453 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#450 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#211 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#214 irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&devlink->rwork)->work) sched_map-wait-type-override &rq->__lock cpu_asid_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 cb_lock rtnl_mutex &cfs_rq->removed.lock irq_context: 0 rtnl_mutex team->team_lock_key#116 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 lweventlist_lock &dir->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#381 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle sched_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle sched_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem team->team_lock_key#117 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#381 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#373 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#371 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#808 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#811 irq_context: 0 (wq_completion)wg-kex-wg0#205 irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-crypt-wg1#106 irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#818 irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#819 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#822 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#817 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#817 &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->info_timer)->work) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#823 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#824 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#824 &rq->__lock irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &rcu_state.expedited_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#827 irq_context: 0 (wq_completion)nfc18_nci_rx_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#833 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#833 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#842 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#842 irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem &sem->wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#843 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#392 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#392 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#851 irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock stock_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock key irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock pcpu_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock percpu_counters_lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle &sbi->s_orphan_lock pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc14_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#869 irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#879 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#878 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#878 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#879 &rq->__lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 stock_lock irq_context: 0 cb_lock genl_mutex &sb->s_type->i_mutex_key#3 pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#879 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#872 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#882 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#880 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#880 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#882 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#883 irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&rdev->wiphy_work) &rdev->wiphy.mtx rcu_read_lock &sta->rate_ctrl_lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 &type->s_umount_key#29 &rq->__lock irq_context: 0 &type->s_umount_key#29 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#29 &dentry->d_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.expedited_wq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#941 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#941 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#941 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#931 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#439 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#439 &rq->__lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &rcu_state.expedited_wq irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#960 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#439 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#885 irq_context: 0 &mm->mmap_lock remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 misc_mtx cpu_hotplug_lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#55 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#914 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#924 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#924 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#942 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#942 irq_context: 0 misc_mtx nfc_devlist_mutex rcu_read_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#936 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#936 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#935 &cfs_rq->removed.lock irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#931 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#2 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx &root->kernfs_rwsem irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 &sb->s_type->i_mutex_key#10 slock-AF_INET#2 pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dev_pm_qos_sysfs_mtx dev_pm_qos_mtx irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem kernfs_idr_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock kernfs_idr_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->k_lock klist_remove_lock irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex subsys mutex#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock subsys mutex#74 &k->k_lock klist_remove_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)nfc2_nci_rx_wq#938 &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &base->lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 remove_cache_srcu &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#932 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#440 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#440 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#440 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &____s->seqcount#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&forw_packet_aggr->delayed_work)->work) &hard_iface->bat_iv.ogm_buff_mutex &bat_priv->tvlv.container_list_lock &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#440 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#432 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#432 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#432 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &tbl->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock &br->multicast_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 rcu_read_lock &pool->lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#118 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy222 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#462 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#462 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#462 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#462 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#454 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#451 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#451 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 tty_mutex &tty->legacy_mutex batched_entropy_u8.lock irq_context: 0 tty_mutex &tty->legacy_mutex kfence_freelist_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#463 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#463 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#463 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#455 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#452 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#452 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#452 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#456 irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#464 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#429 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#429 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#429 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#946 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#946 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#939 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#933 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#172 irq_context: 0 &ep->mtx &mm->mmap_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#172 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#172 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#172 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#170 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#167 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#167 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#167 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &net->ipv6.addrconf_hash_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#947 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#636 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#947 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#947 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#947 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#940 irq_context: 0 rtnl_mutex wq_pool_mutex &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx wq_pool_mutex wq_pool_mutex.wait_lock irq_context: 0 rtnl_mutex wq_pool_mutex.wait_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#940 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#940 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#934 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#948 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#808 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#808 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#808 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#803 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#800 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#809 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#809 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#809 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#741 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#737 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#735 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#345 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#743 irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 rcu_read_lock rcu_node_0 irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 rcu_read_lock &rq->__lock irq_context: 0 &tty->legacy_mutex &tty->legacy_mutex/1 &tty->ldisc_sem &o_tty->termios_rwsem/1 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex nl_table_lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 &dev->mutex (work_completion)(&rfkill->sync_work) &rq->__lock irq_context: 0 &dev->mutex (work_completion)(&rfkill->sync_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &n->list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &type->lock_class keyring_serialise_link_lock remove_cache_srcu &rq->__lock irq_context: 0 &type->lock_class keyring_serialise_link_lock remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->lock_class keyring_serialise_link_lock remove_cache_srcu pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#745 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#745 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#741 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#746 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#747 irq_context: 0 &p->lock fs_reclaim &cfs_rq->removed.lock irq_context: 0 &p->lock fs_reclaim &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#747 &rq->__lock irq_context: 0 &p->lock fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)wg-kex-wg0#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#212 irq_context: 0 &sig->cred_guard_mutex key#5 irq_context: 0 (wq_completion)wg-kex-wg2#212 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &rcu_state.gp_wq irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)writeback (work_completion)(&(&wb->dwork)->work) &type->s_umount_key#40 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)hci4#5 irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_node_0 irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.gp_wq irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.gp_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &base->lock irq_context: softirq &(&group->avgs_work)->timer rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 key_types_sem root_key_user.cons_lock &n->list_lock irq_context: 0 key_types_sem root_key_user.cons_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 &xt[i].mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#809 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#804 irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: softirq (&ndev->rs_timer) rcu_read_lock rcu_read_lock rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &list->lock#5 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#804 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#179 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#179 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#179 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#179 irq_context: 0 rtnl_mutex uevent_sock_mutex pool_lock irq_context: 0 rtnl_mutex uevent_sock_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#177 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#174 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#174 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#563 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#563 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#563 irq_context: 0 key_types_sem &____s->seqcount#2 irq_context: 0 key_types_sem &____s->seqcount irq_context: 0 key_types_sem init_user_ns.keyring_sem rcu_read_lock rcu_read_lock &rq->__lock irq_context: 0 key_types_sem init_user_ns.keyring_sem rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)events_power_efficient (work_completion)(&(&tbl->gc_work)->work) &tbl->lock quarantine_lock irq_context: 0 (wq_completion)events key_gc_work &type->lock_class remove_cache_srcu irq_context: 0 (wq_completion)events key_gc_work &type->lock_class remove_cache_srcu quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#948 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#948 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#948 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#941 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#941 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#935 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#441 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#441 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#433 irq_context: 0 init_user_ns.keyring_sem &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#433 &rq->__lock irq_context: 0 key_types_sem root_key_user.cons_lock &type->lock_class key_construction_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#433 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#430 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#949 irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &sig->cred_guard_mutex tomoyo_ss mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem stock_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem key irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem pcpu_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem percpu_counters_lock irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#974 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#980 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#560 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#980 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#973 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#950 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#950 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#973 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#952 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex failover_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)nfc2_nci_tx_wq#558 irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex pcpu_alloc_mutex irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#981 irq_context: 0 sb_writers#3 jbd2_handle rcu_read_lock pool_lock#2 irq_context: 0 sb_writers#4 oom_adj_mutex stock_lock irq_context: 0 sb_writers#4 oom_adj_mutex key irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#955 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#173 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#19 irq_context: 0 &xa->xa_lock#15 &____s->seqcount irq_context: 0 &type->s_umount_key#64 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#19 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#19 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#14 irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#974 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#967 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#961 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#34 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#961 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#174 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#174 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#174 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#171 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#976 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#976 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#211 irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex gdp_mutex.wait_lock irq_context: 0 misc_mtx nfc_devlist_mutex gdp_mutex.wait_lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex gdp_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#976 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#969 irq_context: 0 sb_writers#4 oom_adj_mutex pcpu_lock irq_context: 0 sb_writers#4 oom_adj_mutex percpu_counters_lock irq_context: 0 sb_writers#4 oom_adj_mutex pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#174 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#981 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#974 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#450 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#450 &rq->__lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 list_lrus_mutex irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 sb_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 sb_lock unnamed_dev_ida.xa_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem inode_hash_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem fs_reclaim irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem &c->lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem &xa->xa_lock#15 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem &xa->xa_lock#15 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem mmu_notifier_invalidate_range_start irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem inode_hash_lock &s->s_inode_list_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem tk_core.seq.seqcount irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_rwsem &sb->s_type->i_lock_key#24 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &xa->xa_lock#15 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &xa->xa_lock#15 pool_lock#2 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &obj_hash[i].lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 stock_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &sb->s_type->i_lock_key#24 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &sb->s_type->i_lock_key#24 &dentry->d_lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 crngs.lock irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &root->kernfs_supers_rwsem irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &dentry->d_lock irq_context: 0 &type->s_umount_key#61 irq_context: 0 &type->s_umount_key#61 &x->wait#23 irq_context: 0 &type->s_umount_key#61 shrinker_mutex irq_context: 0 &type->s_umount_key#61 &obj_hash[i].lock irq_context: 0 &type->s_umount_key#61 &root->kernfs_supers_rwsem irq_context: 0 &type->s_umount_key#61 rename_lock.seqcount irq_context: 0 &type->s_umount_key#61 &dentry->d_lock irq_context: 0 &type->s_umount_key#61 rcu_read_lock &dentry->d_lock irq_context: 0 &type->s_umount_key#61 &sb->s_type->i_lock_key#24 irq_context: 0 &type->s_umount_key#61 &s->s_inode_list_lock irq_context: 0 &type->s_umount_key#61 &xa->xa_lock#8 irq_context: 0 &type->s_umount_key#61 inode_hash_lock irq_context: 0 &type->s_umount_key#61 inode_hash_lock &sb->s_type->i_lock_key#24 irq_context: 0 &type->s_umount_key#61 pool_lock#2 irq_context: 0 &type->s_umount_key#61 &fsnotify_mark_srcu irq_context: 0 &type->s_umount_key#61 &dentry->d_lock/1 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#450 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock (wq_completion)hci3#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#178 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#178 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)events_unbound (work_completion)(&pool->idle_cull_work) wq_pool_attach_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &fc->uapi_mutex &type->s_umount_key#28/1 &xa->xa_lock#15 &c->lock irq_context: 0 &type->s_umount_key#61 &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#61 &rq->__lock irq_context: 0 &type->s_umount_key#61 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &type->s_umount_key#61 &cfs_rq->removed.lock irq_context: softirq (&app->join_timer)#2 &app->lock#2 batched_entropy_u32.lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#176 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#176 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#176 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#176 &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex dev_pm_qos_sysfs_mtx dev_pm_qos_mtx &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 kn->active#4 batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#96 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#191 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#192 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &rq->__lock cpu_asid_lock irq_context: 0 vlan_ioctl_mutex rtnl_mutex rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#963 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#447 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->orig_work)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#447 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#439 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#436 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#177 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#177 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#175 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#983 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#983 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#976 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#804 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex batched_entropy_u8.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#989 irq_context: 0 rtnl_mutex team->team_lock_key#117 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#993 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#206 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#996 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1001 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1001 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1001 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1001 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#994 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#988 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#988 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#988 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#458 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1002 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1003 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1003 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#465 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#465 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#457 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#457 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#215 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#457 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#453 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#183 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#183 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#183 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#183 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#181 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#181 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#181 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#178 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1010 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1002 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1047 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#487 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#703 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#703 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#703 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#324 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#324 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#698 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#318 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#316 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#704 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#704 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#701 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#488 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#488 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#488 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#699 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#705 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#705 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#697 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#702 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#702 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#700 irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#44 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#706 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#706 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#124 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#124 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#122 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#120 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#488 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#325 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#325 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#319 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#319 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#319 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#317 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#707 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#704 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#702 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#326 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#326 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#708 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#705 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#705 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#705 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#320 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#703 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#318 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#125 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#125 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#125 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#125 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#123 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#123 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#123 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#121 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#121 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#121 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#45 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#43 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#42 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#709 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#709 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#706 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#704 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#321 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#319 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#17 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#710 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#707 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#189 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#705 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#328 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1049 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->tt.work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1050 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1050 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1050 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#328 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#322 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#320 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#172 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#62 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#62 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#711 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#711 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#708 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#323 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#321 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#712 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#712 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#712 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#712 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#709 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#709 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#709 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#707 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#126 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#126 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#126 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#126 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#124 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#124 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#124 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#122 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#122 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#122 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#713 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#713 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill_global_led_trigger_work) rfkill_global_mutex sched_map-wait-type-override &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#46 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#46 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#46 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#46 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#44 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#43 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#43 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#43 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#714 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#714 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#714 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#714 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#710 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#710 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#710 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#708 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#708 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#322 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#715 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#711 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#709 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#332 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#332 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#325 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#325 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#325 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#323 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#323 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#716 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#716 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#712 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#712 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#712 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#710 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#127 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#127 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#125 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#125 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#125 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#123 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#123 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#717 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#717 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#713 irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &tn->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1041 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#490 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#490 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#195 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#195 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#479 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1053 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1053 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1054 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#197 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1055 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1045 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#22 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1039 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->nc.work)->work) rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1057 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1058 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1058 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#180 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#179 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 fh->state->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci3#5 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &n->list_lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->tx_work) &list->lock#5 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci3#6 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#6 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci3#6 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#466 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#466 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#466 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#464 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#69 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#69 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#67 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#66 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#470 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#470 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#467 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#467 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#465 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#471 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#471 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#471 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#468 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#468 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#468 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#466 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#466 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#466 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#185 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#185 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#181 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#181 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#181 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#181 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#180 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#472 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#472 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#472 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#186 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#187 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#188 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#188 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#188 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#67 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#476 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#477 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#478 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#478 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#480 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#482 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#479 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#477 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#193 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#193 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#189 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#188 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#483 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#484 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#484 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#195 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#485 irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#108 irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#191 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#192 irq_context: 0 sk_lock-AF_INET fs_reclaim rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#216 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)nfc3_nci_rx_wq#192 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#191 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#486 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#487 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#487 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#484 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#198 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#198 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#482 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#194 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#193 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#73 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#488 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#485 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#490 irq_context: 0 &type->s_umount_key#61 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex (inetaddr_chain).rwsem rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &type->s_umount_key#40 &rq->__lock irq_context: 0 &type->s_umount_key#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#494 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#494 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1049 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#245 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#96 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#96 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#97 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#252 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#990 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#114 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#990 irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#982 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#490 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#490 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#949 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#200 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#949 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#949 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#942 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#942 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#942 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#936 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1042 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#950 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#950 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#943 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#937 irq_context: 0 &hdev->req_lock (wq_completion)hci0#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#951 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#951 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#951 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#951 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#944 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#938 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#938 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#938 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#7 irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &list->lock#5 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc37_nci_cmd_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#241 irq_context: 0 (wq_completion)nfc37_nci_cmd_wq#4 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 wq_pool_mutex rcu_read_lock &rq->__lock irq_context: 0 wq_pool_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#495 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#496 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#496 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#496 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#496 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#488 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#488 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#488 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) rcu_node_0 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex fill_pool_map-wait-type-override &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1061 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock deferred_probe_mutex irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock device_links_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->cmd_work) &rq->__lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1061 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1061 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci1#5 irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle remove_cache_srcu &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#6 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) &list->lock#5 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &n->list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci0#7 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#503 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#503 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#505 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#506 irq_context: 0 pernet_ops_rwsem uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#508 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#508 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#509 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#510 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#510 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#511 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#205 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#81 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#81 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#210 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#33 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#32 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#545 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#545 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#542 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#510 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#79 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#540 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#517 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#517 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#514 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#540 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#518 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#71 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#499 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy227 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#546 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#214 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#546 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#62 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#499 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#500 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#500 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#500 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1063 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1064 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1064 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#501 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#203 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#200 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#200 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1049 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1049 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1049 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1068 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#504 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#496 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#505 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1055 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#218 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#227 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#225 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#550 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#562 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1162 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1162 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1175 irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1177 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1177 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1177 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#584 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#584 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#572 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1182 irq_context: 0 misc_mtx cpu_hotplug_lock &cfs_rq->removed.lock irq_context: 0 misc_mtx cpu_hotplug_lock &obj_hash[i].lock irq_context: 0 misc_mtx cpu_hotplug_lock pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#586 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#586 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1187 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1188 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1188 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#590 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1195 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#3 fs_reclaim &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1196 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1196 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1197 irq_context: 0 nfc_devlist_mutex rcu_read_lock &rq->__lock irq_context: 0 nfc_devlist_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1174 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#250 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#247 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#244 irq_context: 0 rtnl_mutex &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#244 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1198 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#596 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#597 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#598 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1201 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1201 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1189 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1189 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1189 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1202 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#601 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#602 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#602 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1204 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1204 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1204 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#246 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1206 irq_context: 0 sb_writers &type->i_mutex_dir_key#2 key irq_context: 0 sb_writers &type->i_mutex_dir_key#2 pcpu_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 percpu_counters_lock irq_context: 0 sb_writers &type->i_mutex_dir_key#2 pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#23 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#23 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#23 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#21 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#21 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#21 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_rx_wq#21 irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc34_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#19 &rq->__lock irq_context: 0 (wq_completion)nfc19_nci_rx_wq#20 irq_context: 0 tty_mutex remove_cache_srcu &rq->__lock irq_context: 0 tty_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc25_nci_rx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc15_nci_tx_wq#17 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#20 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)nfc9_nci_tx_wq#22 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#23 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#24 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1208 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1208 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#92 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1210 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#588 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#588 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#255 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1188 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1213 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1213 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1214 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1214 &rq->__lock irq_context: 0 sb_writers#4 &mm->mmap_lock &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#594 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#594 &rq->__lock irq_context: 0 kn->active#18 remove_cache_srcu irq_context: 0 kn->active#18 remove_cache_srcu quarantine_lock irq_context: 0 kn->active#18 remove_cache_srcu &c->lock irq_context: 0 kn->active#18 remove_cache_srcu &n->list_lock irq_context: 0 kn->active#18 remove_cache_srcu &obj_hash[i].lock irq_context: 0 kn->active#18 remove_cache_srcu &pcp->lock &zone->lock irq_context: 0 kn->active#18 remove_cache_srcu &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1215 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1191 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#253 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#253 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#250 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#608 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#257 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1217 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1217 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1217 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#609 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#609 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#609 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#609 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#256 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#260 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#260 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#261 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#257 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1221 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#615 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1223 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1223 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1223 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#616 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#263 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#259 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1224 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1200 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#264 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1228 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1232 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1232 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#399 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#399 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#399 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#399 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#400 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#400 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#400 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex gdp_mutex key irq_context: 0 rtnl_mutex gdp_mutex pcpu_lock irq_context: 0 rtnl_mutex gdp_mutex percpu_counters_lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->lock#2 irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &base->lock irq_context: 0 &ndev->req_lock &wq->mutex rcu_node_0 irq_context: 0 &ndev->req_lock &wq->mutex &rcu_state.expedited_wq irq_context: 0 &ndev->req_lock &wq->mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ndev->req_lock &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock &wq->mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock &wq->mutex &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#140 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#402 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#403 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#403 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#404 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#404 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#406 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#406 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#401 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#407 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#407 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#408 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#144 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#144 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#144 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#47 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#215 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#519 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#517 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#217 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#213 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#524 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#525 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#525 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#219 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#33 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#220 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#220 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#81 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#221 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#529 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#529 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#530 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#530 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#543 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#223 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#543 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#543 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#541 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#534 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#541 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#541 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#234 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#234 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#230 irq_context: 0 rcu_read_lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#229 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#411 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#411 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#411 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#411 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#408 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#406 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#229 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#547 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#535 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#547 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#547 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#235 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#235 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#231 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#231 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#231 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#547 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#544 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#544 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#544 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#230 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#542 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#91 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#91 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#91 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#91 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#89 irq_context: 0 &dev->mutex &root->kernfs_rwsem &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#548 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#545 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#543 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#549 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#236 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#236 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#551 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#237 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#552 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#552 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#552 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#549 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#238 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#553 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#233 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#554 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#554 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#556 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#556 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#556 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#557 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#62 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#560 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#560 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#561 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#562 irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#453 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#453 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#453 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex dev_addr_sem remove_cache_srcu &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#442 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#989 rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#989 &rcu_state.expedited_wq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#989 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#146 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#146 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#142 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#141 irq_context: 0 &mm->mmap_lock &mm->mmap_lock/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#50 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#412 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#147 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#983 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#977 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#983 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#117 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#117 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#117 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#992 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#992 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#985 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#147 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#147 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#413 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#414 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#979 irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#448 irq_context: 0 &rs->rs_rdma_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#98 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#98 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#572 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#983 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#977 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#977 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#454 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#454 &rq->__lock irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#110 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#578 irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#100 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#100 irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &mm->mmap_lock fs_reclaim &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#576 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#574 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#416 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#416 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#416 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#417 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#255 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#580 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock remove_cache_srcu pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#581 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#581 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#252 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#580 &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#418 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#418 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#585 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#586 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#587 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#587 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#584 irq_context: 0 rtnl_mutex &hard_iface->bat_iv.ogm_buff_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#996 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#589 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#589 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#586 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#584 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#261 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#257 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#591 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#718 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#150 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#419 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#151 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#421 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#421 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#419 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#417 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#423 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#423 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#420 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#424 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#425 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#425 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#424 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#428 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#148 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#429 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#596 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#429 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#595 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#430 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#155 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#155 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#151 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#432 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#433 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#435 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#435 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#435 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#432 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#157 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#157 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#801 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#810 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#810 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#810 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override rcu_read_lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#436 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#436 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#437 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 fill_pool_map-wait-type-override rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#116 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#116 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#116 console_lock console_srcu console_owner irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#437 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#116 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#116 console_lock console_srcu console_owner console_owner_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#622 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#624 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#281 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#626 irq_context: 0 (wq_completion)bond0#116 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#116 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#438 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#438 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#435 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#56 irq_context: 0 (wq_completion)bond0#116 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#382 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#382 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#382 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#276 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#160 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#160 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#439 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#441 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#442 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#29 irq_context: 0 rtnl_mutex rcu_read_lock &pcp->lock &zone->lock irq_context: 0 rtnl_mutex rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#382 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#374 irq_context: 0 &type->i_mutex_dir_key#4 &root->kernfs_rwsem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#444 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#444 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#374 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#810 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#805 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#372 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#372 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#372 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#811 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#806 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#806 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#812 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#813 irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &ndev->req_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#813 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#808 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#814 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#439 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#815 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#815 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#815 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#816 irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock &obj_hash[i].lock irq_context: 0 &dev->mutex uevent_sock_mutex rcu_read_lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#718 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#714 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#714 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#714 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#712 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#333 irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#326 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#324 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#816 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#637 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#638 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#638 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#635 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#639 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#637 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#641 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#641 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#638 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#639 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#643 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#644 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#163 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#159 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#165 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#447 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#447 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#444 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#445 irq_context: 0 &type->i_mutex_dir_key#4 &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#644 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#644 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#644 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#645 irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#646 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#60 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#60 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#58 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#57 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &cfs_rq->removed.lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &obj_hash[i].lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#168 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#168 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#61 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#648 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#648 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#816 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#817 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#817 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#203 irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#648 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#645 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#291 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#106 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#650 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#650 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sb->s_type->i_mutex_key#3 &sem->wait_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &sem->wait_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &p->pi_lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &p->pi_lock &rq->__lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->debugfs_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#651 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#652 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#653 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#653 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#653 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#288 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#297 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#656 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#656 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#651 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#654 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#658 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#658 irq_context: 0 pernet_ops_rwsem rtnl_mutex smc_ib_devices.mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem k-sk_lock-AF_INET6 &tcp_hashinfo.bhash[i].lock &tcp_hashinfo.bhash2[i].lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#655 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#653 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#293 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#300 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#300 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#300 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#300 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#294 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#658 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#662 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#662 &rq->__lock irq_context: 0 rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 rcu_read_lock &____s->seqcount#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#662 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#665 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#665 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#662 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#660 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#302 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#302 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#302 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#302 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#667 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#667 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#668 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#668 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#665 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#669 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#669 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#666 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#295 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#670 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#671 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#672 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#673 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#673 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#668 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#296 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#676 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#114 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#678 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#678 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#679 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#680 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#680 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#301 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#681 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#681 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#681 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#676 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#682 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#682 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#683 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#684 &rq->__lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci2 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci2#2 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#684 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#111 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#686 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#686 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#683 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#116 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#688 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#688 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#685 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#684 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#458 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#450 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#450 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#450 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#447 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#181 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#181 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#181 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#181 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#179 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#179 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#179 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#176 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1002 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1002 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1002 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#995 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#989 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#989 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#989 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex &root->kernfs_rwsem rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1003 &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1003 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#468 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#460 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#460 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#460 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#456 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#456 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#456 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1011 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1011 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1011 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1011 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1003 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1003 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1003 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#997 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#469 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#469 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#461 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#260 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#260 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#603 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#603 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#267 irq_context: 0 (wq_completion)events_unbound connector_reaper_work &meta->lock irq_context: 0 (wq_completion)events_unbound connector_reaper_work kfence_freelist_lock irq_context: 0 kn->active#14 &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock rcu_node_0 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) &list->lock#5 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci0#8 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&conn->pending_rx_work) &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&conn->pending_rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#4 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock &rq->__lock irq_context: 0 pernet_ops_rwsem rtnl_mutex &pnettable->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#952 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#945 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#939 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#939 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#939 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#267 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 (wq_completion)nfc3_nci_tx_wq#457 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#457 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#457 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1012 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1012 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1004 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1004 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1004 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1004 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#998 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#470 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#470 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#470 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#470 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#462 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#458 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1013 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1013 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1005 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1005 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &idev->mc_lock _xmit_ETHER &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc18_nci_tx_wq#16 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex smc_ib_devices.mutex irq_context: 0 (wq_completion)nfc18_nci_tx_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc18_nci_tx_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &p->pi_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#18 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &root->kernfs_rwsem &sem->wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &root->kernfs_rwsem &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#15 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc19_nci_rx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc20_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc20_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc20_nci_tx_wq#16 irq_context: 0 (wq_completion)nfc20_nci_tx_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc20_nci_tx_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc21_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc21_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc21_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc21_nci_tx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc21_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc22_nci_cmd_wq#15 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc22_nci_cmd_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc22_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc22_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc22_nci_rx_wq#15 &rq->__lock irq_context: 0 sb_writers#3 sb_internal jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc22_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#17 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc23_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc23_nci_rx_wq#17 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)nfc23_nci_rx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc23_nci_rx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy226 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock fs_reclaim &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc23_nci_tx_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc24_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc24_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc24_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc24_nci_tx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc24_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc25_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc25_nci_cmd_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc25_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc25_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc25_nci_tx_wq#14 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc27_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc27_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc27_nci_rx_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc27_nci_rx_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc27_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc28_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc28_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc28_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc29_nci_cmd_wq#16 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex uevent_sock_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex uevent_sock_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex remove_cache_srcu &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc29_nci_rx_wq#16 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex remove_cache_srcu pool_lock#2 irq_context: 0 (wq_completion)nfc29_nci_tx_wq#16 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#115 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#115 &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#115 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc42_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc42_nci_rx_wq#6 irq_context: 0 (wq_completion)nfc42_nci_rx_wq#6 &rq->__lock irq_context: 0 (wq_completion)nfc42_nci_rx_wq#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc42_nci_tx_wq#6 irq_context: 0 (wq_completion)nfc42_nci_tx_wq#6 &rq->__lock irq_context: 0 (wq_completion)nfc42_nci_tx_wq#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#445 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#445 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#437 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#434 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#957 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#957 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#957 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#957 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#957 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#950 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#944 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#28 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#26 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#26 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#26 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#26 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#26 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#26 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#21 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#21 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#21 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#21 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc44_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc44_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc44_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc44_nci_rx_wq#4 &rq->__lock irq_context: 0 (wq_completion)nfc44_nci_rx_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#175 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#175 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#173 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy226 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#170 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#61 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#61 irq_context: 0 rtnl_mutex devnet_rename_sem fs_reclaim &rq->__lock irq_context: 0 rtnl_mutex devnet_rename_sem fs_reclaim &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#59 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#58 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#20 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#20 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#20 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#20 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#17 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#21 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#21 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_rx_wq#21 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock batched_entropy_u8.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock kfence_freelist_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock &n->list_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock crngs.lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock &nsim_dev->fa_cookie_lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock &obj_hash[i].lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 (wq_completion)nfc11_nci_tx_wq#21 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1005 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc12_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc15_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc15_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#18 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc16_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc16_nci_rx_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#999 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#471 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#471 &rq->__lock irq_context: 0 (wq_completion)nfc16_nci_rx_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc16_nci_tx_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc17_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc17_nci_cmd_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc17_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc17_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc17_nci_tx_wq#15 irq_context: 0 (wq_completion)nfc17_nci_tx_wq#15 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#19 &rq->__lock irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc18_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc18_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc18_nci_tx_wq#17 irq_context: 0 (wq_completion)nfc18_nci_tx_wq#17 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#471 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#471 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#463 irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc19_nci_tx_wq#16 irq_context: 0 (wq_completion)nfc19_nci_tx_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc20_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc20_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc20_nci_tx_wq#17 irq_context: 0 (wq_completion)nfc20_nci_tx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc20_nci_tx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc21_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#459 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#459 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#459 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#184 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#606 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#606 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#607 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#607 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#605 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#603 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#609 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#609 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#206 irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#184 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#184 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#184 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#182 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 kn->active#46 &rq->__lock irq_context: 0 kn->active#46 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#203 irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#204 irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#802 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#806 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#803 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#812 irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_ifc_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh rcu_read_lock_bh fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 &mm->mmap_lock rcu_read_lock fill_pool_map-wait-type-override &pcp->lock &zone->lock irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: softirq rcu_read_lock &local->rx_path_lock rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#201 irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#201 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 rtnl_mutex &wg->device_update_lock cpu_hotplug_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#199 irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#202 irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex fs_reclaim irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx pcpu_alloc_mutex &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#606 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#612 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#612 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#612 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#612 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#266 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1014 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &dentry->d_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#202 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#200 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#185 irq_context: 0 rtnl_mutex devnet_rename_sem uevent_sock_mutex quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)events wireless_nlevent_work net_rwsem rcu_node_0 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock uevent_sock_mutex rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#615 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#615 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#274 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#617 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#199 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#204 irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)nfc2_nci_tx_wq#612 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#275 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#276 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#102 irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#807 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#804 irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc2_nci_tx_wq#805 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#814 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#809 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#806 irq_context: 0 (wq_completion)wg-kex-wg2#199 irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#199 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#815 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#200 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#810 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#807 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#807 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)nfc2_nci_tx_wq#807 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#816 irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#200 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex dev_addr_sem rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock uevent_sock_mutex &____s->seqcount irq_context: 0 (wq_completion)nfc21_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc21_nci_tx_wq#15 irq_context: 0 (wq_completion)nfc21_nci_tx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc45_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc45_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc45_nci_tx_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc43_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc43_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc43_nci_rx_wq#5 irq_context: 0 (wq_completion)nfc43_nci_tx_wq#5 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy225 irq_context: 0 &ndev->req_lock (wq_completion)nfc41_nci_cmd_wq#6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy225 &rq->__lock irq_context: 0 (wq_completion)nfc41_nci_cmd_wq#6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy225 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc41_nci_rx_wq#6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy225 &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy225 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy225 pool_lock#2 irq_context: 0 (wq_completion)nfc41_nci_tx_wq#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc40_nci_cmd_wq#8 irq_context: 0 (wq_completion)nfc40_nci_cmd_wq#8 irq_context: 0 (wq_completion)nfc40_nci_rx_wq#8 irq_context: 0 (wq_completion)nfc40_nci_tx_wq#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc35_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc35_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc35_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc35_nci_tx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc35_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 nfc_devlist_mutex remove_cache_srcu &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc39_nci_cmd_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc39_nci_cmd_wq#7 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc39_nci_cmd_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc39_nci_cmd_wq#7 irq_context: 0 (wq_completion)nfc39_nci_rx_wq#7 irq_context: 0 (wq_completion)nfc39_nci_tx_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc38_nci_cmd_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc38_nci_cmd_wq#7 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc38_nci_cmd_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc38_nci_cmd_wq#7 irq_context: 0 (wq_completion)nfc38_nci_rx_wq#7 irq_context: 0 (wq_completion)nfc38_nci_tx_wq#7 irq_context: 0 (wq_completion)nfc38_nci_tx_wq#7 &rq->__lock irq_context: 0 (wq_completion)nfc38_nci_tx_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc37_nci_cmd_wq#8 irq_context: 0 (wq_completion)nfc37_nci_cmd_wq#8 irq_context: 0 (wq_completion)nfc37_nci_rx_wq#8 irq_context: 0 (wq_completion)nfc37_nci_tx_wq#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc36_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc36_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc36_nci_rx_wq#8 irq_context: 0 (wq_completion)nfc36_nci_rx_wq#8 &rq->__lock irq_context: 0 (wq_completion)nfc36_nci_tx_wq#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc34_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc34_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc34_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc34_nci_tx_wq#13 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem uevent_sock_mutex &pcp->lock &zone->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc33_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc33_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc33_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc32_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc32_nci_rx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc31_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc31_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc31_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc30_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc30_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc30_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc26_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#20 irq_context: 0 (wq_completion)nfc26_nci_tx_wq#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc22_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc22_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc22_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc22_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#958 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#958 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#951 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#951 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#945 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#959 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#952 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#946 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#946 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#946 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock kfence_freelist_lock irq_context: 0 &hdev->req_lock &hdev->lock &meta->lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci1#8 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &list->lock#5 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#960 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#960 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#953 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#947 irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->cmd_work) &rq->__lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#102 irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#102 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#818 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#818 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#818 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#813 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#810 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#819 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#819 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#819 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#814 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#820 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#820 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#820 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &dev->mutex uevent_sock_mutex mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 &dev->mutex uevent_sock_mutex mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#823 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#824 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#824 irq_context: 0 (wq_completion)ipv6_addrconf (work_completion)(&(&ifa->dad_work)->work) rtnl_mutex &idev->mc_lock rcu_read_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex rcu_read_lock &net->sctp.local_addr_lock &net->sctp.addr_wq_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#827 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#828 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#829 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#830 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#831 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#832 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#832 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#828 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#825 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#826 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#384 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#384 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#826 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#384 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#827 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#836 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#836 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#837 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#837 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#833 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#830 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#839 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#840 irq_context: 0 tomoyo_ss pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#841 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#841 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#841 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#386 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#837 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#843 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#843 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#843 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#838 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#835 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#388 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#845 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#845 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#840 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#837 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#837 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#144 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#390 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#390 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#847 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#847 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#847 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#847 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#848 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#848 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#848 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#849 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#850 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#850 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#851 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#393 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#852 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#852 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#852 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#847 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#853 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#854 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#854 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#849 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#846 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#855 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#855 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#149 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#856 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#856 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#851 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#848 irq_context: 0 &fc->uapi_mutex &type->s_umount_key#67/1 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#396 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#396 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#396 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#396 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#388 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#849 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#15 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rcu_state.expedited_wq irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock irq_context: 0 nfc_devlist_mutex &root->kernfs_rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#15 &rq->__lock irq_context: 0 tty_mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 tty_mutex fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 (wq_completion)nfc29_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc29_nci_tx_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#12 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc36_nci_cmd_wq#7 irq_context: 0 (wq_completion)nfc35_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#858 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#853 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#853 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#853 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#850 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#51 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc20_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc39_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#155 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#856 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#861 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#862 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#862 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#863 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#863 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#864 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#865 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#390 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#866 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#866 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#401 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#156 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#858 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#868 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#863 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#869 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#869 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#394 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#870 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#870 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#861 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#158 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#871 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#871 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#872 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#867 irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#875 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#875 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#406 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#406 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#864 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#407 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#407 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#878 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#878 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#871 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#866 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#866 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#866 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#867 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#880 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#880 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#873 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#868 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#408 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#408 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#881 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#881 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#409 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#410 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#883 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#883 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#883 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#872 irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#183 irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#4 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) quarantine_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu &c->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#878 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#886 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#886 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#886 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#886 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#879 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#874 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#887 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#887 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#887 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#887 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#888 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#888 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#881 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#876 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#889 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#889 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#882 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#877 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#890 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#890 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#883 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#891 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#891 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#891 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#891 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#413 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#413 &rq->__lock irq_context: 0 &mm->mmap_lock &folio_wait_table[i] &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#414 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#414 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#414 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#892 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#892 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#892 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#886 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#886 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#886 &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#886 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#887 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#895 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#895 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#895 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#896 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#415 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#415 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#415 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#885 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#885 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#885 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#891 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#891 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#891 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#898 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#898 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#898 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#899 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#899 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#899 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#900 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#416 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#416 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#901 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#902 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#417 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#417 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#409 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#904 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#904 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#897 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#891 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#905 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#906 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#906 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#906 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#899 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#418 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#418 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_rx_wq#160 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#163 irq_context: 0 &dev->mutex &base->lock irq_context: 0 &dev->mutex &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#420 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#420 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#420 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#412 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#421 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#421 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#421 irq_context: 0 (crypto_chain).rwsem remove_cache_srcu irq_context: 0 (crypto_chain).rwsem remove_cache_srcu quarantine_lock irq_context: 0 (crypto_chain).rwsem remove_cache_srcu &c->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#412 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#910 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#910 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#910 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#910 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#903 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#55 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#904 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#425 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#912 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#905 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#913 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#913 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#906 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#900 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#900 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#900 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#426 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss remove_cache_srcu fill_pool_map-wait-type-override &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#914 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#914 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#914 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#907 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#907 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#901 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#427 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#427 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#419 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#416 irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &obj_hash[i].lock pool_lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &rq->__lock irq_context: 0 &type->s_umount_key#77 &sn->pipefs_sb_lock (rpc_pipefs_notifier_list).rwsem &sb->s_type->i_mutex_key#20/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#902 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#902 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#902 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#908 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#915 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#915 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#915 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#916 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#916 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#916 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#916 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#909 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#428 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#428 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#428 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#428 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#917 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#917 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#918 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#918 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#918 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc6_nci_cmd_wq#24 irq_context: 0 misc_mtx (wq_completion)nfc6_nci_cmd_wq#24 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc6_nci_cmd_wq#24 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#430 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#430 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#919 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#920 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#920 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#920 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#920 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#921 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#921 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#914 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#922 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#922 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#915 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#909 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#909 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#923 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#923 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#923 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#923 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#916 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#910 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#924 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#924 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#917 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#917 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#917 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#911 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#925 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#925 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#925 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#169 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#432 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#432 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#432 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#927 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#927 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#927 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#928 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#928 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#915 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#929 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#929 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#929 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#929 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#930 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#930 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#931 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#931 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#435 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#427 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#424 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#932 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#925 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#933 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#933 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#926 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#920 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#934 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#101 &nsim_trap_data->trap_lock batched_entropy_u8.lock crngs.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#935 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#935 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#935 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->tx_work) &list->lock#5 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci1#6 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex cpu_hotplug_lock wq_pool_mutex.wait_lock irq_context: 0 rtnl_mutex cpu_hotplug_lock &p->pi_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#203 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#204 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#102 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)bond0#113 irq_context: 0 (wq_completion)bond0#113 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#113 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#113 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#113 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#208 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#207 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#107 irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#60 irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#59 irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &rcu_state.expedited_wq irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 misc_mtx nfc_devlist_mutex remove_cache_srcu &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#977 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#977 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#89 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#977 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#977 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#970 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#964 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#964 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#964 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#971 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#965 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#448 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#448 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#448 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#440 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#437 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &meta->lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rdev->wiphy.mtx (wq_completion)phy230 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#966 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#449 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#449 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#207 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#205 irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#208 irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#206 irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#106 irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#205 irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu quarantine_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 rtnl_mutex &root->kernfs_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 irq_context: 0 (wq_completion)wg-crypt-wg2#103 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)bond0#113 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#113 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#113 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#113 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#113 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &c->lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex rcu_read_lock &pool->lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#103 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx &sb->s_type->i_mutex_key#3 remove_cache_srcu &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#183 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#183 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#180 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#29 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#996 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#989 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#29 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#29 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#29 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#27 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#27 &rq->__lock irq_context: 0 sb_writers#8 &of->mutex cgroup_mutex cpu_hotplug_lock cgroup_threadgroup_rwsem css_set_lock cgroup_file_kn_lock kernfs_notify_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#27 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#953 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#953 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#946 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#940 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#442 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#997 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#63 &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#63 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#63 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#61 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#60 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#60 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#60 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) rcu_node_0 irq_context: 0 &xt[i].mutex batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &base->lock irq_context: 0 (wq_completion)mld (work_completion)(&(&idev->mc_dad_work)->work) &idev->mc_lock rcu_read_lock rcu_read_lock rcu_read_lock_bh dev->qdisc_tx_busylock ?: &qdisc_tx_busylock _xmit_ETHER#2 &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#89 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-kex-wg2#215 irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#442 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#442 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#442 irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1004 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1004 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1004 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#182 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#459 irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#434 irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 misc_mtx rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 misc_mtx rfkill_global_mutex fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#93 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock base_crng.lock irq_context: 0 (wq_completion)wg-kex-wg2#185 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1071 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1055 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1073 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1073 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1063 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1074 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1074 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#509 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#501 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#502 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#502 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1075 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1075 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1077 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1078 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1078 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#512 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#504 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1079 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1079 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1080 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1080 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1080 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1080 &rq->__lock irq_context: 0 &type->s_umount_key#64 pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#515 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#515 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#515 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#515 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#70 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1083 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1083 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1084 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#516 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#516 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#508 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1076 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1087 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1087 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1088 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1088 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#518 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1089 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1089 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1091 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1091 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1092 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1092 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#521 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#521 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1094 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1094 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1094 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#523 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#523 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1077 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#215 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1097 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1098 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1088 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1079 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1099 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#527 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1101 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1101 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1102 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1102 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1102 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1092 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1103 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1103 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1085 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1085 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#531 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#531 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#216 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#213 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#519 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1109 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1110 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1110 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1110 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1091 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#533 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#533 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#215 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#534 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1112 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1112 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#522 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1103 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1095 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#536 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#536 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#222 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#222 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#220 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1117 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1117 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1098 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1098 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#538 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#530 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1119 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1119 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#531 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#540 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#226 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#224 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1121 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1121 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#533 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#528 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#543 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#434 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#543 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#543 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#35 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1124 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1124 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1114 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#536 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1127 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1128 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1128 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1128 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#546 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#228 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#228 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#228 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#80 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#548 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#549 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#549 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#549 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1131 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1132 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1133 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#550 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#541 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1135 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1136 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1136 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1126 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#552 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#227 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1138 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1138 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1140 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1141 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1131 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1121 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1142 irq_context: 0 misc_mtx &wq->mutex &rq->__lock irq_context: 0 misc_mtx &wq->mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1122 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1144 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1144 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1145 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1145 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#226 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#82 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1147 irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 &u->iolock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#543 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#543 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1149 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1151 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1151 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1151 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1151 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1140 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#550 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#550 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#545 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1152 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#560 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#560 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#551 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#561 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#235 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1155 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1155 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1155 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#564 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#564 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1156 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#565 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1146 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1158 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1158 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#238 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#86 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#567 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#567 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1160 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#87 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#240 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#240 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1162 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1162 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1163 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#569 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#569 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#570 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#570 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#555 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1153 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#84 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#235 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1165 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1165 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1165 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1165 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#571 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1166 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1166 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#562 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#562 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1167 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1168 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1168 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_tx_wq#559 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#431 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#431 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#431 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 crypto_alg_sem &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1169 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1171 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1171 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1171 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1171 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1159 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1172 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1172 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#578 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#578 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1174 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1175 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#205 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock kfence_freelist_lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1176 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1176 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#581 irq_context: 0 (wq_completion)hci1#7 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#115 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1178 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1178 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1166 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1155 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#246 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#246 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#246 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#583 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#566 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#566 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#566 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1180 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1182 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1182 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#585 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1171 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1184 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#955 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#173 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#173 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1176 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1176 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1165 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1189 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#591 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#591 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#173 irq_context: 0 (wq_completion)nfc26_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#578 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#591 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#591 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#579 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1193 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#19 &rq->__lock irq_context: 0 (wq_completion)nfc26_nci_rx_wq#19 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc26_nci_tx_wq#19 irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 sb_internal jbd2_handle &sbi->s_orphan_lock rcu_read_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 &root->deactivate_waitq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc35_nci_cmd_wq#14 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &sig->cred_guard_mutex &sig->exec_update_lock &mm->mmap_lock &anon_vma->rwsem &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &n->list_lock irq_context: 0 wq_pool_mutex &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1194 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1194 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1194 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy227 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#87 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#548 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#548 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#548 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#549 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#546 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#546 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#546 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#544 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#544 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#544 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#236 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#236 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#232 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#232 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#232 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#550 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#550 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#231 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#547 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#547 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#547 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#545 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#551 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#548 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#546 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#237 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#237 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#233 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#232 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#232 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#549 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#547 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#238 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#238 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#553 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#553 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#553 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#550 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#548 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#548 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#548 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#27 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#238 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#234 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#551 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#549 irq_context: softirq (&in_dev->mr_ifc_timer) rcu_read_lock &in_dev->mc_tomb_lock quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#555 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#555 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#555 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#555 &rcu_state.expedited_wq irq_context: 0 (wq_completion)nfc9_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#19 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#19 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#555 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#555 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#552 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#550 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 pin_fs_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 sk_lock-AF_PACKET rcu_state.exp_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &rnp->exp_wq[2] irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#556 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#553 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#551 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#239 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#239 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#235 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#234 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#557 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#557 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#557 irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#554 irq_context: 0 (wq_completion)nfc37_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc37_nci_tx_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc36_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc36_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc36_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc36_nci_tx_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc34_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc34_nci_tx_wq#9 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#552 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#610 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#611 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#598 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#598 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1218 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#598 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#612 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#612 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#599 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#264 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1227 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1227 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1214 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#606 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1205 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1231 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1231 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#621 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1233 irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#206 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#394 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#394 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#139 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#139 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#135 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#134 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#400 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#397 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#395 irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->timeout_work)->work) irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#2 irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#401 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#401 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#398 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#396 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#140 irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci1#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#136 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#135 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#402 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#399 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#397 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#397 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#397 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#403 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#403 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#403 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#403 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#400 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#400 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#400 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#398 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#401 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#399 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#141 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#141 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#137 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#136 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#405 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#405 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#402 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#400 irq_context: 0 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#406 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#406 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#403 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#403 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#403 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#404 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#142 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#402 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#142 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#142 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#142 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#138 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#137 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#46 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#46 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#45 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#45 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#408 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#408 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#408 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#405 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#403 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#143 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#143 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#139 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#138 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#144 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#144 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#144 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#140 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#139 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#139 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#139 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#409 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#409 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#409 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#409 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#406 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#404 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#47 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#47 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#47 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#47 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#46 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#46 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#46 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#48 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#49 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#49 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#49 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#145 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#145 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#145 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#141 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#140 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#410 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#410 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#407 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#405 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#405 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#50 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#48 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#47 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#412 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#409 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#407 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#147 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#143 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#142 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#413 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#410 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#51 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#51 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#408 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#49 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#48 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#48 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#48 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#414 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#414 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#414 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#411 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#409 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#415 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#415 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#412 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#410 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#148 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#410 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#148 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#410 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#144 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#143 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#416 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#413 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#411 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#149 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#149 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#145 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#144 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#52 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#52 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#50 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#49 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#417 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#417 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#417 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#414 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#412 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#418 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#418 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#415 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#413 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#150 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#146 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#145 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#419 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#419 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#419 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#416 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#414 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#151 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#151 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#147 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#146 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#420 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#420 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#420 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#420 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#417 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#415 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#421 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#421 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#418 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#416 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#152 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#152 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#148 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#147 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#422 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#422 irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim &cfs_rq->removed.lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim &obj_hash[i].lock irq_context: 0 sb_writers#4 &sb->s_type->i_mutex_key#9 fs_reclaim pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#418 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#424 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#424 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#424 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#421 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#421 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#421 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#419 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#425 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#422 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#420 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#426 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#426 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#423 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#421 irq_context: 0 sched_map-wait-type-override &obj_hash[i].lock pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#427 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#427 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#422 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#428 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#425 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#423 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#153 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#153 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#149 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#149 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#149 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#429 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#426 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#424 irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#430 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#430 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#53 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#53 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#430 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#51 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#427 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#50 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#427 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#425 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#154 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#154 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#150 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#149 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#24 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#24 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#24 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#23 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#150 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#431 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#431 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#431 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#431 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#428 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#426 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#426 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#426 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#25 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#25 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#25 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#54 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#54 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#54 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#24 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#52 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#51 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#156 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#156 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#152 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#151 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#151 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#151 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#432 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#429 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#427 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#427 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#427 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#430 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#428 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#434 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#434 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#431 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#429 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#435 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#432 &rq->__lock irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) quarantine_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#430 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#157 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#157 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#153 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#152 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#433 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#431 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#55 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#55 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#55 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#55 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#53 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#52 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#26 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#26 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#26 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#26 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#158 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#158 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#26 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#25 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#154 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#154 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#154 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#153 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#437 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#437 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#434 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#432 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#159 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#159 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#155 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#154 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#154 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#154 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#433 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#56 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#54 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#54 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#54 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#27 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#53 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#53 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#53 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#156 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#155 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#439 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#439 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#439 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#28 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#28 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#27 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#27 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#27 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#26 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#436 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#434 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#57 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#57 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#57 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#57 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#55 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#54 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#440 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#161 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#161 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#157 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#437 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#156 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#156 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#156 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#435 irq_context: 0 &sb->s_type->i_mutex_key#18 &n->list_lock irq_context: 0 &sb->s_type->i_mutex_key#18 &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#58 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#58 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#56 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#55 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#438 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#436 irq_context: 0 (wq_completion)hci1#6 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#442 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#442 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#442 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#439 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#439 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#439 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#437 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#59 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#59 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#162 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#162 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#162 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#57 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#162 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#158 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#56 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#157 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#29 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#29 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#29 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#28 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#27 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#443 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#443 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#440 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#438 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#444 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#444 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#444 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#444 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#441 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#441 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#441 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#445 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#445 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#445 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#442 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#440 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#163 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#163 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#163 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#159 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#159 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#158 irq_context: 0 (wq_completion)nfc34_nci_tx_wq#9 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#164 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#164 irq_context: 0 &ndev->req_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#446 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#446 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#443 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#443 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#443 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#160 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#159 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#441 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#165 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#165 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#165 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#161 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#160 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#442 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#442 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#442 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#166 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#166 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#162 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#161 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#448 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#448 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#445 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#445 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#443 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#167 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#167 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#167 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#167 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#163 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#162 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#164 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#163 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#163 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#163 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#61 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#59 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#58 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#165 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#164 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#170 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#170 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#170 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#166 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#165 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#165 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#165 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#171 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#167 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#172 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#168 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#167 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#472 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#472 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#472 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#469 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#469 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#469 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#467 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#186 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#182 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#181 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#473 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#473 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#470 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#468 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#187 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#183 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#182 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#184 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#184 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#184 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#183 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#474 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#474 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#471 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#469 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#70 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#70 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#70 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#70 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#68 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#68 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#68 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#475 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#475 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#189 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#189 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#185 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#184 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#472 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#470 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#476 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#476 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#476 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#473 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#471 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#190 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#190 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#186 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#185 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#185 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#185 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#477 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#474 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#472 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#71 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#71 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#71 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#69 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#68 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#191 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#191 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#191 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#475 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#473 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#473 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#473 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#191 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#187 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#187 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#187 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#186 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#186 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#186 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#30 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#30 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#30 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#72 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#72 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#70 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#30 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#29 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#28 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#69 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#479 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#479 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#476 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#476 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#476 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#474 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#480 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#480 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#480 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#477 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#475 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#475 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#475 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#192 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#192 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#192 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#192 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#188 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#188 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#188 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#481 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#481 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#187 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#478 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#476 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#483 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#480 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#478 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#481 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#479 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#194 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#194 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#190 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#189 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#195 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#191 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#190 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#485 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#485 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#485 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#482 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#480 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#486 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#483 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#481 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#197 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#197 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#193 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#193 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#193 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#192 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#73 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#71 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#70 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#483 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#199 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#199 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#195 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#194 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#489 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#489 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#486 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#484 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#490 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#487 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#485 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#200 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#196 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#195 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#491 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#491 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#488 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#486 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 &hdev->req_lock (wq_completion)hci1#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#502 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#502 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &hdev->req_lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &c->lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#499 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#499 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#499 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#497 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#497 &rq->__lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &c->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &n->list_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex uevent_sock_mutex.wait_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &rq->__lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock hci_cb_list_lock.wait_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &rq->__lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &hdev->req_lock &hdev->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->tx_work) &list->lock#5 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci4#4 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#500 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#498 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#208 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#208 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#204 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#203 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#504 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#504 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#504 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#504 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#501 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#501 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#499 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#505 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#505 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#505 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#502 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#500 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#506 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#503 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#501 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#507 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#507 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#507 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#507 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#504 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#502 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#508 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#508 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#505 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#503 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#503 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#509 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#509 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#509 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#506 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#504 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#507 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#505 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#511 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#508 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#506 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#512 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#512 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#512 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#512 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#512 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#509 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#507 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#209 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#209 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#204 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#79 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#78 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#513 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#513 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#510 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#508 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#210 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#210 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#206 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#206 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#205 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#511 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#509 irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex rfkill_global_mutex rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc33_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc33_nci_rx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#247 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#247 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#240 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#240 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#240 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#240 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#240 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#240 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc33_nci_tx_wq#10 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#243 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#242 irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#558 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#236 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#558 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#235 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#235 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#235 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#555 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#553 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#553 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#553 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#449 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#559 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#559 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#559 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#559 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#556 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#554 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#557 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#555 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#241 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#241 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#237 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#236 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#561 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#561 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#561 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#558 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#556 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#556 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#556 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#242 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#242 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#238 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#238 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#238 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#237 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#562 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#562 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#562 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#559 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#559 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#559 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#557 irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#9 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc32_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc32_nci_rx_wq#9 irq_context: 0 (wq_completion)nfc32_nci_tx_wq#9 irq_context: 0 (wq_completion)nfc32_nci_tx_wq#9 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc30_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc30_nci_cmd_wq#10 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#9 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc35_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc30_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc35_nci_rx_wq#8 irq_context: 0 (wq_completion)nfc30_nci_tx_wq#10 irq_context: 0 (wq_completion)nfc35_nci_tx_wq#8 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#10 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc29_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc29_nci_tx_wq#10 irq_context: 0 (wq_completion)nfc29_nci_tx_wq#10 &rq->__lock irq_context: 0 (wq_completion)nfc29_nci_tx_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq#9 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc31_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc31_nci_rx_wq#9 irq_context: 0 (wq_completion)nfc31_nci_rx_wq#9 &rq->__lock irq_context: 0 (wq_completion)nfc31_nci_rx_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc31_nci_tx_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#12 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc26_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc28_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc28_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc28_nci_tx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc27_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc27_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc27_nci_rx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc27_nci_rx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc27_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#10 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc24_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc24_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc24_nci_rx_wq#10 &rq->__lock irq_context: 0 (wq_completion)nfc24_nci_rx_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc29_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc29_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc29_nci_rx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc29_nci_rx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc24_nci_tx_wq#10 irq_context: 0 (wq_completion)nfc29_nci_tx_wq#11 irq_context: 0 (wq_completion)nfc29_nci_tx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc29_nci_tx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#248 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#248 irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc23_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc23_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#244 irq_context: 0 (wq_completion)nfc23_nci_tx_wq#12 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#243 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#243 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#243 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc25_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc25_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc25_nci_rx_wq#9 irq_context: 0 (wq_completion)nfc25_nci_rx_wq#9 &rq->__lock irq_context: 0 (wq_completion)nfc25_nci_rx_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc25_nci_tx_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc22_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc22_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc22_nci_rx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc20_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc20_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc20_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc21_nci_cmd_wq#9 irq_context: 0 (wq_completion)nfc21_nci_rx_wq#9 irq_context: 0 (wq_completion)nfc21_nci_rx_wq#9 &rq->__lock irq_context: 0 (wq_completion)nfc21_nci_rx_wq#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc21_nci_tx_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc19_nci_rx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc19_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc18_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc18_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc18_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc17_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc17_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc17_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc17_nci_tx_wq#10 irq_context: 0 (wq_completion)nfc17_nci_tx_wq#10 &rq->__lock irq_context: 0 (wq_completion)nfc17_nci_tx_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc17_nci_tx_wq#10 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc16_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc16_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc16_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc15_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc15_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#10 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#10 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#10 &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_tx_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc12_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#94 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#94 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#94 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#94 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#92 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#90 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#90 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#569 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#569 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#566 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#564 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#515 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#515 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#515 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#515 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#512 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#512 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#512 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#35 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#35 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#34 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#33 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#33 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#33 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#249 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#570 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#567 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#249 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#245 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#565 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#244 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#95 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#95 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#93 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#91 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#571 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#211 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#571 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#211 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#571 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#82 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#82 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#82 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#207 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#568 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#566 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#572 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#569 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#567 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#516 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#516 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#250 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#206 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#513 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#511 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#250 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#94 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#82 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#80 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#80 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#80 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#79 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#94 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#94 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#92 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#92 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#92 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#514 rcu_node_0 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#514 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#514 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#512 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#212 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#212 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#208 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#207 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#573 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#573 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#518 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#515 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#513 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#570 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#568 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#213 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#213 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#209 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#208 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#208 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#208 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#568 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#568 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#97 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#95 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#519 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#519 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#516 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#514 irq_context: 0 (wq_completion)events (work_completion)(&barr->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#214 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#214 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#214 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#210 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#209 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#251 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#251 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#247 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#246 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#36 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#36 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#35 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#93 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#574 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#574 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#34 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#34 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#114 irq_context: 0 (wq_completion)bond0#114 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#114 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#114 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#114 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#520 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#520 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#517 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#515 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#215 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#521 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#521 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#518 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#518 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#518 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#516 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#516 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#516 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#211 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#210 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#571 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#569 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#569 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg1#104 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#106 irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#109 irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#522 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#522 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#216 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#216 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#212 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#211 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#211 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#211 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#523 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#523 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#520 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#518 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#252 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#252 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#252 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#248 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#247 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#247 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#575 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#572 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#572 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#572 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#570 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#576 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#573 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#571 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#212 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#253 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#253 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#98 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#96 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#94 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#94 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#249 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#248 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#577 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#577 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#574 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#572 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#572 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#99 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#99 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#37 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#37 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#97 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#254 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#254 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#95 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#250 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#524 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#524 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#524 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#521 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#521 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#521 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#519 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#36 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#35 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#249 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#578 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#578 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#578 irq_context: 0 (wq_completion)bond0#114 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)nfc2_nci_rx_wq#575 irq_context: 0 (wq_completion)bond0#114 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#114 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#114 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#114 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#573 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#579 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#579 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#98 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#96 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#255 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#251 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#250 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#38 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#38 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#37 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#36 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#36 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#580 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#577 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#575 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#256 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#256 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#256 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#256 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#252 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#251 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#578 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#576 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#582 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#582 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#579 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#577 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#577 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#577 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#257 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#257 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#253 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#583 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#583 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#583 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#583 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#580 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#580 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#580 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#580 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#578 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#258 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#258 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#254 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#254 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#254 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#253 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#584 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#584 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#581 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#579 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#579 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#579 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#585 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#585 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#585 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#585 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#582 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#580 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#580 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#580 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#259 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#259 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#525 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#525 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#522 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#520 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#520 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#520 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#218 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#218 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#214 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#214 rcu_node_0 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#214 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#214 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#213 irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#109 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#586 irq_context: 0 &sb->s_type->i_mutex_key#18 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#255 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#583 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#254 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#581 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#582 irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#260 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#260 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#256 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#255 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#588 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#588 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#588 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#588 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#6 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#6 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#6 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#6 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#219 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#526 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#526 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#523 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#215 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#83 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#83 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#521 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#83 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#521 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#521 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#214 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#83 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#83 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#83 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#81 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#80 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#32 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#32 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#32 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#31 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#527 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#527 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#524 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#522 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#216 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#215 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#215 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#215 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#84 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#84 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#84 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#84 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#82 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#221 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#221 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#221 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#528 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#528 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#217 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#525 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#525 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#525 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#216 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#523 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#529 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#529 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#526 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#526 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#526 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#5 &sb->s_type->i_mutex_key#13 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#524 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#222 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#222 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#222 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#222 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#218 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#217 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#530 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#585 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#583 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#583 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#583 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#261 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#261 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#261 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#590 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#590 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#587 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#585 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#256 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#591 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#591 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#591 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#588 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#586 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#586 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#586 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#592 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#592 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#592 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#592 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#589 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#587 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#262 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#593 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#593 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#590 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#588 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#594 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#594 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#589 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#595 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#595 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#530 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#527 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#525 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#595 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#85 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#85 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#83 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#83 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#82 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#595 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#592 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#592 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#592 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#590 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#263 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#263 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#263 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#258 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#257 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#593 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#591 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#597 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#597 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#597 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#597 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#594 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#592 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#264 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#264 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#264 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#259 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#258 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#258 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#258 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#598 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#598 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#595 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#593 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#101 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#101 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#99 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#99 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#599 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#599 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#599 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#97 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#594 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#600 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#600 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#597 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#597 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#597 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#601 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#601 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#598 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#596 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#265 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#265 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#265 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#260 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#259 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#602 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#602 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#599 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#597 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#266 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#266 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#266 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#266 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem stock_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem key irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem pcpu_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem percpu_counters_lock irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#261 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#260 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#603 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#603 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#600 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#600 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#600 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#598 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#262 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#261 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#261 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#261 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#604 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#604 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#601 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#599 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#605 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#605 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#605 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#605 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#602 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#268 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#268 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#268 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#268 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#263 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#262 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#600 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#603 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#601 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#607 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#607 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#604 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#604 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#604 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#602 irq_context: 0 nfc_devlist_mutex rcu_read_lock rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#269 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#269 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#269 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#269 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#264 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#264 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#264 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#263 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#608 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#608 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#609 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#609 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#606 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#606 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#606 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#604 irq_context: 0 &dev->mutex kn->active#4 rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex kn->active#4 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#270 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#265 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#264 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#610 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#610 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#610 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#610 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#607 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#605 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#271 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#271 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#266 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#265 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#265 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#611 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#611 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#611 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#608 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#608 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#608 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#609 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#607 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#607 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#613 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#613 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#610 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#608 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#608 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#608 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#272 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#272 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#267 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#614 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#614 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#614 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#531 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#531 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#528 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#526 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#223 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#219 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#219 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#219 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#614 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#611 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#611 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#611 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#273 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#273 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#268 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#609 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#267 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#612 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#610 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#274 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#274 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#274 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#269 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#268 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#616 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#616 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#616 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#532 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#532 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#532 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#532 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#616 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#613 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#529 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#529 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#613 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#527 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#613 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#527 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#527 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#218 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#611 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#611 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) rcu_node_0 irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#611 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#617 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#617 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#617 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#614 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#275 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#270 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#269 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#276 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#276 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#276 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#271 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#533 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#533 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#533 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#533 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#530 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#528 irq_context: 0 misc_mtx &dev->mutex rfkill_global_mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#534 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#271 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#271 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#270 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#277 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#531 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#529 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#86 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#86 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#84 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#224 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#224 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#83 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#220 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#219 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#535 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#535 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#532 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#530 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) &n->list_lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&hdev->cmd_work) &n->list_lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#277 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#277 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#618 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#618 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#615 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#277 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#613 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#272 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#272 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#536 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#536 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#533 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#533 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#533 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#531 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#225 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#225 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#225 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#225 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#221 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#272 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#102 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#102 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#271 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 jbd2_handle &ei->i_data_sem rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#220 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#87 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#87 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#87 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#87 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#87 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#87 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#85 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#84 irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &rcu_state.expedited_wq irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &f->f_pos_lock &type->i_mutex_dir_key#5 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#102 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#100 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#98 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#449 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#616 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#537 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#537 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#537 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#537 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#534 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#532 irq_context: 0 &dev->mutex rcu_node_0 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#226 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#226 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#226 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#226 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#222 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#221 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#614 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#620 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#620 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#620 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#620 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#617 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#617 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex fill_pool_map-wait-type-override &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#441 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#438 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#980 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#980 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#973 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)phy229 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#538 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#538 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#974 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#974 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#968 irq_context: 0 (wq_completion)events deferred_process_work rtnl_mutex sched_map-wait-type-override pool_lock#2 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#450 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#442 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#439 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#176 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#173 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3 jbd2_handle mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#533 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#227 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#227 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#227 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#227 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &base->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &n->list_lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#223 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#222 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &____s->seqcount#2 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->cmd_work) &____s->seqcount irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &n->list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci4#5 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) &list->lock#5 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci4#6 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci4#6 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#207 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock &rnp->exp_wq[1] irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#988 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#988 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg1#208 irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#8 &type->i_mutex_dir_key#6/1 &root->kernfs_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#981 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#975 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#975 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#975 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#453 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#445 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#445 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#989 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#989 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#989 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#989 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#989 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#107 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#982 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#209 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#982 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#975 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#969 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#451 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#451 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#451 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#451 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#451 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#443 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#440 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#440 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#440 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#970 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#970 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#970 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#210 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#210 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#210 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#976 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#976 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#976 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#454 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#454 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#446 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#443 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#617 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#615 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#278 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#278 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#278 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#273 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#272 irq_context: 0 (wq_completion)wg-crypt-wg2#104 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#621 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#621 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#618 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#618 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#618 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#616 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#103 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#103 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#103 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#103 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#101 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#99 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#99 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#99 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#279 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#279 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#274 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#273 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#622 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#619 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#617 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#280 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#280 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#275 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#275 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#275 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#623 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#623 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#274 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#620 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#618 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#624 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#281 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#621 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#621 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#621 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#619 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#619 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#619 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#276 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#275 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#104 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#104 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#102 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#100 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#625 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#625 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#622 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#620 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#626 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#626 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#626 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#623 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#623 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#623 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#621 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#282 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#282 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#277 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#627 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#627 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#624 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#622 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#622 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#622 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#12 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#11 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#11 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_tx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#11 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#11 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc14_nci_tx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc16_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc16_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc16_nci_rx_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc16_nci_rx_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc16_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc18_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc18_nci_tx_wq#12 irq_context: 0 (wq_completion)nfc21_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc21_nci_tx_wq#10 irq_context: 0 (wq_completion)nfc26_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc26_nci_rx_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc26_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc28_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc25_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc25_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc25_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc25_nci_rx_wq#10 &rq->__lock irq_context: 0 (wq_completion)nfc25_nci_rx_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc25_nci_tx_wq#10 irq_context: 0 (wq_completion)nfc25_nci_tx_wq#10 &rq->__lock irq_context: 0 (wq_completion)nfc25_nci_tx_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#14 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#14 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc26_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc26_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc26_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc26_nci_rx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc48_nci_cmd_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc47_nci_cmd_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc47_nci_cmd_wq#3 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc47_nci_cmd_wq#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc47_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc47_nci_rx_wq#3 irq_context: 0 (wq_completion)nfc47_nci_tx_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc46_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc46_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc46_nci_rx_wq#3 irq_context: 0 (wq_completion)nfc46_nci_tx_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc45_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc45_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc45_nci_rx_wq#3 irq_context: 0 (wq_completion)nfc45_nci_tx_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc44_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc44_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc44_nci_rx_wq#3 irq_context: 0 (wq_completion)nfc44_nci_tx_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc43_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc43_nci_cmd_wq#3 irq_context: 0 (wq_completion)nfc43_nci_rx_wq#3 irq_context: 0 (wq_completion)nfc43_nci_tx_wq#3 irq_context: 0 &ndev->req_lock (wq_completion)nfc42_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc42_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc42_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc42_nci_tx_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc41_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc41_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc41_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc41_nci_rx_wq#4 &rq->__lock irq_context: 0 (wq_completion)nfc41_nci_tx_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc40_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc40_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc40_nci_rx_wq#5 irq_context: 0 (wq_completion)nfc40_nci_rx_wq#5 &rq->__lock irq_context: 0 (wq_completion)nfc40_nci_rx_wq#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc40_nci_tx_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc39_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc39_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc39_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc39_nci_rx_wq#4 &rq->__lock irq_context: 0 (wq_completion)nfc39_nci_rx_wq#4 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc39_nci_tx_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc38_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc38_nci_cmd_wq#4 irq_context: 0 (wq_completion)nfc38_nci_rx_wq#4 irq_context: 0 (wq_completion)nfc38_nci_tx_wq#4 irq_context: 0 &ndev->req_lock (wq_completion)nfc37_nci_cmd_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc37_nci_cmd_wq#5 &rq->__lock irq_context: 0 (wq_completion)nfc37_nci_cmd_wq#5 irq_context: 0 (wq_completion)nfc37_nci_rx_wq#5 irq_context: 0 (wq_completion)nfc37_nci_tx_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc36_nci_cmd_wq#5 irq_context: 0 &ndev->req_lock (wq_completion)nfc36_nci_cmd_wq#5 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc36_nci_cmd_wq#5 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc36_nci_rx_wq#5 irq_context: 0 (wq_completion)nfc36_nci_tx_wq#5 irq_context: 0 (wq_completion)nfc36_nci_tx_wq#5 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#10 &rq->__lock irq_context: 0 (wq_completion)nfc35_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc35_nci_rx_wq#9 irq_context: 0 (wq_completion)nfc35_nci_rx_wq#9 &rq->__lock irq_context: 0 (wq_completion)nfc35_nci_tx_wq#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc34_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc34_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc34_nci_tx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc33_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc33_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc33_nci_tx_wq#11 irq_context: 0 (wq_completion)nfc33_nci_tx_wq#11 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#10 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#10 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc32_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc32_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc32_nci_tx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc31_nci_cmd_wq#10 irq_context: 0 (wq_completion)nfc31_nci_rx_wq#10 irq_context: 0 (wq_completion)nfc31_nci_tx_wq#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc26_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc30_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc30_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc30_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#12 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc29_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc29_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc29_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc25_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc25_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc25_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc25_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc28_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc28_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc28_nci_rx_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc28_nci_rx_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc28_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#12 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc27_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc27_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc27_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc24_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc24_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc24_nci_rx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc24_nci_rx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc24_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc21_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc21_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc21_nci_rx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc21_nci_rx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc21_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc23_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc23_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc23_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc22_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc22_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc22_nci_rx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc20_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc20_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc20_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc19_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc18_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc18_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc18_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc18_nci_tx_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc18_nci_tx_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc16_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc16_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc16_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc17_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc17_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc17_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc17_nci_tx_wq#11 irq_context: 0 (wq_completion)nfc17_nci_tx_wq#11 &rq->__lock irq_context: 0 (wq_completion)nfc17_nci_tx_wq#11 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#12 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc15_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#11 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#12 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#39 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#39 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#38 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#38 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#38 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#37 irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#105 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#105 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#105 irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#103 irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock pool_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#101 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#101 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#207 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#208 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#283 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#283 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#278 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#278 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#277 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#629 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#626 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#626 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#626 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#624 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#624 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#624 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &pipe->mutex/1 fs_reclaim mmu_notifier_invalidate_range_start &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#631 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#628 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#628 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#628 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#626 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#626 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#626 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#284 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#284 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#284 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#279 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#278 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#106 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#106 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#632 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#632 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#40 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#40 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#40 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#39 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#104 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#38 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#629 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#102 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#102 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#102 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#627 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#630 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#628 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#285 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#285 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#285 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#280 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#279 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#634 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#634 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#634 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#634 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#631 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#629 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#629 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#629 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#635 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#635 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#635 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#632 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#630 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#636 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#633 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#633 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#633 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#631 irq_context: 0 &mm->mmap_lock &mapping->i_mmap_rwsem &anon_vma->rwsem kfence_freelist_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#637 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#634 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#632 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#286 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#286 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#281 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#280 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#635 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#635 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#633 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#636 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#634 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#287 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#287 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#287 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#287 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#282 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#282 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#282 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#281 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#281 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#281 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#640 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#640 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#107 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#107 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#635 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#105 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#103 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#641 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#641 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#636 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#642 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#642 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#637 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#643 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#643 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#643 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#640 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#640 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#638 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#638 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#641 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#639 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#639 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#639 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#645 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#642 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#640 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#640 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#288 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#283 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#282 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#646 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#646 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#646 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#643 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#643 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#643 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#641 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#289 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#289 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#284 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#283 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#647 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#647 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#647 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#644 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#644 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#642 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#648 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#643 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#290 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#290 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#290 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#285 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#285 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#284 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#291 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#286 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#285 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#649 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#649 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#649 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#646 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#644 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#644 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#644 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#292 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#292 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#292 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#292 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#287 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#286 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#108 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#108 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#104 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#41 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#41 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#41 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#41 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#40 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#39 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#39 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#39 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#647 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#647 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#647 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#645 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#293 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#293 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#288 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#287 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#287 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#287 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#651 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#651 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#651 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#648 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#646 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#289 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#294 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#294 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#294 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#109 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#109 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#107 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#105 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#295 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#652 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#652 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#652 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#652 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#652 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#649 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#649 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#649 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#647 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#653 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#650 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#648 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#110 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#110 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#110 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#296 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#296 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#108 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#290 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#288 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#288 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#106 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#654 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#654 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#654 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#651 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#649 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#655 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#655 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#655 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#655 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#652 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#650 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#297 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#291 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#291 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#289 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#656 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#656 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#653 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#653 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#653 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#298 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#298 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#292 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#290 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#290 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#290 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#657 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#657 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#652 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#653 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#299 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#299 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#293 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#293 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#659 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#293 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#659 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#659 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#291 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#659 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#656 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#654 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#111 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#111 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#109 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#107 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#107 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#107 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#660 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#660 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#292 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#657 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#655 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#112 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#112 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#110 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#661 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#661 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#108 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#656 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#662 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#659 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#657 irq_context: 0 &tty->ldisc_sem &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#663 irq_context: 0 &tty->ldisc_sem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#663 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#660 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#301 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#301 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#301 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#658 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#301 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#295 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#293 irq_context: 0 (wq_completion)wg-kex-wg1#186 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#664 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#664 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#661 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#659 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#296 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#294 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#666 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#666 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#666 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#666 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#663 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#661 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#661 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#661 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#664 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#662 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#663 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#669 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#669 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#669 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#664 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#303 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#303 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#303 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#303 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#297 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#670 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#667 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#665 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#671 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#668 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#666 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#672 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#672 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#672 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#669 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#669 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#669 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#667 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#673 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#673 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#670 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#670 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#670 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#668 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#668 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#304 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#304 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#304 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#304 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#298 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#298 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#298 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#296 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#296 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#674 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#674 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#671 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#669 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#113 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#113 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#111 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#109 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#675 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#675 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#675 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#675 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#672 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#670 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#676 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#676 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#676 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#673 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#673 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#673 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#671 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#114 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#112 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#110 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#110 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#110 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#305 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#305 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#299 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#299 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#299 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#297 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#297 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#297 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#677 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#677 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#677 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#677 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#674 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#672 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#672 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#678 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#675 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#673 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#679 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#679 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#306 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#306 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#306 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#306 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#679 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#679 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#679 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#300 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#298 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#676 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#676 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#676 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#674 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#677 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#675 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#675 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#675 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#307 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#307 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#299 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#681 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#678 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#682 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#682 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#682 key irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#682 pcpu_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#682 percpu_counters_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#682 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#679 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#677 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#683 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#683 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#683 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#680 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#680 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#680 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#678 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#678 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#678 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#308 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#308 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#308 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#308 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#302 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#300 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#684 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#681 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#679 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#679 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#679 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#115 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#115 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#113 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#309 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#309 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#303 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#301 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#685 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#685 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#682 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#680 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#680 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#680 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#681 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#116 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#116 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#310 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#310 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#116 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#116 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#116 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#304 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#302 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#687 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#687 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#684 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#684 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#684 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#114 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#112 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#682 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#683 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#311 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#311 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#305 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#305 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#305 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#303 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#312 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#306 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#306 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#306 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#304 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#690 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#690 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#690 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#687 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#685 irq_context: 0 (wq_completion)bat_events (work_completion)(&(&bat_priv->mcast.work)->work) rcu_read_lock &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#313 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#313 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#307 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#307 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#307 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#117 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#117 irq_context: 0 sb_writers#3 &type->i_mutex_dir_key#3/1 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#305 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#115 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#691 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#691 irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start &cfs_rq->removed.lock irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start &obj_hash[i].lock irq_context: 0 &type->i_mutex_dir_key#4 fs_reclaim mmu_notifier_invalidate_range_start pool_lock#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#688 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#686 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#113 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#692 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#689 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#687 irq_context: 0 misc_mtx nfc_devlist_mutex &root->kernfs_rwsem &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#314 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#314 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#308 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#306 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#306 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#306 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#693 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#690 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#690 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#690 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#688 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#118 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#118 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#116 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#315 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#315 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#114 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#114 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#114 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#309 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#309 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#309 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#307 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#694 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#694 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#694 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#691 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#691 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#691 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#689 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#690 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#316 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#316 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#310 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#308 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#119 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#119 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#119 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#119 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#117 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#115 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#317 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#317 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#317 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#311 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#309 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#318 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#318 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#312 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#310 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#310 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#310 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#696 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#693 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#120 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#120 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#118 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#118 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#118 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#691 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#116 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#121 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#121 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#119 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#119 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#119 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#117 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#319 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#319 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#319 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#313 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#311 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#698 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#698 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#696 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#694 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#320 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#320 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#314 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#312 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#312 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#312 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#15 irq_context: 0 misc_mtx (wq_completion)nfc8_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#42 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#42 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#40 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#40 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#122 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#122 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#122 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#122 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#120 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#120 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#120 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#118 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#118 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#118 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#321 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#321 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#315 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#315 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#313 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#43 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#43 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#700 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#700 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#42 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#697 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#695 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#695 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#695 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#41 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#701 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#701 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#701 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#698 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#696 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#322 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#322 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#316 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#316 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#314 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#323 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#317 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#315 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#123 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#121 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#119 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#128 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#18 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#126 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#124 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#719 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#719 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#715 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#334 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#713 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#47 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#47 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#45 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#45 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#45 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#44 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#720 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#716 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#716 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#714 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#714 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#714 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#721 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#721 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#717 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#715 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#335 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#335 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#335 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#325 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#722 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#722 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#718 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#716 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#716 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#716 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#336 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#336 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#336 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#336 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#336 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#336 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#328 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#326 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#723 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#723 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#723 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#723 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#719 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#719 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#719 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#717 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#717 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#717 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#329 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#327 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#327 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#327 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#724 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#724 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#720 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#718 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#338 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#338 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#330 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#328 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#725 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#725 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#721 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#721 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#721 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#719 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#726 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#726 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#726 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#726 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#722 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#722 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#722 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#720 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#727 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#727 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#727 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#727 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#727 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#727 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#727 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#723 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#721 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#728 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#728 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#728 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#724 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#724 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#724 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#722 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#729 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#725 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#723 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#339 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#331 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#331 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#331 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#329 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#329 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#329 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#724 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#731 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#727 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#727 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#340 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#725 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#725 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#340 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#340 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#340 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#332 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#332 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#332 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#330 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#726 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#341 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#333 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#333 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#333 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#331 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#331 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#331 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#727 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#734 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#734 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#728 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#735 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#735 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#731 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#731 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#731 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#729 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#342 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#342 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#342 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#342 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#334 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#334 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#334 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#332 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#736 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#732 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#730 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#737 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#737 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#737 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#733 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#731 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#731 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#731 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#343 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#343 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#343 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#335 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#335 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#335 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#333 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#738 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#738 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#732 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#129 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#129 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#129 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#129 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#127 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#125 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#125 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#125 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#739 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#739 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#735 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#733 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#344 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#336 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#336 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#336 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#334 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#740 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#740 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#740 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#740 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#740 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#740 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#734 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#130 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#130 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#128 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#128 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#128 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#126 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#735 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#735 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#345 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#337 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#335 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#742 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#742 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#742 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#742 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#738 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#736 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#743 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#743 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#737 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#737 &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#744 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#744 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#744 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#744 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#740 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#738 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#346 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#346 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#338 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#336 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#739 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#746 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#746 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#742 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#740 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#747 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#747 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#743 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#741 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#748 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#744 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#742 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#347 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#347 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#347 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#745 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#749 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#347 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#339 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#337 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#340 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#340 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#340 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#338 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#747 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#744 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#752 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#752 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#748 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#745 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#349 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#349 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#341 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#339 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#129 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#127 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#127 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#127 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#753 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#753 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#753 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#749 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#746 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#754 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#754 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#754 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#750 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#750 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#750 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#747 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#350 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#350 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#342 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#755 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#351 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#351 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#351 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#751 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#748 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#351 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#343 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#341 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#341 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#341 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#132 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#132 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#132 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#132 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#130 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#756 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#756 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#752 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#749 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#133 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#133 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#131 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#129 irq_context: 0 tty_mutex &tty->legacy_mutex &tty->ldisc_sem &o_tty->termios_rwsem/1 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#352 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#352 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#352 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#352 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#344 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#344 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#344 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#342 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#757 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#757 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#753 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#750 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#758 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#758 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#758 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#758 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#751 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#353 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#353 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#353 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#345 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#343 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#343 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#343 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#759 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#759 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#759 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#134 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#134 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#132 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#759 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#130 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#755 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#354 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#354 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#346 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#344 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#344 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#344 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#760 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#760 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#756 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#756 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#756 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#753 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#761 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#757 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#754 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#754 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#355 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#355 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#762 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#355 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#347 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#347 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#347 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#345 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#356 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#356 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#346 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#357 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#357 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#763 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#763 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#758 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#349 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#755 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#347 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#764 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#759 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#756 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#756 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#756 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#348 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#351 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#349 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#135 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#135 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#765 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#765 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#133 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#760 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#760 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#760 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#131 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#757 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#131 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#131 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#136 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#136 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#360 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#360 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#352 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#352 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#352 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#136 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#134 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#350 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#132 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#766 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#766 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#766 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#761 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#761 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#761 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#761 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#758 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#758 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#758 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#811 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#808 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#817 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#817 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#817 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#817 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#812 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#812 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#812 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#809 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#811 irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#4 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->lock#2 irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->timeout_work)->work) irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#2 irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#820 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#815 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#815 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#815 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#812 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#812 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#812 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci4#3 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#821 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#821 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#821 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#821 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#821 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#821 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#816 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#816 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#816 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#813 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#822 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#822 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#817 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#814 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#823 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#383 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#383 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#383 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#823 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#818 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#383 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#375 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#375 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#815 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#815 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#815 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#373 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#819 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#819 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#819 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#816 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#825 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#820 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#820 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#820 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#817 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#817 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#817 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#826 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#826 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#821 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#821 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#821 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#818 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#827 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#822 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#822 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#822 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#819 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#819 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#819 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#828 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#823 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#820 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#824 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#824 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#824 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#821 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#830 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#825 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#822 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#831 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#831 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#826 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#823 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#827 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#824 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#826 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#826 &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#826 &obj_hash[i].lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#829 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#834 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#834 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#834 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#384 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#376 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#374 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#374 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#374 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#835 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#835 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#835 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#835 stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#835 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#835 key irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#835 pcpu_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#835 percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#835 pcpu_lock stock_lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#835 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#830 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#146 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#146 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#144 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#142 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#836 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#836 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#831 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#831 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#831 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#828 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#828 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#828 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#837 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#837 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#832 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#829 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#147 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#829 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#829 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#147 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#147 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#385 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#385 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#385 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#147 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#145 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#143 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#377 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#375 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#375 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#838 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#838 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#839 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#839 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#839 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#834 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#831 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#840 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#835 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#832 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#841 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#836 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#833 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#386 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#386 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#386 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#378 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#376 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#837 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#837 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#837 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#837 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#837 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#834 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#387 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#387 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#387 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#387 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#379 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#377 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#377 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#377 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#388 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#388 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#388 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#380 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#378 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#378 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#378 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#844 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#844 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#844 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#844 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#839 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#836 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#836 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#836 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#837 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#389 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#389 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#389 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#389 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#148 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#148 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#148 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#381 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#379 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#148 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#146 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#146 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#146 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#144 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#144 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#846 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#846 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#841 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#838 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#390 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#390 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#390 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#390 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#382 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#380 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#842 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#842 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#842 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#839 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#848 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#843 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#843 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#840 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#849 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#849 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#849 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#844 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#841 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#850 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#850 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#845 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#845 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#845 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#842 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#149 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#149 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#149 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#149 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#391 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#391 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#147 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#145 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#383 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#381 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#384 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#382 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#851 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#851 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#846 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#843 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#843 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#843 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#393 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#393 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#393 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#385 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#150 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#150 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#150 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#383 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#150 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#148 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#146 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#852 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#847 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#847 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#844 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#844 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#394 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#394 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#394 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#394 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#386 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#384 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#853 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#848 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#845 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#845 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#850 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#850 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#850 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#847 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#847 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#847 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#395 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#395 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#395 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#395 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#387 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#385 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#151 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#151 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#151 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#151 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#147 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#388 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#386 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#857 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#857 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#857 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#857 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#852 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#15 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#15 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#16 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc14_nci_rx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#15 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#15 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc14_nci_tx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc23_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc23_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc23_nci_rx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc23_nci_rx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc23_nci_tx_wq#15 irq_context: 0 tty_mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 tty_mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc29_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc29_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc29_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc35_nci_rx_wq#11 irq_context: 0 (wq_completion)nfc35_nci_tx_wq#11 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#858 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#858 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#858 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#858 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#858 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#397 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#397 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#389 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#389 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#389 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#387 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#152 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#152 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#152 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc40_nci_cmd_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc40_nci_cmd_wq#7 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc40_nci_cmd_wq#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc40_nci_cmd_wq#7 irq_context: 0 (wq_completion)nfc40_nci_rx_wq#7 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#152 irq_context: 0 (wq_completion)nfc40_nci_tx_wq#7 irq_context: 0 (wq_completion)nfc40_nci_tx_wq#7 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#150 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#148 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#148 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#148 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#51 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#49 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#48 irq_context: 0 &type->i_mutex_dir_key#4 remove_cache_srcu &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#21 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#21 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#21 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#21 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#21 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#17 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_rx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#153 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#153 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#151 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#18 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#149 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_rx_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc13_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_tx_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc15_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc15_nci_rx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&rfkill->sync_work) rfkill_global_mutex sched_map-wait-type-override &pool->lock &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc15_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc16_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc16_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc16_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc17_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc17_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc17_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc17_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc18_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc18_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc18_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc19_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#16 irq_context: 0 (wq_completion)nfc14_nci_tx_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc14_nci_tx_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc20_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc20_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc21_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc21_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc21_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc22_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc22_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc22_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc22_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc22_nci_tx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc22_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc24_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc24_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc24_nci_tx_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc25_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc25_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc25_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc25_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc39_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc39_nci_rx_wq#6 irq_context: 0 (wq_completion)nfc39_nci_tx_wq#6 irq_context: 0 (wq_completion)nfc39_nci_tx_wq#6 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#859 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#859 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#859 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#854 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#851 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#398 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#398 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#390 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#390 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#388 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#52 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#52 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#50 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#49 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#154 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#154 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#152 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#150 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#22 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#22 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#22 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#22 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#18 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#18 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#19 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_rx_wq#19 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_tx_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc38_nci_cmd_wq#6 irq_context: 0 &ndev->req_lock (wq_completion)nfc38_nci_cmd_wq#6 &rq->__lock irq_context: 0 (wq_completion)nfc38_nci_cmd_wq#6 irq_context: 0 (wq_completion)nfc38_nci_rx_wq#6 irq_context: 0 (wq_completion)nfc38_nci_tx_wq#6 irq_context: 0 (wq_completion)nfc38_nci_tx_wq#6 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc37_nci_cmd_wq#7 irq_context: 0 (wq_completion)nfc37_nci_cmd_wq#7 irq_context: 0 (wq_completion)nfc37_nci_rx_wq#7 irq_context: 0 (wq_completion)nfc37_nci_tx_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc35_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc35_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc35_nci_tx_wq#12 irq_context: 0 (wq_completion)nfc35_nci_tx_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc35_nci_tx_wq#12 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc36_nci_cmd_wq#8 irq_context: 0 (wq_completion)nfc36_nci_cmd_wq#8 irq_context: 0 (wq_completion)nfc36_nci_rx_wq#7 irq_context: 0 (wq_completion)nfc36_nci_tx_wq#7 irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc34_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc34_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc34_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc33_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc33_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc33_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc32_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc32_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc32_nci_rx_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc32_nci_tx_wq#12 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#15 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc29_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc29_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc29_nci_tx_wq#15 irq_context: 0 (wq_completion)nfc29_nci_tx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc29_nci_tx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq#12 &rq->__lock irq_context: 0 (wq_completion)nfc31_nci_cmd_wq#12 irq_context: 0 (wq_completion)nfc31_nci_rx_wq#12 irq_context: 0 (wq_completion)nfc31_nci_tx_wq#12 irq_context: 0 (wq_completion)nfc31_nci_tx_wq#12 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq#13 irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc30_nci_cmd_wq#13 irq_context: 0 (wq_completion)nfc30_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc30_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc30_nci_tx_wq#13 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc28_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc28_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc28_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc28_nci_tx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc28_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc27_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc27_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc27_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#18 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc26_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc26_nci_tx_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc23_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc23_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc23_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_tx_wq#18 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#23 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#23 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#23 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#23 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#53 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#53 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#51 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#50 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#155 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#153 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#151 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#399 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#399 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#399 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#399 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#391 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#389 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#860 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#860 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#855 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#852 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#861 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#862 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#862 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#857 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#857 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#857 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#853 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#853 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#853 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#858 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#858 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#858 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#854 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#864 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#859 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#855 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#865 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#865 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#865 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#860 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#856 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#400 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#400 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#400 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#400 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#392 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#392 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#392 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#861 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#857 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#401 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#393 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#391 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#867 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#867 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#867 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#156 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#156 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#156 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#867 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#862 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#402 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#402 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#154 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#152 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#394 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#392 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#392 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#868 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#868 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#403 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#403 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#395 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#393 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#868 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#863 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#863 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#859 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#869 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#864 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#864 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#864 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#860 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#404 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#404 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#404 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#404 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#396 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#396 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#396 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 kn->active#4 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#394 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#394 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#54 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#54 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#52 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#157 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#157 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#155 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#155 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#155 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#51 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#153 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#153 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#153 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#870 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#870 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#865 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#405 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#405 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#397 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#395 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#395 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#395 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#158 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#158 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#158 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#158 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#158 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#156 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#156 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#156 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#154 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#154 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#154 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#871 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#871 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#871 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#866 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#862 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#862 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#862 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &xt[i].mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#867 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#867 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#873 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#873 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#873 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#874 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#874 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#874 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#875 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#868 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#863 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#398 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#396 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#876 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#876 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#869 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#869 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#869 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#864 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#407 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#399 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#399 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#399 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#397 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#397 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#397 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#877 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#877 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#870 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#865 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#865 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#865 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#159 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#159 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#157 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#155 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#155 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#155 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#117 irq_context: 0 (wq_completion)bond0#117 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#408 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#408 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#400 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#398 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#881 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#881 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#874 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#874 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#869 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#409 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#409 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#409 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#401 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#399 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#410 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#410 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#875 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#875 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#875 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#410 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#402 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#400 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#400 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#400 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#870 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#876 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#876 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#876 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#871 irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#884 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#877 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#877 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#877 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#872 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#872 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#885 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#885 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#885 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#873 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#873 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#873 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#411 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#411 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#411 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#411 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#403 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#401 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#401 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#401 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#880 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#875 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#412 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#412 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#404 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#402 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#878 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#878 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#878 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#884 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#884 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#884 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#879 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#879 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#879 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#413 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#413 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#405 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#405 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#405 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#403 irq_context: 0 nfc_devlist_mutex uevent_sock_mutex &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#414 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#406 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#406 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#406 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#404 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#404 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#404 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#892 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#885 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#885 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#885 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#880 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#880 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#880 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#886 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#886 &obj_hash[i].lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#886 pool_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#893 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#894 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#894 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#894 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#894 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#887 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#887 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#881 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#881 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#881 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#895 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#888 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#888 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#888 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#882 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#896 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#896 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#896 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#889 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#889 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#889 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#883 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#415 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#407 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#405 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#897 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#897 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#890 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#890 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#890 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#884 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#160 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#160 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#158 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#156 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#156 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#156 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#899 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#892 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#892 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#892 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#886 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#900 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#893 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#887 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#408 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#406 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#901 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#901 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#901 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#894 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#888 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#161 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#161 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#159 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#157 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#157 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#902 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#902 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#902 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#895 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#889 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#889 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#889 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#407 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#903 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#903 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#896 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#890 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#891 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#891 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#905 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#905 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#898 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#892 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#893 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#907 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#907 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#907 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#907 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#900 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#894 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#410 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#410 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#410 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#408 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#162 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#162 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#908 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#908 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#901 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#901 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#901 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#895 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#163 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#161 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#419 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#419 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#158 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#411 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#409 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#409 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#409 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#420 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#412 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#412 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#410 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#421 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#413 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#413 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#413 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#411 irq_context: 0 (crypto_chain).rwsem remove_cache_srcu &n->list_lock irq_context: 0 (crypto_chain).rwsem remove_cache_srcu &obj_hash[i].lock irq_context: 0 (crypto_chain).rwsem remove_cache_srcu pool_lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#909 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#909 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#909 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#909 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#902 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#896 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#896 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#896 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#422 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#422 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#414 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#164 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#164 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#162 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#159 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#897 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#423 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#423 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#165 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#165 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#163 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#415 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#413 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#160 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#53 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#52 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#52 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#52 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#416 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#424 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#911 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#911 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#904 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#904 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#898 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#425 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#425 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#425 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#417 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#414 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#912 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#912 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#912 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#899 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#899 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#899 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#426 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#426 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#426 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#418 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#415 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#166 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#166 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#166 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#166 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#164 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#164 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#164 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#161 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#909 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#903 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#420 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#417 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#917 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#917 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#917 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#910 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#904 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#167 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#167 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#165 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#162 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#429 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#429 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#421 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#56 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#56 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#56 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#56 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#54 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#54 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#54 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#53 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#53 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#53 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#418 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#918 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#911 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#905 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#422 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#419 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#419 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#419 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#57 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#168 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#57 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#168 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#57 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#166 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#163 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#163 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#163 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#25 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#25 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#57 stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#57 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#57 key irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#57 pcpu_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#57 percpu_counters_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#57 pcpu_lock stock_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#57 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#57 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#55 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#919 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#55 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#55 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_rx_wq#24 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#24 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#24 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#912 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#54 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#906 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#24 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#913 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#913 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#913 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#907 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#908 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#911 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#911 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#431 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#431 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#423 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#420 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#925 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#918 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#918 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#912 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#912 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#912 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#169 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#167 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#164 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#926 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#926 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#926 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#926 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#919 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#913 irq_context: 0 misc_mtx (wq_completion)nfc6_nci_cmd_wq#26 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#58 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#58 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#58 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#58 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#56 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#55 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#432 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#424 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#424 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#424 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#421 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#421 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#927 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#920 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#433 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#433 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#914 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#425 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#422 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#170 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#170 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#168 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#165 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#165 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#165 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#928 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#928 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#921 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#921 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#921 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#922 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#916 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#434 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#434 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#434 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#434 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#426 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#423 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#930 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#930 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#923 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#917 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#924 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#918 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#435 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#435 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#435 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#932 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#925 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#925 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#919 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#436 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#436 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#428 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#425 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#171 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#171 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#169 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#166 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#934 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#927 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#921 irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5#2 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci5 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#935 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#928 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#922 irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#437 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#437 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#437 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#437 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#429 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#426 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#426 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#426 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#936 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#936 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#929 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#929 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#929 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#923 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#937 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#924 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#925 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#931 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#938 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#938 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#938 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#939 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#939 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#932 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#932 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#932 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#926 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#940 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#940 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#933 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#927 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#934 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#934 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#934 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#928 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#438 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#438 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#438 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#438 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#430 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#427 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#427 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#439 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#935 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#935 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#431 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#935 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#431 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#431 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#935 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#428 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#428 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#428 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#929 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#434 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 crypto_alg_sem &obj_hash[i].lock irq_context: 0 &xt[i].mutex fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &xt[i].mutex fill_pool_map-wait-type-override &c->lock irq_context: 0 &xt[i].mutex fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &xt[i].mutex fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &xt[i].mutex fill_pool_map-wait-type-override pool_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock rcu_read_lock rcu_read_lock quarantine_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount#2 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->rx_work) &hdev->lock &____s->seqcount irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)bond0#115 irq_context: 0 (wq_completion)bond0#115 (work_completion)(&(&slave->notify_work)->work) irq_context: 0 (wq_completion)bond0#115 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#115 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#115 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#115 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#115 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#115 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#954 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#954 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#947 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#941 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) &list->lock#5 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci1#8 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci1#8 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#972 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 nl_table_wait.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#961 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#961 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#961 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#961 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#961 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#954 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#954 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#948 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem rtnl_mutex &idev->mc_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#116 &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#965 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#959 irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci1#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock key irq_context: 0 (wq_completion)nfc2_nci_tx_wq#950 irq_context: 0 &pipe->mutex/1 &mm->mmap_lock pcpu_lock irq_context: 0 &pipe->mutex/1 &mm->mmap_lock percpu_counters_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#966 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#966 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#966 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#961 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#117 (work_completion)(&(&slave->notify_work)->work) &obj_hash[i].lock irq_context: 0 (wq_completion)bond0#117 (work_completion)(&(&slave->notify_work)->work) &base->lock irq_context: 0 (wq_completion)bond0#117 (work_completion)(&(&slave->notify_work)->work) &base->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#117 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#117 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#117 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#117 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#117 _xmit_ETHER irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#964 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#964 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#957 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#957 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#957 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#951 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#951 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#951 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 rtnl_mutex team->team_lock_key#117 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#117 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#117 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#117 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#117 &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#117 &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#117 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#117 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#117 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#117 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#117 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#117 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#117 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#117 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#117 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#117 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#117 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#117 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#117 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#117 lock irq_context: 0 rtnl_mutex team->team_lock_key#117 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#117 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#117 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#117 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#117 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#117 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#117 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#117 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#117 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#117 console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)bond0#117 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#117 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#117 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#117 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#991 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#991 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#455 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#455 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#447 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#444 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#984 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#978 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#115 irq_context: 0 rtnl_mutex team->team_lock_key#115 fs_reclaim irq_context: 0 rtnl_mutex team->team_lock_key#115 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 rtnl_mutex team->team_lock_key#115 &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#115 netpoll_srcu irq_context: 0 rtnl_mutex team->team_lock_key#115 net_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#115 net_rwsem &list->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#115 &tn->lock irq_context: 0 rtnl_mutex team->team_lock_key#115 _xmit_ETHER irq_context: 0 rtnl_mutex team->team_lock_key#115 &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#115 input_pool.lock irq_context: 0 rtnl_mutex team->team_lock_key#115 rcu_read_lock &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#115 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#115 nl_table_lock irq_context: 0 rtnl_mutex team->team_lock_key#115 nl_table_wait.lock irq_context: 0 rtnl_mutex team->team_lock_key#115 rcu_read_lock &pool->lock irq_context: 0 rtnl_mutex team->team_lock_key#115 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#115 &in_dev->mc_tomb_lock irq_context: 0 rtnl_mutex team->team_lock_key#115 &im->lock irq_context: 0 rtnl_mutex team->team_lock_key#115 _xmit_ETHER &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#115 cbs_list_lock irq_context: 0 rtnl_mutex team->team_lock_key#115 &ndev->lock irq_context: 0 rtnl_mutex team->team_lock_key#115 sysfs_symlink_target_lock irq_context: 0 rtnl_mutex team->team_lock_key#115 lock irq_context: 0 rtnl_mutex team->team_lock_key#115 lock kernfs_idr_lock irq_context: 0 rtnl_mutex team->team_lock_key#115 &root->kernfs_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#115 &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 rtnl_mutex team->team_lock_key#115 &n->list_lock irq_context: 0 rtnl_mutex team->team_lock_key#115 &n->list_lock &c->lock irq_context: 0 rtnl_mutex team->team_lock_key#115 lweventlist_lock irq_context: 0 rtnl_mutex team->team_lock_key#115 lweventlist_lock &dir->lock#2 irq_context: 0 rtnl_mutex team->team_lock_key#115 (console_sem).lock irq_context: 0 rtnl_mutex team->team_lock_key#115 console_lock console_srcu console_owner_lock irq_context: 0 rtnl_mutex team->team_lock_key#115 console_lock console_srcu console_owner irq_context: 0 rtnl_mutex team->team_lock_key#115 console_lock console_srcu console_owner &port_lock_key irq_context: 0 rtnl_mutex team->team_lock_key#115 console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 &obj_hash[i].lock irq_context: 0 rtnl_mutex team->team_lock_key#115 rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 rtnl_mutex team->team_lock_key#115 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 rtnl_mutex team->team_lock_key#115 rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#948 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#942 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#443 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#443 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 fs_reclaim irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 rcu_read_lock &ndev->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#435 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 net_rwsem &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 net_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 net_rwsem &cfs_rq->removed.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 net_rwsem &obj_hash[i].lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 net_rwsem pool_lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 &tn->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 &n->list_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 &n->list_lock &c->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#115 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)bond0#115 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#115 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#432 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#432 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#432 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#432 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#432 &obj_hash[i].lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#7 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#7 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#7 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#171 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#168 irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 nl_table_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 nl_table_wait.lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 net_rwsem irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 net_rwsem &list->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 &tn->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci4#6 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#993 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#993 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#986 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#980 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#994 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#994 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#987 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#987 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#987 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#981 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#981 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#981 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#456 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#456 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#445 irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#108 irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#216 irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 rds_sock_lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 clock-AF_RDS irq_context: 0 &rs->rs_recv_lock irq_context: 0 rds_cong_monitor_lock irq_context: 0 rds_cong_lock irq_context: 0 &rs->rs_lock irq_context: 0 &q->lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#216 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#108 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#111 (work_completion)(&peer->transmit_packet_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx &root->kernfs_rwsem &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1018 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1018 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1018 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1018 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1009 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1009 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1003 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-crypt-wg0#99 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem &wg->device_update_lock (wq_completion)wg-kex-wg0#194 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1019 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1019 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1019 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1010 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1004 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1020 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1020 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1020 irq_context: 0 (wq_completion)nfc35_nci_rx_wq#13 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1011 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1005 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1005 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1005 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#473 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#465 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#461 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#461 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#461 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &tty->legacy_mutex &tty->ldisc_sem &buf->lock &tty->termios_rwsem &rq->__lock irq_context: 0 &tty->legacy_mutex &tty->ldisc_sem &buf->lock &tty->termios_rwsem &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#186 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#186 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#184 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#181 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#62 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#61 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#61 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#61 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#187 irq_context: 0 (wq_completion)wg-crypt-wg0#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#187 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#187 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#187 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#182 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#65 irq_context: 0 (wq_completion)wg-crypt-wg0#99 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#65 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#65 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#62 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#110 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#110 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)bond0#110 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#183 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#466 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#462 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1013 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1007 irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 lweventlist_lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 lweventlist_lock pool_lock#2 irq_context: 0 (wq_completion)nfc35_nci_tx_wq#13 irq_context: 0 (wq_completion)nfc35_nci_tx_wq#13 &rq->__lock irq_context: 0 (wq_completion)nfc35_nci_tx_wq#13 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 lweventlist_lock &dir->lock#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 rcu_read_lock &pool->lock irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) &rq->__lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci0#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#209 irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#444 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#444 irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc3_nci_rx_wq#436 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#433 irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#209 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#59 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#59 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#57 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#56 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#956 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#956 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#949 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#943 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#27 irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#209 irq_context: 0 (wq_completion)wg-kex-wg2#209 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#25 irq_context: 0 &vma->vm_lock->lock rcu_read_lock fill_pool_map-wait-type-override pool_lock#2 irq_context: 0 &vma->vm_lock->lock rcu_read_lock fill_pool_map-wait-type-override &____s->seqcount irq_context: 0 &vma->vm_lock->lock rcu_read_lock fill_pool_map-wait-type-override &c->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock fill_pool_map-wait-type-override &n->list_lock &c->lock irq_context: 0 &vma->vm_lock->lock rcu_read_lock fill_pool_map-wait-type-override pool_lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#25 irq_context: 0 (wq_completion)wg-kex-wg2#209 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#20 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)nfc7_nci_tx_wq#20 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 &dev->mutex uevent_sock_mutex &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#20 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#60 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#60 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#60 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#174 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#174 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#60 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#172 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#169 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#169 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#58 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#57 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#19 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)nfc9_nci_tx_wq#19 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc9_nci_tx_wq#19 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#20 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#20 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#17 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc13_nci_tx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc14_nci_rx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc14_nci_rx_wq#17 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc14_nci_rx_wq#17 key irq_context: 0 (wq_completion)nfc14_nci_rx_wq#17 pcpu_lock irq_context: 0 (wq_completion)nfc14_nci_rx_wq#17 percpu_counters_lock irq_context: 0 (wq_completion)nfc14_nci_rx_wq#17 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc14_nci_rx_wq#17 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#209 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)nfc15_nci_cmd_wq#15 irq_context: 0 (wq_completion)wg-kex-wg1#210 irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 devlinks.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &xa->xa_lock#19 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc15_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc15_nci_rx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc15_nci_tx_wq#15 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#17 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#17 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc16_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc16_nci_cmd_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc16_nci_cmd_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 stack_depot_init_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)nfc16_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc16_nci_rx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc16_nci_rx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc16_nci_tx_wq#17 irq_context: 0 (wq_completion)nfc16_nci_tx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc16_nci_tx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &xa->xa_lock#4 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex net_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &tn->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex subsys mutex#20 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#115 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc17_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc17_nci_cmd_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc17_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg2#210 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#212 irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#209 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#108 irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock irq_context: 0 cb_lock genl_mutex rtnl_mutex &rdev->wiphy.mtx rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#210 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#105 irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1014 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1004 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1004 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1004 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#996 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#996 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#996 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#990 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#990 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#990 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#182 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#182 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1014 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1008 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1008 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1008 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#475 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#475 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#118 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#182 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#180 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#180 rcu_node_0 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#180 &rcu_state.expedited_wq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#180 &rcu_state.expedited_wq &p->pi_lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#180 &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#180 &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#180 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#180 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#177 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#189 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#189 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#189 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#475 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#467 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#459 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#459 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 &____s->seqcount#2 irq_context: 0 (wq_completion)events (linkwatch_work).work rtnl_mutex team->team_lock_key#117 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 devlinks.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &xa->xa_lock#19 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 stack_depot_init_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &sb->s_type->i_mutex_key#3 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &xa->xa_lock#4 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex net_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &tn->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &root->kernfs_rwsem &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex subsys mutex#20 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex failover_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex lock kernfs_idr_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex lock kernfs_idr_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex lock kernfs_idr_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex uevent_sock_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 rtnl_mutex uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#116 &rq->__lock irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#116 irq_context: 0 cb_lock nlk_cb_mutex-GENERIC &devlink->lock_key#116 &devlink_port->type_lock irq_context: 0 (wq_completion)bond0#118 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem irq_context: 0 (wq_completion)bond0#118 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 (wq_completion)bond0#118 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex &tn->lock irq_context: 0 (wq_completion)bond0#118 (work_completion)(&(&slave->notify_work)->work) rtnl_mutex.wait_lock irq_context: 0 (wq_completion)bond0#118 (work_completion)(&(&slave->notify_work)->work) &p->pi_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#459 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#451 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#451 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#448 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#189 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#187 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#187 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#187 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#463 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1015 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1009 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1025 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1016 irq_context: 0 &hdev->req_lock (wq_completion)hci5#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1016 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1016 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1010 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#476 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#190 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#190 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#476 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#188 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#468 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#468 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#468 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#185 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#464 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#464 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1017 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1011 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#477 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#477 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#469 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1012 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#478 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#478 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#478 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#470 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#466 irq_context: 0 &dev->mutex uevent_sock_mutex &meta->lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1029 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1029 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1029 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1029 irq_context: 0 rtnl_mutex per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu) irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1019 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 rcu_node_0 irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#995 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#995 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#988 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#982 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#982 &rq->__lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock &pcp->lock &zone->lock irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock quarantine_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1019 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1019 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1013 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1015 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) (console_sem).lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&hdev->cmd_timer)->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1032 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1022 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1016 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#467 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1033 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1033 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1033 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1023 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1023 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1023 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1017 irq_context: 0 (wq_completion)events fqdir_free_work quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#480 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#480 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1024 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1018 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1025 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1019 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#481 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#481 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#481 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#983 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#481 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#473 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#469 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1026 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1020 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1037 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1038 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1038 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1038 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1028 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1028 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1028 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#997 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#990 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#984 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1021 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1039 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1039 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1039 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1022 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1040 irq_context: 0 &sb->s_type->i_mutex_key#17 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1040 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1040 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1040 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1030 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1023 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1023 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1041 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1031 irq_context: 0 (wq_completion)hci3#7 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#6 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &list->lock#7 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &hdev->req_wait_q irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &list->lock#5 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &data->read_wait irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &data->read_wait &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) lock#6 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) lock#6 kcov_remote_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) (console_sem).lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner &port_lock_key irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) console_lock console_srcu console_owner console_owner_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->req_wait_q &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock (&timer.timer) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) lock#6 &kcov->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) rcu_read_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &hdev->req_lock &c->lock irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#100 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1024 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#483 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#483 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &____s->seqcount#2 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->cmd_work) &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_rx_wq#475 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#484 &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) fs_reclaim irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) hci_sk_list.lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->power_on) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) chan_list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &x->wait#9 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &xa->xa_lock#17 &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock lock kernfs_idr_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock bus_type_sem irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock sysfs_symlink_target_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &root->kernfs_rwsem irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &dev->power.lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock dpm_list_mtx irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rlock-AF_NETLINK irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex rcu_read_lock &ei->socket.wq.wait &ep->lock &ep->wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex nl_table_wait.lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &k->k_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock subsys mutex#74 &k->k_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &list->lock#7 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock chan_list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->ident_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &list->lock#8 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_cb_list_lock &conn->chan_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock hci_sk_list.lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &n->list_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->rx_work) &hdev->lock uevent_sock_mutex &n->list_lock &c->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) &list->lock#8 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) &list->lock#5 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) &data->read_wait irq_context: 0 (wq_completion)hci3#8 (work_completion)(&hdev->tx_work) &list->lock#7 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&conn->pending_rx_work) irq_context: 0 (wq_completion)hci3#8 (work_completion)(&conn->pending_rx_work) &list->lock#9 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#484 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#484 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1032 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#476 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#476 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#476 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1025 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1025 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1025 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#472 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#472 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#472 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#472 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#472 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock stock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#46 nsim_bus_dev_list_lock &dev->mutex uevent_sock_mutex rcu_read_lock pcpu_lock stock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 devlinks.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &xa->xa_lock#19 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &xa->xa_lock#19 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &xa->xa_lock#19 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &xa->xa_lock#19 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rcu_read_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 stack_depot_init_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex bpf_devs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex bpf_devs_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex bpf_devs_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex bpf_devs_lock rcu_read_lock rhashtable_bucket irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &xa->xa_lock#4 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex (work_completion)(&(&devlink_port->type_warn_dw)->work) irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &devlink_port->type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex net_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex net_rwsem &list->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &tn->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &x->wait#9 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex gdp_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex gdp_mutex &k->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex gdp_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex gdp_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex gdp_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex gdp_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex gdp_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex gdp_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex gdp_mutex kobj_ns_type_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex lock kernfs_idr_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &root->kernfs_rwsem &root->kernfs_iattr_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex bus_type_sem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex sysfs_symlink_target_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &root->kernfs_rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex dpm_list_mtx irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex uevent_sock_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex uevent_sock_mutex fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex uevent_sock_mutex fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex uevent_sock_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex uevent_sock_mutex &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex uevent_sock_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex subsys mutex#20 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex subsys mutex#20 &k->k_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &dir->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex uevent_sock_mutex &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex dev_hotplug_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex dev_hotplug_mutex &dev->power.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex dev_base_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex input_pool.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#998 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#998 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#991 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &tbl->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex sysctl_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex nl_table_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex nl_table_wait.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex failover_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex proc_inum_ida.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex proc_subdir_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#985 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock remove_cache_srcu irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock remove_cache_srcu quarantine_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock remove_cache_srcu &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock remove_cache_srcu &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock remove_cache_srcu &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock _xmit_ETHER irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &pnettable->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex smc_ib_devices.mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &vn->sock_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex rcu_read_lock &ndev->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex batched_entropy_u32.lock crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &idev->mc_lock _xmit_ETHER &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#117 rtnl_mutex &obj_hash[i].lock pool_lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1033 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1026 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1026 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1026 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#191 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#485 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#485 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#485 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#191 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#189 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#189 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#189 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#477 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#186 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#473 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1044 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1044 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1044 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1034 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1027 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1027 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#66 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#66 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#66 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#66 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#64 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#64 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#64 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#63 irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#108 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 misc_mtx (wq_completion)nfc6_nci_cmd_wq#30 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#486 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#193 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#193 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#486 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#478 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#474 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#193 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1045 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1045 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1035 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1028 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#191 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#67 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#67 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#67 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#67 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#188 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#65 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#64 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#31 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#31 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#31 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#31 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#28 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#28 irq_context: 0 (wq_completion)wg-crypt-wg0#108 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1036 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1029 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1047 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1037 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1030 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#487 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#487 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#487 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#487 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#479 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#475 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1048 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1048 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1048 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1048 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1038 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1038 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1038 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1031 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#480 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#194 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#194 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#476 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#476 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#476 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#192 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#192 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#192 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1049 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1049 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1049 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1039 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1032 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1032 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1032 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#489 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#489 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#481 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#477 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1050 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1040 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1033 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1051 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1051 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1034 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#482 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#478 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#193 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#190 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1052 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1052 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1042 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1035 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#491 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#491 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#491 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#491 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#483 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#483 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#483 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#479 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#479 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#196 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#196 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#194 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex defrag4_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#191 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#191 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#191 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1053 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1053 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1043 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &____s->seqcount#2 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1043 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem ovs_mutex nf_ct_proto_mutex nf_hook_mutex &____s->seqcount irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1043 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1036 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1036 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#9 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1036 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#492 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#492 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#9 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#484 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#484 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#484 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_rcv#10 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#480 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1054 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_send#10 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1054 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1054 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1044 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1044 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1037 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1037 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1037 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &f->f_pos_lock sb_writers#14 quarantine_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#197 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#197 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#197 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#195 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#195 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#192 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1055 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1055 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1055 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc8_nci_rx_wq#21 irq_context: 0 misc_mtx (wq_completion)nfc8_nci_cmd_wq#22 irq_context: 0 misc_mtx (wq_completion)nfc8_nci_cmd_wq#22 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc8_nci_cmd_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc8_nci_cmd_wq#22 &rq->__lock cpu_asid_lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#493 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#493 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#485 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#481 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#68 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#68 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#68 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#68 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#66 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#65 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1045 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1045 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#22 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#22 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1038 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1038 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1038 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_rx_wq#22 irq_context: 0 (wq_completion)nfc7_nci_rx_wq#22 &rq->__lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc7_nci_tx_wq#22 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#32 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#32 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#32 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#198 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#198 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#32 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#29 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#29 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#29 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#9 irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#9 &rq->__lock irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#9 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)netns net_cleanup_work pernet_ops_rwsem (wq_completion)tipc_crypto#10 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#196 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#193 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1056 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1056 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1046 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1039 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1039 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1057 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1057 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1057 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1047 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1040 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1040 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1058 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1058 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1048 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1048 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1048 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#199 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#199 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1041 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#197 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#194 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#494 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#494 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#486 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#482 irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#216 (work_completion)(&peer->transmit_handshake_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#217 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#103 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#215 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#69 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#69 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#69 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg0#218 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &n->list_lock &c->lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#69 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#67 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#67 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#67 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#66 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#66 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#66 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#200 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#200 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1059 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1059 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1059 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#198 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex wq_pool_mutex.wait_lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#195 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#195 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#195 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1059 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#495 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#487 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#487 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#487 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#483 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#483 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#483 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#70 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#70 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#70 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#70 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#68 irq_context: 0 (wq_completion)wg-kex-wg0#215 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#67 irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#33 irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &tbl->lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &n->lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#8 irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#215 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#33 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#30 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#30 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#484 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#201 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#201 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#199 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#196 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1060 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1060 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1050 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1043 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#202 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#202 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#497 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#200 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#497 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#200 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#213 irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock key#27 irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &nf_conntrack_locks[i] &nf_conntrack_locks[i]/1 batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#213 (work_completion)(&peer->transmit_handshake_work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg0#216 irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#214 irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc4_nci_rx_wq#200 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#489 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#485 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#197 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1061 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1061 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1061 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#110 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#498 irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#498 irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#490 irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc3_nci_tx_wq#486 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#486 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#486 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1051 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1044 irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1044 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1044 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#71 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#69 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#68 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1062 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1062 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1052 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#203 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#203 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1045 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#201 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#198 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#499 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#72 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#72 irq_context: 0 (wq_completion)wg-kex-wg2#213 irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg2#213 (work_completion)(&peer->transmit_handshake_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock crngs.lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &cookie->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg1#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-kex-wg0#216 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#214 irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &wg->static_identity.lock &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount#2 irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &table->lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &list->lock#14 irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh &r->producer_lock#2 irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg2#214 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &r->consumer_lock#2 irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#7 irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &____s->seqcount#9 irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock rcu_read_lock_bh &list->lock#12 irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh tk_core.seq.seqcount irq_context: 0 cb_lock genl_mutex nl_table_wait.lock &p->pi_lock irq_context: 0 cb_lock rtnl_mutex &rdev->wiphy.mtx nl_table_wait.lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) tk_core.seq.seqcount irq_context: 0 (wq_completion)wg-crypt-wg1#107 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#70 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#70 &rq->__lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#107 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#70 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#69 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#499 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#491 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#487 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#500 irq_context: 0 sb_writers#8 &of->mutex kn->active#49 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#8 &of->mutex kn->active#49 rcu_read_lock &rcu_state.expedited_wq irq_context: 0 sb_writers#8 &of->mutex kn->active#49 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock irq_context: 0 sb_writers#8 &of->mutex kn->active#49 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock irq_context: 0 sb_writers#8 &of->mutex kn->active#49 rcu_read_lock &rcu_state.expedited_wq &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#492 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#492 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 crngs.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 devlinks.xa_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &xa->xa_lock#19 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &xa->xa_lock#19 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 pcpu_alloc_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 pcpu_alloc_mutex pcpu_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &base->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 pin_fs_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 fs_reclaim irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 rcu_read_lock rename_lock.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &dentry->d_lock &wq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 mmu_notifier_invalidate_range_start irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &s->s_inode_list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 tk_core.seq.seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &sb->s_type->i_lock_key#7 &dentry->d_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 pool_lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &n->list_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 batched_entropy_u32.lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex rtnl_mutex.wait_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rtnl_mutex &(&net->nexthop.notifier_chain)->rwsem irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock &data->fib_event_queue_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock rcu_read_lock &pool->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock rcu_node_0 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &n->list_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock &tb->tb6_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock &tb->tb6_lock &net->ipv6.fib6_walker_lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &table->lock#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock &tb->tb6_lock &data->fib_event_queue_lock irq_context: 0 (wq_completion)wg-kex-wg2#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock rcu_read_lock_bh &peer->keypairs.keypair_update_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock &tb->tb6_lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 rcu_read_lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &(&fn_net->fib_chain)->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &handshake->lock &c->lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &sb->s_type->i_mutex_key#3 &rq->__lock irq_context: 0 (wq_completion)wg-kex-wg1#212 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) rcu_read_lock_bh rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)wg-kex-wg1#211 (work_completion)(&peer->transmit_handshake_work) &cfs_rq->removed.lock irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &dir->lock#2 irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh rcu_read_lock &ul->lock irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) &base->lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) &base->lock &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh &obj_hash[i].lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) rcu_read_lock_bh &base->lock &obj_hash[i].lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &obj_hash[i].lock pool_lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &rq->__lock irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &____s->seqcount#2 irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 &____s->seqcount irq_context: 0 sb_writers#7 &of->mutex kn->active#47 nsim_bus_dev_list_lock &dev->mutex &devlink->lock_key#118 stack_depot_init_mutex irq_context: 0 (wq_completion)nfc3_nci_rx_wq#492 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#488 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#488 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#488 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1063 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#204 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#204 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#204 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1053 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1053 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1053 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#204 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#202 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#202 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#202 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#199 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#199 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#199 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1046 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1064 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1064 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1054 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#501 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#501 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1047 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#501 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#493 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#489 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#205 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#205 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#203 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#203 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#200 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1065 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1065 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1065 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1065 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1055 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1048 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#502 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#502 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#502 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#502 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#494 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#494 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#490 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1066 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1066 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1056 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1067 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1067 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1057 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1050 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#503 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#503 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#495 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#495 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#495 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#491 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#491 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#491 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1068 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1058 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1051 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#504 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#504 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#504 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#492 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1069 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1069 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1059 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1052 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1052 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1052 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1070 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1070 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1060 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1053 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#505 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#505 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#505 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#497 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#493 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#493 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#493 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#206 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#206 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#204 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#201 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#201 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1071 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1071 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1071 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1061 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1061 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1054 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#506 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#506 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#506 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1072 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1072 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1072 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#506 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#498 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#494 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1072 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1062 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1055 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#207 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#207 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#207 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#207 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#205 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#202 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1056 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1056 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1056 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#507 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#507 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#507 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#507 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#499 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#495 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1074 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#508 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#508 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#508 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#508 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#500 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#496 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1074 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1064 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1057 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1057 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1057 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1057 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#509 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#497 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#502 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#510 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1075 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1075 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1065 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1058 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#511 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#511 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#511 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#511 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#503 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#503 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#498 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#498 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1076 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1076 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1066 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1059 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1077 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1077 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1077 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1067 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1060 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1078 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1078 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1068 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1068 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1068 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1061 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#512 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#499 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#499 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#499 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&({ do { const void *__vpp_verify = (typeof((worker) + 0))((void *)0); (void)__vpp_verify; } while (0); ({ unsigned long __ptr; __ptr = (unsigned long) ((typeof(*((worker))) *)((worker))); (typeof((typeof(*((worker))) *)((worker)))) (__ptr + (((__per_cpu_offset[(cpu)])))); }); })->work) &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#513 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#513 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#505 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#500 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#500 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#500 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1079 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1069 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1062 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#208 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#208 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#208 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#208 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#206 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#203 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#203 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1080 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1080 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1080 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1070 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1063 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#514 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#514 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#514 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#514 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#506 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#501 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1071 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1081 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#209 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#209 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#209 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#209 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#207 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#204 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#507 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#502 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1082 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1082 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1072 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1064 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1064 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1064 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#210 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#210 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#208 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#73 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#73 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#71 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#205 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1083 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1083 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1073 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1065 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1065 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1065 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1084 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1084 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1084 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1074 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1066 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1066 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1066 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#503 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#503 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#503 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1085 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1085 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1085 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#211 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#211 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#211 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1085 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#211 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#211 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#211 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1075 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#209 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#206 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1067 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1067 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1067 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1076 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1086 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1087 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#517 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#517 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#517 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1087 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1077 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1068 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1068 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1068 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#517 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#509 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#504 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1078 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1069 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#518 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#518 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#518 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#510 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#510 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#510 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#505 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#505 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#505 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#212 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#212 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#212 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#212 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#210 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#210 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#210 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#207 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#207 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#207 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1089 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1089 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1079 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1070 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1070 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#519 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#519 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#511 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#506 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1090 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1090 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1090 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1090 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1080 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1071 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1071 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1091 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1091 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1081 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#520 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#520 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#520 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1072 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#520 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#512 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#507 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1092 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1092 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1082 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1082 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1073 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#521 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#521 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#513 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#508 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1093 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1093 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1093 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1093 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg0#110 (work_completion)(&peer->transmit_packet_work) &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1083 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1074 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1094 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1084 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1075 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#74 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#74 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#72 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#72 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#72 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#71 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#71 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#71 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#522 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#522 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#522 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#522 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#514 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#509 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#213 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#213 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#211 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#211 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#211 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#208 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#515 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#510 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#214 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#214 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#212 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#209 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1095 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1095 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1095 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1095 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1085 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1076 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#524 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#524 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#524 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#524 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#516 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1096 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1096 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1096 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#511 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#511 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#511 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1096 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1086 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1086 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1086 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#215 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#213 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#210 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#75 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#75 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#73 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#72 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#72 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1097 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1087 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1078 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#525 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#525 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#517 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#512 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#512 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#512 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1098 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#526 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#526 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#526 &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#526 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#526 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#518 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#513 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#216 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#216 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#214 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#211 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1099 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1099 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1099 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1089 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1080 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1100 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1100 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1090 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1081 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#527 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#527 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#527 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#519 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#514 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1101 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1091 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1082 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#528 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#528 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#520 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#515 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1102 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1092 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1092 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#529 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#529 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#529 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#529 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1083 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#521 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#521 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#521 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#516 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#217 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#217 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#215 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#215 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#215 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#212 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#212 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#212 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1093 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1084 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1084 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_tx_wq#1085 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1094 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1104 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1105 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1105 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1105 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1105 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#530 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1095 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#530 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1095 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#530 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1095 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#530 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#522 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1095 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1095 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1086 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#517 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1086 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1086 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1106 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1106 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1106 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1106 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1096 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1087 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#531 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#531 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#523 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#518 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#518 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#518 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#218 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#218 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#76 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#76 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#76 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#76 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#74 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#73 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#73 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1107 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1107 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1107 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1107 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1097 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1097 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1088 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1088 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1088 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#532 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#532 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#524 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#524 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#524 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#219 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#219 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#217 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1108 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1108 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1108 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1108 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1098 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1089 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#214 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1089 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1089 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1109 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1109 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1109 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1099 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1090 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1110 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1110 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1110 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1100 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#525 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#520 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1111 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1111 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1111 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1111 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1101 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1092 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#220 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#220 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#534 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#534 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#534 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#526 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#521 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#221 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#221 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#221 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1112 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1112 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1112 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#221 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#219 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#219 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#219 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1112 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#216 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1102 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1102 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1102 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1093 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1093 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1093 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#535 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#535 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#535 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#535 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#527 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1113 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1113 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1113 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1113 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1094 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1094 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1094 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1114 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1114 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1114 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1114 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1104 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1104 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1104 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#536 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#536 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#528 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#523 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#523 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#523 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#217 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#217 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#217 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#537 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#537 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#537 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1115 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1115 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#537 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1105 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1096 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#529 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#529 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#529 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#524 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#77 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#77 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#77 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#77 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#77 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#77 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#75 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#75 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#75 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#74 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#74 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#74 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1116 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1116 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1116 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#34 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#34 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1116 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1106 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#34 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#34 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#34 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#31 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1097 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#31 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1117 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1117 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1107 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1107 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1107 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#223 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#223 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#221 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#221 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#221 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#538 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#538 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#538 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#538 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#538 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#218 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#530 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#530 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#525 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1118 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1118 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1118 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#224 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#224 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#224 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#224 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#222 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#219 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1118 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1108 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1099 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1119 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1119 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1119 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1119 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#539 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#539 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1109 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#531 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1109 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#531 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1100 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#526 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#225 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#225 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#225 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#225 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#223 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#220 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1120 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1120 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#541 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#541 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#541 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1110 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#541 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#532 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#532 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#532 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1101 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#532 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#532 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#527 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#226 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#226 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#226 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#78 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#78 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#78 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#78 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#76 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#75 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#224 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#221 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#221 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#221 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1111 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1111 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1111 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#542 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#542 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1111 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1111 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1102 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#543 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#534 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#529 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#227 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#227 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#227 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1122 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1122 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#222 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1112 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1112 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1112 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1103 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#79 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#79 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#77 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#77 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#77 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#76 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#544 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#76 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#76 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#544 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#544 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#35 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#35 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#544 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#35 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#535 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#535 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#535 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#530 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#32 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#32 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#530 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#32 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#530 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#32 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#32 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#32 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1123 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1123 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1123 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1123 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1113 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1104 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1124 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1124 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1114 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1105 irq_context: 0 misc_mtx &rq->__lock &obj_hash[i].lock irq_context: 0 misc_mtx &rq->__lock &base->lock irq_context: 0 misc_mtx &rq->__lock &base->lock &obj_hash[i].lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1115 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1125 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#545 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#545 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#531 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1126 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1126 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1126 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1126 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1116 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1116 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1116 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1106 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1106 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1127 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1117 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1107 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#546 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1128 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1118 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#537 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1108 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#532 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1129 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1129 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1119 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1109 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#547 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#547 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#547 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#547 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#538 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#538 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#538 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#533 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#533 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#533 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#80 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#80 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#80 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#78 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#78 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#78 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#77 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#77 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#77 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#228 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#226 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#533 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#533 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#223 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#223 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#223 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1130 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1130 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1130 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1130 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1120 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1120 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1120 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1110 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1110 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1110 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#548 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#539 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#534 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1131 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1121 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1111 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1111 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#549 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#540 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#535 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1132 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1122 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1112 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1133 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1123 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1113 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#550 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#550 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#550 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#541 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#536 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#536 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1134 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1134 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1134 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1134 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#551 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#551 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#551 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#551 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1124 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1124 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1124 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#542 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#537 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#537 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#537 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1114 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1114 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1114 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1135 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1135 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1135 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1125 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1125 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1125 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1115 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1116 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1137 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1137 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1127 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1127 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1127 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1117 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#552 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1117 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1117 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#552 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#552 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#543 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#538 irq_context: 0 misc_mtx (wq_completion)nfc5_nci_tx_wq#78 irq_context: 0 misc_mtx (wq_completion)nfc5_nci_rx_wq#79 irq_context: 0 misc_mtx (wq_completion)nfc5_nci_rx_wq#79 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc5_nci_rx_wq#79 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#81 irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#81 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#229 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#229 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#224 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1138 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1138 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1128 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1128 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1128 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1118 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#553 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#553 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#553 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#553 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#544 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#539 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#539 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#539 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#230 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#230 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#228 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#225 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1139 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1139 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1139 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1139 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1129 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1129 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1129 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1119 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1119 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1119 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1140 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1140 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1140 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1130 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1120 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1120 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1120 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1141 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1141 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1141 irq_context: 0 misc_mtx &wq->mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#554 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#554 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#545 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#540 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1143 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1143 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1143 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1143 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1132 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1133 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1123 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1145 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1145 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1134 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1124 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1146 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1146 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1146 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1146 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1135 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1125 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#231 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#231 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#229 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#82 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#82 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#555 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#555 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#555 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#82 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#80 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#555 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#546 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#79 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#541 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#36 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#36 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#36 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#33 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#33 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#33 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#33 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1147 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1147 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1147 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1136 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1136 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1136 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1126 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1126 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1126 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#556 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#556 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#556 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#556 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#547 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#542 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#232 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#232 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#232 &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#232 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1148 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1148 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1148 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1148 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1137 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1127 irq_context: 0 (wq_completion)wg-crypt-wg1#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 (wq_completion)wg-crypt-wg2#106 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#557 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#557 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#548 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#548 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#548 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#543 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1149 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1149 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1149 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1138 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1128 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1128 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1128 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#558 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#558 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#558 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#558 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#558 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#558 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1150 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1150 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1150 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1150 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#549 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#544 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1139 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1139 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1129 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1140 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1130 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1130 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1130 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#559 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#559 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#559 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#559 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#233 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#233 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#230 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#227 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1152 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1141 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1131 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#546 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#83 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#83 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1153 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#83 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1153 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1153 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1153 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1142 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1132 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1132 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#83 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#234 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#234 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#81 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#231 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#228 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#228 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#228 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#80 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#561 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#561 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#561 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#552 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#552 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#552 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#547 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#547 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#547 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1154 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#84 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1154 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1154 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#84 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#82 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1154 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1143 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1143 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1133 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#81 irq_context: 0 &type->i_mutex_dir_key#5 rcu_read_lock &rcu_state.gp_wq &p->pi_lock &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#562 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#562 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#553 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#548 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#548 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#548 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#235 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#232 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#229 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1155 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1144 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1144 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1134 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#563 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1134 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1134 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#563 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#563 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#563 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#554 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#549 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#236 irq_context: 0 misc_mtx (wq_completion)nfc5_nci_cmd_wq#85 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#236 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#236 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#236 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#233 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#230 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#230 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#230 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1156 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1156 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1145 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1135 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1135 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#565 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#565 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#565 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#555 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#555 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#555 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#550 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#550 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#550 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1157 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1157 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#237 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#237 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#234 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1136 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#231 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1147 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1147 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1147 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1137 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1137 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1137 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#566 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#566 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#566 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#566 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#556 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#556 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#556 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#37 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#37 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#34 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#34 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#34 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#551 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#551 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#551 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#34 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#238 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#238 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#238 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#235 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#232 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#232 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#232 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc2_nci_rx_wq#1148 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1159 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1159 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1159 &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1159 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#86 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#83 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#82 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#557 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#557 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#557 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#552 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#552 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#552 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#552 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#239 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#239 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#236 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#233 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1160 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1160 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1160 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1149 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1149 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1149 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1138 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1138 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#87 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#87 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#87 irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &cfs_rq->removed.lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex &obj_hash[i].lock irq_context: 0 &q->sysfs_dir_lock &q->sysfs_lock &q->rq_qos_mutex pool_lock#2 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#84 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#83 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#568 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#568 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#568 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#568 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#568 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#568 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1161 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1161 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1161 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1161 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1150 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1150 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1150 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#558 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#558 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#558 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1139 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#553 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1139 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#553 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1139 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#553 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#240 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#240 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#237 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#234 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#234 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#234 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1151 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1140 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1163 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1163 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1163 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1152 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1141 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#569 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#569 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#559 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#559 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#554 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#570 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#570 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#560 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#560 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1164 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1164 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1153 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1153 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#88 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#88 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#85 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1142 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#241 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#241 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#238 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#238 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#238 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1154 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1154 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1154 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#571 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#571 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#571 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1143 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#561 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#556 irq_context: 0 misc_mtx cpu_hotplug_lock wq_pool_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#572 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#572 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#572 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1166 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1166 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1155 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1155 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1155 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1144 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#573 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#573 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#573 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#242 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#242 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#239 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#239 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#239 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#236 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#573 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#557 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#557 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#557 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#38 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#38 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#38 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#35 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#35 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#35 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#89 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#89 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#89 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#89 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#86 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#86 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#86 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#85 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1167 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1167 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1167 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1156 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1145 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#574 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#574 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#563 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#558 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1168 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1168 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1157 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1146 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_tx_wq#559 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_tx_wq#559 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#564 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#564 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_rx_wq#564 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#575 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#575 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#575 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1170 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1170 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1170 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#576 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1170 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#577 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1158 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#577 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1158 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1158 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#565 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#560 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1147 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1148 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1172 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1172 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1160 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1149 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1173 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1173 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1173 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1173 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1161 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1161 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1161 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#578 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#578 &rq->__lock cpu_asid_lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1150 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#578 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#566 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#561 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1174 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1162 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1151 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1151 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1151 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#243 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#243 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#579 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#579 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#240 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#240 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#240 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#237 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#567 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#562 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1175 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1175 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1163 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1152 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#580 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#580 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#580 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#580 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#568 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#563 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#244 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#244 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#241 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#238 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#238 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#238 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1164 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1153 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#581 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1177 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1165 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#569 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1154 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1154 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1154 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#564 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1166 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1166 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1155 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1155 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#582 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#582 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#582 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#582 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#582 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#582 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#570 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#570 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#570 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#565 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#565 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#565 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#245 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#245 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#245 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#245 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#242 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#242 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#239 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#583 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#583 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#246 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#243 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#240 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#240 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#240 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#583 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#571 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1179 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1179 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1179 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1179 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1167 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1156 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#90 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#90 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#90 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#90 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#87 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#86 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#86 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1180 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1180 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1180 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1168 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1168 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1168 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1157 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1157 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1157 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#584 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#584 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#584 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#584 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#572 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#572 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#567 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1181 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1181 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1169 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1158 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#247 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#247 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#247 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1182 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1182 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1182 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1170 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#247 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#244 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#244 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#244 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#241 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1159 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#585 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#585 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#585 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#573 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#568 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#568 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#568 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#117 &nsim_trap_data->trap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1183 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1183 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#248 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#248 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#248 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#248 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#245 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#245 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#245 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1171 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1171 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#242 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1160 irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock irq_context: 0 &dev->mutex dev_pm_qos_sysfs_mtx &root->kernfs_rwsem rcu_read_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#115 &nsim_trap_data->trap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc3_nci_rx_wq#574 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#569 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &hdev->cmd_sync_work_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &conn->lock#2 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_cb_list_lock (work_completion)(&(&conn->timeout_work)->work) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &x->wait#2 irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1184 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1184 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &obj_hash[i].lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock rcu_read_lock &pool->lock &p->pi_lock &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->disc_work)->work) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->auto_accept_work)->work) irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock (work_completion)(&(&conn->idle_work)->work) irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1184 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1172 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1172 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1172 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1161 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1161 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1161 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock fs_reclaim mmu_notifier_invalidate_range_start irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock &c->lock irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock tk_core.seq.seqcount irq_context: 0 (wq_completion)hci3#7 (work_completion)(&hdev->cmd_sync_work) &hdev->req_lock &hdev->lock hci_sk_list.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#587 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#587 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#587 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1185 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1185 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1185 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#587 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#587 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#587 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1173 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#575 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#570 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1162 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#249 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#249 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#246 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#243 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1186 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1186 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1186 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1186 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1174 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1174 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1174 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1163 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#588 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#588 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#588 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#588 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#576 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#571 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1187 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1175 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1164 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1164 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1164 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1165 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1165 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#589 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#589 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#577 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#577 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#572 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1189 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1189 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1189 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1189 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1189 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1177 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1166 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1190 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1190 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1190 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1190 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1178 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1178 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1167 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#578 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#578 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#573 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#573 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#573 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1191 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1191 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1191 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1191 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1179 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1179 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1168 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1168 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#579 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#574 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1192 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1192 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1192 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)wg-crypt-wg1#106 (work_completion)(&peer->transmit_packet_work) batched_entropy_u8.lock crngs.lock irq_context: 0 (wq_completion)wg-crypt-wg2#105 (work_completion)(&peer->transmit_packet_work) &peer->endpoint_lock rcu_read_lock_bh batched_entropy_u32.lock crngs.lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1192 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1180 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#592 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1180 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1180 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#592 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#592 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#592 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#580 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#575 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1169 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1193 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1193 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1193 irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &cfs_rq->removed.lock irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &obj_hash[i].lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1181 irq_context: 0 sb_writers#7 tomoyo_ss rcu_read_lock &rcu_state.gp_wq &p->pi_lock &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1170 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1194 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1182 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1182 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1182 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1171 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1195 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1195 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1195 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1183 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1172 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1196 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1196 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1184 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1184 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1184 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1173 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1173 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1173 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#593 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#593 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#581 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#576 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#576 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#576 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1197 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1197 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1197 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1185 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#594 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#594 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#582 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#582 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#582 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#577 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#250 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#250 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#250 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#244 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1198 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1186 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1175 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#595 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#595 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#583 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#578 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1199 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1199 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1187 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1176 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#596 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#584 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#579 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#597 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#597 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#585 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#585 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#585 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#580 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#580 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#580 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#251 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#251 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#251 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#251 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#248 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#245 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#599 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#599 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#586 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1200 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#586 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#586 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1200 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1200 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#581 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#581 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#581 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1200 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1188 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1188 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1188 &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1188 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1177 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1177 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1177 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1178 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#600 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#600 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#587 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#587 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#587 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#582 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1202 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1202 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1202 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1190 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1179 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#601 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#601 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#601 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#588 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#583 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#583 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#583 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1203 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1203 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1191 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1180 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#602 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#602 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#602 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#602 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#589 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#589 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#589 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#584 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1204 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1192 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1192 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1192 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1181 irq_context: 0 &fc->uapi_mutex virtio_fs_mutex irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#252 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#252 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#252 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#252 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#249 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#249 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#249 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#246 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#246 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#246 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#246 &obj_hash[i].lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1205 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1205 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1205 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1205 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1193 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1193 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1182 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1206 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1206 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1206 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1194 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1194 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1194 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1183 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#603 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#603 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#590 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#585 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#585 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#585 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#585 &rq->__lock cpu_asid_lock irq_context: 0 &fc->uapi_mutex &rq->__lock &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#23 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#22 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#22 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#21 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#21 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_tx_wq#21 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#21 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#21 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)events (work_completion)(&(&nsim_dev->trap_data->trap_report_dw)->work) &devlink->lock_key#116 &nsim_trap_data->trap_lock &pcp->lock &zone->lock &____s->seqcount irq_context: 0 (wq_completion)nfc9_nci_tx_wq#21 irq_context: 0 (wq_completion)nfc9_nci_tx_wq#21 &rq->__lock irq_context: 0 (wq_completion)nfc21_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc21_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc21_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc19_nci_rx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc19_nci_tx_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc19_nci_rx_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc19_nci_tx_wq#18 irq_context: 0 (wq_completion)nfc34_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc34_nci_rx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc34_nci_rx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc34_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#19 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc19_nci_tx_wq#19 irq_context: 0 (wq_completion)nfc19_nci_tx_wq#19 &rq->__lock irq_context: 0 (wq_completion)nfc19_nci_tx_wq#19 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc33_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc33_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc33_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc33_nci_rx_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc33_nci_rx_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc33_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc32_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc32_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc32_nci_rx_wq#14 irq_context: 0 (wq_completion)nfc32_nci_tx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc31_nci_cmd_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc31_nci_cmd_wq#14 irq_context: 0 (wq_completion)nfc31_nci_rx_wq#14 irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc31_nci_rx_wq#14 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc35_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc31_nci_tx_wq#14 irq_context: 0 (wq_completion)nfc31_nci_tx_wq#14 &rq->__lock irq_context: 0 (wq_completion)nfc31_nci_tx_wq#14 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc30_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc30_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc30_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc30_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc30_nci_tx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc30_nci_tx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#20 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc29_nci_cmd_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc29_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc29_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc29_nci_rx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc29_nci_rx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc35_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc29_nci_tx_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc35_nci_rx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc28_nci_cmd_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc28_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc28_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc28_nci_rx_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc28_nci_rx_wq#16 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc28_nci_tx_wq#16 irq_context: 0 (wq_completion)nfc35_nci_tx_wq#15 irq_context: 0 (wq_completion)nfc28_nci_tx_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#20 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#20 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#20 &rq->__lock irq_context: 0 (wq_completion)nfc19_nci_rx_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc27_nci_cmd_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc27_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc27_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc27_nci_rx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc27_nci_tx_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#21 irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq#15 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc34_nci_cmd_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#21 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc26_nci_cmd_wq#21 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc26_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc26_nci_rx_wq#21 irq_context: 0 (wq_completion)nfc26_nci_tx_wq#21 irq_context: 0 (wq_completion)nfc34_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc34_nci_rx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc25_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc34_nci_rx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc25_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc34_nci_rx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc25_nci_rx_wq#15 &rq->__lock irq_context: 0 (wq_completion)nfc25_nci_rx_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc25_nci_tx_wq#15 irq_context: 0 (wq_completion)nfc34_nci_tx_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#15 irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#15 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc24_nci_cmd_wq#15 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc24_nci_cmd_wq#15 irq_context: 0 (wq_completion)nfc24_nci_rx_wq#15 irq_context: 0 (wq_completion)nfc24_nci_tx_wq#15 irq_context: 0 (wq_completion)nfc19_nci_tx_wq#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc23_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc23_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc23_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc23_nci_tx_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#17 irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#17 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc21_nci_cmd_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc21_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc21_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc21_nci_tx_wq#17 irq_context: 0 (wq_completion)nfc21_nci_tx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc21_nci_tx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc22_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc22_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc22_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc22_nci_tx_wq#17 irq_context: 0 (wq_completion)nfc22_nci_tx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc22_nci_tx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc20_nci_cmd_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc20_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc20_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc20_nci_rx_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc20_nci_tx_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#18 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc18_nci_cmd_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc18_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc18_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc18_nci_tx_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc17_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc17_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc17_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc17_nci_tx_wq#16 irq_context: 0 (wq_completion)nfc17_nci_tx_wq#16 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc19_nci_cmd_wq#21 irq_context: 0 &ndev->req_lock (wq_completion)nfc16_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc16_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc16_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc16_nci_rx_wq#19 &rq->__lock irq_context: 0 (wq_completion)nfc16_nci_tx_wq#19 irq_context: 0 (wq_completion)nfc19_nci_cmd_wq#21 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#21 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#21 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc19_nci_rx_wq#21 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc15_nci_cmd_wq#17 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#17 irq_context: 0 (wq_completion)nfc15_nci_tx_wq#17 &rq->__lock irq_context: 0 (wq_completion)nfc15_nci_tx_wq#17 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc15_nci_tx_wq#17 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc15_nci_tx_wq#17 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc19_nci_tx_wq#21 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#20 irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#20 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc14_nci_cmd_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc14_nci_cmd_wq#20 irq_context: 0 (wq_completion)nfc14_nci_rx_wq#20 irq_context: 0 (wq_completion)hci3#8 (work_completion)(&(&conn->disc_work)->work) &hdev->unregister_lock &rq->__lock irq_context: 0 (wq_completion)nfc14_nci_tx_wq#20 &rq->__lock irq_context: 0 (wq_completion)nfc14_nci_tx_wq#20 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#18 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc15_nci_cmd_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc15_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc15_nci_rx_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc15_nci_rx_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc15_nci_tx_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#22 irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#22 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc9_nci_cmd_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_cmd_wq#22 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#22 irq_context: 0 (wq_completion)nfc9_nci_rx_wq#22 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_rx_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc9_nci_tx_wq#22 &rq->__lock irq_context: 0 (wq_completion)nfc9_nci_tx_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc13_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc13_nci_cmd_wq#16 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#16 irq_context: 0 (wq_completion)nfc13_nci_rx_wq#16 &rq->__lock irq_context: 0 (wq_completion)nfc13_nci_tx_wq#16 irq_context: 0 &ndev->req_lock (wq_completion)nfc12_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc12_nci_cmd_wq#19 irq_context: 0 (wq_completion)nfc12_nci_rx_wq#19 irq_context: 0 (wq_completion)nfc12_nci_tx_wq#19 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#24 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#24 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#22 irq_context: 0 misc_mtx (wq_completion)nfc9_nci_rx_wq#23 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#22 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc9_nci_rx_wq#23 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc9_nci_rx_wq#23 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc9_nci_rx_wq#23 &cfs_rq->removed.lock irq_context: 0 misc_mtx (wq_completion)nfc9_nci_rx_wq#23 &obj_hash[i].lock irq_context: 0 misc_mtx (wq_completion)nfc9_nci_cmd_wq#23 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#22 irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#22 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc11_nci_cmd_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_cmd_wq#22 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#22 irq_context: 0 (wq_completion)nfc11_nci_rx_wq#22 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_rx_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc11_nci_tx_wq#22 irq_context: 0 (wq_completion)nfc11_nci_tx_wq#22 &rq->__lock irq_context: 0 (wq_completion)nfc11_nci_tx_wq#22 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (work_completion)(&local->sdreq_timeout_work) &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc10_nci_cmd_wq#18 irq_context: 0 (wq_completion)nfc10_nci_cmd_wq#18 irq_context: 0 &ndev->req_lock (wq_completion)nfc8_nci_cmd_wq#25 irq_context: 0 (wq_completion)nfc8_nci_cmd_wq#25 irq_context: 0 (wq_completion)nfc10_nci_rx_wq#18 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#18 irq_context: 0 (wq_completion)nfc10_nci_tx_wq#18 &rq->__lock irq_context: 0 (wq_completion)nfc10_nci_tx_wq#18 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc7_nci_cmd_wq#23 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#24 irq_context: 0 (wq_completion)nfc7_nci_cmd_wq#23 irq_context: 0 (wq_completion)nfc8_nci_rx_wq#24 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_rx_wq#24 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_rx_wq#24 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc7_nci_rx_wq#23 irq_context: 0 (wq_completion)nfc7_nci_tx_wq#23 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#39 irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#39 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#36 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#36 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_rx_wq#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc8_nci_tx_wq#23 irq_context: 0 (wq_completion)nfc8_nci_tx_wq#23 &rq->__lock irq_context: 0 (wq_completion)nfc8_nci_tx_wq#23 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_tx_wq#36 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#36 &rq->__lock irq_context: 0 (wq_completion)nfc6_nci_tx_wq#36 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#91 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#91 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#88 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#88 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#88 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_rx_wq#88 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#88 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#87 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#87 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_tx_wq#87 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#253 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#253 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#253 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#253 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#250 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#247 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#247 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#247 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#604 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#604 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#591 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#586 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1207 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1207 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1195 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1184 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1208 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1208 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1196 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1196 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1185 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1185 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#92 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#92 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#92 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1209 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1209 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1209 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#254 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#254 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#254 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#254 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#251 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#251 &rq->__lock irq_context: 0 (wq_completion)nfc5_nci_rx_wq#89 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1209 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1197 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#88 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1186 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#248 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#605 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#605 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#592 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#587 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1210 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1210 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1210 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1198 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#606 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#606 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1187 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1187 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1187 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_rx_wq#593 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#588 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#255 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#255 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#255 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#252 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#252 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#249 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1211 irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1211 &rq->__lock irq_context: 0 misc_mtx (wq_completion)nfc2_nci_cmd_wq#1211 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1212 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1212 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1199 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1199 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1199 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1213 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1213 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1200 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1189 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1214 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1214 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1201 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1201 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1190 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#607 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#607 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#589 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1215 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1202 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1202 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1202 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#256 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#256 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#608 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#608 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1216 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1216 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#608 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#595 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1203 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1192 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1192 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1192 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#590 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#257 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#257 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#257 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#254 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#254 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#254 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#251 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#251 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1217 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1204 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1193 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1193 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1193 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1193 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1193 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#596 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#591 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#591 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#591 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#610 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#610 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#610 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#597 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#592 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#592 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#592 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1218 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1218 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1218 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1205 irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#40 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1205 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1205 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#40 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc6_nci_cmd_wq#40 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1194 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1194 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1194 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc6_nci_cmd_wq#40 irq_context: 0 (wq_completion)nfc6_nci_rx_wq#37 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#258 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#258 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#258 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#258 irq_context: 0 (wq_completion)nfc6_nci_tx_wq#37 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#255 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#252 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#252 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#252 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#611 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#611 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#611 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#593 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#259 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#259 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1219 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#259 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1219 irq_context: 0 misc_mtx (wq_completion)nfc3_nci_cmd_wq#612 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1206 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1195 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#93 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#93 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#93 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#93 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#90 irq_context: 0 (wq_completion)nfc5_nci_tx_wq#89 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#613 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#613 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#594 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#614 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#614 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#600 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#600 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#600 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#595 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1220 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1220 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1220 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1220 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1207 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1207 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1207 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1196 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#261 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#261 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#261 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#257 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#253 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#253 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#253 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1221 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1208 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1208 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1197 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#615 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#615 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#615 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#601 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#596 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#262 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#262 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#258 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#258 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#258 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#254 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#254 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1222 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1222 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1209 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1198 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1223 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1210 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1199 irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#616 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#602 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#602 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_rx_wq#602 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc3_nci_tx_wq#597 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#597 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#597 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#263 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#263 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#263 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#259 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#259 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#617 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#617 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#617 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#259 &cfs_rq->removed.lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#259 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#617 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#603 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#598 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#255 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1224 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1211 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1211 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1211 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_rx_wq#260 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#260 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_rx_wq#260 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#94 irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#94 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#256 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#91 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1225 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1225 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1225 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_tx_wq#90 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#618 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#618 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#618 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1225 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1212 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#618 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#618 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#618 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1201 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#604 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#599 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1226 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1226 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1226 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1226 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1213 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1213 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1202 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1202 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1203 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1228 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1228 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1228 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#265 irq_context: 0 misc_mtx (wq_completion)nfc4_nci_cmd_wq#266 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1215 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1215 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1215 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1204 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1204 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1204 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#619 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#619 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#619 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#605 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#600 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#267 irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#267 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#267 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#267 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#261 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#261 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#620 irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#620 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#620 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc4_nci_tx_wq#257 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1229 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1229 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1229 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1229 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1216 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1216 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1216 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#620 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#620 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#620 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#601 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1230 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1230 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1230 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1230 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1217 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1206 irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1231 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1231 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1218 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1218 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1207 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1232 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1219 irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1219 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_rx_wq#1219 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1208 irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1208 &rq->__lock irq_context: 0 (wq_completion)nfc2_nci_tx_wq#1208 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc4_nci_cmd_wq#268 irq_context: 0 (wq_completion)nfc4_nci_cmd_wq#268 irq_context: 0 (wq_completion)nfc4_nci_rx_wq#262 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#258 irq_context: 0 (wq_completion)nfc4_nci_tx_wq#258 &rq->__lock irq_context: 0 (wq_completion)nfc4_nci_tx_wq#258 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#621 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#621 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#621 &cfs_rq->removed.lock irq_context: 0 &ndev->req_lock (wq_completion)nfc3_nci_cmd_wq#621 &obj_hash[i].lock irq_context: 0 (wq_completion)nfc3_nci_cmd_wq#621 irq_context: 0 (wq_completion)nfc3_nci_rx_wq#607 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#602 irq_context: 0 (wq_completion)nfc3_nci_tx_wq#602 &rq->__lock irq_context: 0 (wq_completion)nfc3_nci_tx_wq#602 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#95 irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#95 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc5_nci_cmd_wq#95 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1233 &rq->__lock irq_context: 0 &ndev->req_lock (wq_completion)nfc2_nci_cmd_wq#1233 &rq->__lock &per_cpu_ptr(group->pcpu, cpu)->seq irq_context: 0 (wq_completion)nfc5_nci_cmd_wq#95 irq_context: 0 (wq_completion)nfc5_nci_rx_wq#92 irq_context: 0 (wq_completion)nfc2_nci_cmd_wq#1233